About
Taming Large Language Models @ SIGDIAL 2023 & INLG 2023
The goal of our workshop is to provide a common ground for experts working on different models and tasks related to controllable text generation. The interplay between research on controlled-generation and instruction-based learning paradigms sits at the heart of our discussions. Here, they can share task-agnostic solutions, engage in evaluations, and discuss task-specific challenges. By encouraging communication, collaboration, and research amongst attendees, the workshop intends to foster insightful debates about the recent progress of the NLG community.
Next iteration of this workshop series is to be held at ICLR 2024. Checkout LLM Agents
The proceedings of the workshop will be published by the ACL Anthology. We eagerly anticipate your valuable contributions and look forward to seeing you there!
Topics of interest include, but are not limited to:
- Alignment: Investigating techniques to better align LLMs with human values and intentions, including reward modeling, human-in-the-loop systems, and quantifying alignment metrics. Understanding the objectives pursued by a model and aligning them with human preferences are key challenges.
- In-context Learning: Exploring the role of context in LLMs, including how to improve context understanding, manage context drift, and enhance context-aware responses.
- Instruction-based Control: Comparing popular controlling mechanisms, including approaches such as logit manipulation, decoder mixing, and classifier guidance, amongst others, against the simpler instruction-based control.
- Generality: Investigating controllable techniques that work cross-task and dataset.
- Safety and Robustness: Assessing potential risks and vulnerabilities in LLMs, along with solutions such as adversarial training, safe exploration, and monitoring model behavior during deployment.
- Controllability vs. Robustness: Developing methods to better understand LLMs' decision-making processes, and how it acts in grounded scenarios. Understanding its reliance on implicit vs. explicit memory.
- Scalability and efficiency: Investigating novel approaches for reducing computational requirements for achieving control in LLMs.
The TamingLLM Workshop will be co-located with SIGDial 2023 and INLG 2023! We look forward to your participation in this exciting workshop!
Venue:
The SIGDial-INLG 2023 conference will take place in Prague, Czech Republic in OREA Hotel Pyramida, Prague.
The hotel is a 20-minute tram ride away from the city center, within walking distance of the Prague Castle.
Address:
OREA Hotel Pyramida, Bělohorská 24, 169 01 Praha 6, Czechia.
CFP
Call for Papers
- Submission deadline: July 7, 2023
- Author notification: July 21, 2023
- Camera-ready deadline: August 14, 2023
- Workshop date: September 12, 2023
-
All deadlines are 11.59 pm AOE time.
- Long papers (8 pages + references)
- Short papers (4 pages + references)
- The proceedings will be published by ACL Anthology.
- All long, short, and abstract submissions must follow the two-column ACL format, which is available as an Overleaf template and also downloadable directly (Latex and Word). Please refer to the SIGDIAL 2023 website for the most recent version of the templates.
- Submissions must conform to the official ACL style guidelines, which are contained in these templates. Submissions must be electronic, in PDF format.
- All submissions should be anonymized to facilitate double-blind reviewing.
- Submissions that do not adhere to the author guidelines or ACL policies will be rejected without review.
- Appendix should be added to the main document after references.
- Appendix does not count towards the page length.
Talks
Keynote Speakers
Shafiq Joty
Research Director @ Salesforce AI & Associate Professor @ NTUDaphne Ippolito
Assistant Professor @ CMUNancy F. Chen
Senior Principal Scientist, Principal Investigator, Group Leader @ I2R, A*STARNanyun (Violet) Peng
Assistant Professor @ UCLAEric Malmi
Research Scientist @ Google ZürichYuandong Tian
Research Scientist @ Meta AI ResearchNazneen Fatema Rajani
Robustness Research Lead @ Hugging FaceAccepted Papers
Accepted Papers
Schedule
Workshop Schedule
12 Sep 2023 |
|||
Time in Prague | PST | Time in China | |
09:30-12:00 | 00:30-03:00 | 15:30-18:00 | Keynote Session I |
09:30-10:00 | 00:30-01:00 | 15:30-16:00 | Shafiq Joty |
10:00-10:30 | 01:00-01:30 | 16:00-16:30 | Daphne Ippolito |
10:30-11:00 | 01:30-02:00 | 16:30-17:00 | Nancy F. Chen |
11:00-11:30 | 02:00-02:30 | 17:00-17:30 | Break (coffee provided) |
11:30-12:00 | 02:30-03:00 | 17:30-18:00 | Nanyun (Violet) Peng |
12:00-13:45 | 03:00-04:45 | 18:00-19:45 | Lunch Break |
13:45-15:30 | 04:45-06:30 | 19:45-21:30 | Oral Session |
13:45-14:00 | 04:45-05:00 | 19:45-20:00 | Style Locality for Controllable Generation with kNN Language Models Gilles Nawezi, Lucie Flek, Charles Welch |
14:00-14:15 | 05:00-05:15 | 20:00-20:15 | CST5: Data Augmentation for Code-Switched Semantic Parsing Anmol Agarwal, Jigar Gupta, Rahul Goel, Shyam Upadhyay, Pankaj Joshi, Rengarajan Aravamudhan |
14:15-14:30 | 05:15-05:30 | 20:15-20:30 | PandaGPT: One Model To Instruction-Follow Them All Yixuan Su, Tian Lan, Huayang Li, Jialu Xu, Yan Wang, Deng Cai |
14:30-14:45 | 05:30-05:45 | 20:30-20:45 | Emotion-Conditioned Text Generation through Automatic Prompt Optimization Yarik Menchaca Resendiz, Roman Klinger |
14:45-15:00 | 05:45-06:00 | 20:45-21:00 | Mitigating Harms of LLMs via Knowledge Distillation for a Virtual Museum Tour Guide Ashley Lewis, Michael White |
15:00-15:15 | 06:00-06:15 | 21:00-21:15 | Evaluating Large Language Models for Document-grounded Response Generation in Information-Seeking Dialogues Norbert Braunschweiler, Rama Doddipatla, Simon Keizer, Svetlana Stoyanchev |
15:15-15:30 | 06:15-06:30 | 21:15-21:30 | Enhancing Pipeline-Based Conversational Agents with Large Language Models Mina Foosherian, Hendrik Purwins, Purna Rathnayake, Touhidul Alam, Rui Teimao, Klaus-Dieter Thoben |
15:30-16:30 | 06:30-07:30 | 21:30-22:30 | Break (coffee provided) |
16:30-18:00 | 07:30-09:00 | 22:30-24:00 | Keynote Session II |
16:30-17:00 | 07:30-08:00 | 22:30-23:00 | Eric Malmi |
17:00-17:30 | 08:00-08:30 | 23:00-23:30 | Yuandong Tian |
17:30-18:00 | 08:30-09:00 | 23:30-24:00 | Nazneen Fatema Rajani |
We accommodate hybrid presentations: the workshop rooms will include a camera & audio equipment.
Zoom link: ZOOM (https://yale.zoom.us/my/yale.cs).
We welcome anyone to be a remote participant via Zoom and virtual registration is free (which will be available formally after August 10).
Organization
Workshop Organizers
Organizing Commitee
Xiangru Robert Tang
Ph.D. student @ Yale UniversityDevamanyu Hazarika
Applied Scientist @ AmazonDi Jin
Senior Applied Scientist @ AmazonChao-Wei Huang
PhD Student @ National Taiwan UniversitySherry Tongshuang Wu
Assistant Professor @ CMUAdvising Commitee
Bill Yuchen Lin
Postdoctoral Young Investigator @ AI2Deng Cai
Senior Researcher @ Tencent AI LabArman Cohan
Assistant Professor @ Yale UniversityAsli Celikyilmaz
Senior Research Manager @ MetaDilek Hakkani-Tür
Senior Principal Scientist @ AmazonPublicity Committee
Publicity Committee
Chunlin Lu, Central South University
Jinxuan Zhou, Central South University
Program Committee
Program Committee
Guo Zhang, Massachusetts Institute of Technology
Prakhar Gupta, Carnegie Mellon University
Yilun Zhao, Yale University
Yichen Jiang, University of North Carolina at Chapel Hill
Minghao Guo, Zhejiang University
Yiming Zong, Zhejiang University
Yanjun Shao, Fudan University
Ziming Li, Hong Kong University of Science and Technology
Yuliang Liu, Nanjing University
Behnam Hedayatnia, Amazon
Xiao Zhou, Hong Kong University of Science and Technology
Boru Lu, University of Washington
Sha Li, University of Illinois at Urbana-Champaign
Deepanway Ghosal, Singapore University of Technology and Design