Conference

Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts - ACL Anthology

· · 6 min read · 7 views

Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts Qi Zhang , Hassan Sajjad (Editors) Anthology ID: 2023.emnlp-tutorial Month: December Year: 2023 Address: Singapore Venue: EMNLP SIG: Publisher: Association for Computational Linguistics URL: https://aclanthology.org/2023.emnlp-tutorial/ DOI: Bib Export formats: BibTeX MODS XML EndNote PDF: https://aclanthology.org/2023.emnlp-tutorial.pdf PDF (full) Bib TeX Search Show all abstracts Hide all abstracts pdf bib Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts Qi Zhang | Hassan Sajjad pdf bib abs NLP + V is: NLP Meets Visualization Shafiq Joty | Enamul Hoque | Jesse Vig Natural language and visualization (Vis) are two powerful modalities of human communication. The goal of this tutorial is to push forward the agenda of tightly integrating these two modalities. To this end, the tutorial will introduce NLP+Vis with a focus on two main threads of work: (i) NLP for Vis: How to develop and adapt state-of-the-art NLP models for solving various visualization tasks? and (ii) Vis for NLP: How to leverage visualization techniques to interpret and explain complex NLP models effectively? The tutorial will first motivate why NLP+Vis is an important area of research and provide an overview of research topics on combining NLP and Vis techniques. Then an overview of state-of-the-art deep learning models for NLP will be covered. Next, we will provide an overview of applying visualization techniques to help make NLP models more interpretable and explainable. In the final part, we will focus on various application tasks at the intersection of NLP and Vis. We will conclude with an interactive discussion of future challenges for NLP+Vis applications. The audience will include researchers interested in applying NLP for visualizations as well as others who focus more generally at the intersection of machine learning and visualization. pdf bib abs Security Challenges in Natural Language Processing Models Qiongkai Xu | Xuanli He Large-scale natural language processing models have been developed and integrated into numerous applications, given the advantage of their remarkable performance. Nonetheless, the security concerns associated with these models prevent the widespread adoption of these black-box machine learning models. In this tutorial, we will dive into three emerging security issues in NLP research, i.e., backdoor attacks, private data leakage, and imitation attacks. These threats will be introduced in accordance with their threatening usage scenarios, attack methodologies, and defense technologies. pdf bib abs Designing, Evaluating, and Learning from Humans Interacting with NLP Models Tongshuang Wu | Diyi Yang | Sebastin Santy The rapid advancement of natural language processing (NLP) research has led to various applications spanning a wide range of domains that require models to interact with humans – e.g., chatbots responding to human inquiries, machine translation systems assisting human translators, designers prompting Large Language Models for co-creation or prototyping AI-infused applications, etc. In these cases, humans interaction is key to the success of NLP applications; any potential misconceptions or differences might lead to error cascades at the subsequent stages. Such interaction involves a lot of design choices around models, e.g. the sensitivity of interfaces, the impact of design choice and evaluation questions, etc. This tutorial aims to provide a systematic and up-to-date overview of key considerations and effective approaches for studying human-NLP model interactions. Our tutorial will focus specifically on the scenario where end users – lay people and domain experts who have access to NLP models but are less familiar with NLP techniques – use or collaborate with deployed models. Throughout the tutorial, we will use five case studies (on classifier-assisted decision making, machine-aided translation, dialog systems, and prompting) to cover three major themes: (1) how to conduct human-in-the-loop usability evaluations to ensure that models are capable of interacting with humans; (2) how to design user interfaces (UIs) and interaction mechanisms that provide end users with easy access to NLP models; (3) how to learn and improve NLP models through the human interactions. We will use best practices from HCI to ground our discussion, and will highlight current challenges and future directions. pdf bib abs LLM -driven Instruction Following: Progresses and Concerns Wenpeng Yin | Qinyuan Ye | Pengfei Liu | Xiang Ren | Hinrich Schütze The progress of natural language processing (NLP) is primarily driven by machine learning that optimizes a system on a large-scale set of task-specific labeled examples. This learning paradigm limits the ability of machines to have the same capabilities as humans in handling new tasks since humans can often solve unseen tasks with a couple of examples accompanied by task instructions. In addition, we may not have a chance to prepare task-specific examples of large-volume for new tasks because we cannot foresee what task needs to be addressed next and how complex to annotate for it. Therefore, task instructions act as a novel and promising resource for supervision. This tutorial targets researchers and practitioners who are interested in AI and ML technologies for NLP generalization in a low-shot scenario. In particular, we will present a diverse thread of instruction-driven NLP studies that try to answer the following questions: (i) What is task instruction? (ii) How is the process of creating datasets and evaluating systems conducted? (iii) How to encode task instructions? (iv) When and why do some instructions work better? (v) What concerns remain in LLM-driven instruction following? We will discuss several lines of frontier research that tackle those challenges and will conclude the tutorial by outlining directions for further investigation. pdf bib abs Mitigating Societal Harms in Large Language Models Sachin Kumar | Vidhisha Balachandran | Lucille Njoo | Antonios Anastasopoulos | Yulia Tsvetkov Numerous recent studies have highlighted societal harms that can be caused by language technologies deployed in the wild. While several surveys, tutorials, and workshops have discussed the risks of harms in specific contexts – e.g., detecting and mitigating gender bias in NLP models – no prior work has developed a unified typology of technical approaches for mitigating harms of language generation models. Our tutorial is based on a survey we recently wrote that proposes such a typology. We will provide an overview of potential social issues in language generation, including toxicity, social biases, misinformation, factual inconsistency, and privacy violations. Our primary focus will be on how to systematically identify risks, and how eliminate them at various stages of model development, from data collection, to model development, to inference/language generation. Through this tutorial, we aim to equip NLP researchers and engineers with a suite of practical tools for mitigating safety risks from pretrained language generation models. pdf bib abs Creative Natural Language Generation Tuhin Chakrabarty | Vishakh Padmakumar | He He | Nanyun Peng Large language models such as GPT-3, GPT4, Claude etc., have advanced the state of the art in several natural language generation tasks such as text summarization and machine translation. However when it comes to open-ended tasks with a focus on creativity such as generating stories, poetry, or various forms of figurative language, these state-of-the-art language models are often found to be inadequate. This tutorial aims to bring awareness of the important and emerging research area of open-domain creative generation, with a focus on language generation while also touching on multi-modal generation (e.g., image captioning, visual metaphors). It targets natural language processing (NLP) and artificial intelligence (AI) researchers as well as creative writing practitioners who are interested in building systems that are capable of emulating as well as augmenting human creativity. In particular, we will review recent studies on creative language generation both at the sentence level as well as longer forms of text. We will provide the audiences with a holistic view of 1) the importance and challenges of building creative language generation systems; 2) how we incorporate content planning, domain knowledge and creativity specific heuristics for different forms of creative language generation such as story, poetry, humor, metaphors etc 3) how can we build better evaluation methods for creative text generation? In particular, how could the recent advancement of AI shape the future workforce for creativity? We will conclude the tutorial by outlining future research directions in this area.

Executive Summary

The Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP) Tutorial Abstracts present a comprehensive overview of cutting-edge research and tutorials in the field of NLP. The tutorials cover a range of topics, including the intersection of NLP and visualization, security challenges in NLP models, and human interaction with NLP systems. These abstracts highlight the advancements and emerging issues in NLP, providing valuable insights for researchers and practitioners in the field.

Key Points

  • Integration of NLP and visualization techniques
  • Security challenges in NLP models, including backdoor attacks, private data leakage, and imitation attacks
  • Human interaction with NLP models and the importance of designing and evaluating these interactions

Merits

Comprehensive Coverage

The proceedings cover a wide range of topics, providing a holistic view of current research and challenges in NLP.

Interdisciplinary Approach

The tutorials emphasize the intersection of NLP with other fields such as visualization and human-computer interaction, fostering interdisciplinary research.

Practical Applications

The tutorials focus on practical applications and real-world challenges, making the research relevant to industry practitioners.

Demerits

Limited Depth

The abstracts provide a high-level overview but lack detailed discussions on specific methodologies and results.

Focus on Emerging Issues

While the tutorials address emerging issues, they may not cover foundational aspects of NLP comprehensively.

Expert Commentary

The Proceedings of the 2023 EMNLP Tutorial Abstracts offer a valuable snapshot of the current state and future directions of NLP research. The tutorials highlight the growing importance of interdisciplinary approaches, particularly the integration of NLP with visualization techniques and human-computer interaction. The focus on security challenges is timely, given the increasing adoption of NLP models in various applications. However, the abstracts could benefit from more detailed discussions on specific methodologies and results to provide deeper insights. Overall, these proceedings are a valuable resource for researchers and practitioners seeking to stay at the forefront of NLP advancements.

Recommendations

  • Encourage more detailed discussions on specific methodologies and results in future tutorials to provide deeper insights.
  • Foster interdisciplinary collaboration by organizing more joint workshops and tutorials that bridge NLP with other fields such as visualization and human-computer interaction.

Sources

Related Articles