Tilburg center for Cognition and Communication (TiCC), Tilburg University
Department of Linguistics, Federal University of Minas Gerais
Data-to-text Natural Language Generation: Traditional, Novel and Future Methods
Data-to-text Natural Language Generation (NLG) is a consolidated field of research which normally combines Computational Linguistics, Software Engineering and Artificial Intelligence methods to generate natural language from non-linguistic representations. Traditionally, most data-to-text applications have been designed using a modular pipeline architecture, in which the non-linguistic input data is converted into natural language through several intermediate transformations. In contrast, influenced by the phenomenon of deep learning, recent neural models for data-to-text generation have been proposed as end-to-end approaches, where the non-linguistic input is rendered in natural language with much less explicit intermediate representations in-between. Theoretically, we know that pipeline approaches are more transparent and their modules can be reused across applications, whereas neural end-to-end approaches may demand less manual labor and has registered state-of-the-art results in other text generation tasks, such as Machine Translation. Although we know these pros and cons in theory, the question about which kind of model empirically generates the most fluent and semantic texts from non-linguistic representations still remains unanswered. This lack of an empirical comparison is partially caused by the fact that traditional benchmarks for the task only consist of raw non-linguistic representations in parallel with their textual realizations, benefiting the evaluation of end-to-end approaches but not of pipeline ones, since explicit intermediate representations are missing for the study of particular modules of the latter architecture. In this presentation, I will introduce an annotation framework to enrich popular benchmarks with explicit intermediate representations, which will help the development and evaluation of particular pipeline modules as Discourse Ordering, Aggregation, Lexicalization, Referring Expressions Generation and Textual Realization. Based on a version of a popular data-to-text benchmark enriched with our framework, I will also present the results of a comparison between pipeline and end-to-end approaches. Finally, based on the findings of this project, I will project future challenges in the research of data-to-text NLG in terms of data, applications and evaluation.
Convince Me If You Can: Argument Generation with Content Planning and Style Specification
Understanding, evaluating, and generating arguments are crucial elements of the decision-making and reasoning process. A multitude of arguments and counter-arguments are constructed on a daily basis to persuade and inform us on a wide range of issues. However, constructing persuasive arguments is a challenging task for both human and computers, as it requires credible evidence, rigorous logical reasoning, and sometimes emotional appeals.
In this talk, I will introduce our neural network-based argument generation model. It consists of a powerful retrieval system and a novel two-step generation model, where a text planning decoder first decides on the main talking points and a proper language style for each sentence, then a content realization component constructs an informative and fluent paragraph-level argument. We believe that the proposed argument generation framework will enable many compelling applications, including providing unbiased perspectives on complex issues, debate coaching, and essay writing tutoring. Our framework is also generic and has been applied to other text generation problems, such as Wikipedia article paragraph generation and scientific paper abstract writing.
CCIS, Northeastern University
Facebook AI
Hierarchical Structure in Story Generation
We explore the task of story generation: creative systems that can build coherent and fluent passages of text about a topic. Using a dataset of 300k human-written stories paired with writing prompts, we investigate hierarchical story generation. Our models first generate a premise and then transform it into a passage of text. We develop models that improve the relevance of the story to the premise using a novel form of model fusion and present improvements to self-attention that better capture long-range context. Then, we build upon this work by proposing a coarse to fine mechanism for story generation, decomposing the task into several steps. We first explicitly generate logical verb sequences to model action in stories, then form these into sentences, and finally fill-in character names. We show that such decompositions improve the consistency and diversity of generated stories.
Topical Chat: On the Structure of Knowledge Grounded Conversations
Conversational agents like Amazon Alexa, Google Assistant and Apple Siri have been exploding in popularity over the past few years. However, much work remains in the area of social conversation over a broad range of domains and topics. To advance the state of the art in open domain dialog, Amazon launched the Alexa Prize, a 2.5-million-dollar university competition where selected university teams were challenged to build conversational agents, known as “socialbots”, to converse coherently and engagingly with humans on popular topics such as Sports, Politics, Entertainment, Fashion and Technology for 20 minutes. The Alexa Prize offers a unique opportunity to perform research and interact with real user conversational data at scale. Over the past two years, we have learned that there are certain areas that these bots could improve on such as topical depth, breadth and smooth topical transitions in order to have deep and engaging conversations. Given this information, we formed a conversational dataset where we can study how to create engaging conversations. We introduce Topical-Chat: a knowledge-grounded human-human conversation dataset, where the underlying knowledge spans 8 broad topics. Our dataset enables models to leverage world knowledge while conversing with humans leading to more coherent and interesting conversations. We will present some modeling work using generative encoder-decoder conversational models trained on Topical-Chat and perform automated and human evaluation for benchmarking. Additionally we will present an analysis of Topical-Chat based on the knowledge content selected and presented to humans as background knowledge.
Amazon Alexa AI