Dialogue summarization for automatic report generation

Dialogue summarization for automatic report generation

Research questions: 

– How to control output of a pre-trained language model and include necessary input in the correct order? E.g., experiments with different prompts for generation vs. post-editing. 

– What is the best way to encode the previous context? Can we just concatenate the strings or is a more structured representation needed? e.g., one could experiment with using AMRs for previous turns and provide them as additional inputs. 

Option #1: we can use our in-house EveEnti annotations as a formal representation of the mission content and implement a model that will use such representations as input and generate fluent text describing the mission. 

Option #2: use AMRs as formal input representations. 

In both cases we will need an annotator to produce reference descriptions for each of the dialogue threads. We are working together with firefighters and may be able to get their support to create the reference descriptions. 

Paper suggestions: 

  1. Improving abstractive dialogue summarization with graph structures and topic words, Zhao et al., 2020 Link: https://www.aclweb.org/anthology/2020.coling-main.39/ 
  1. Multi-view sequence-to-sequence models with conversational structure for abstractive dialogue summarization, Chen and Yang, 2021. Link: https://arxiv.org/pdf/2010.01672 
  1. An Exploratory Study on Long Dialogue Summarization: What Works and What’s Next, Zhang et al., 2021. Link: https://arxiv.org/abs/2109.04609 
  1. A survey on dialogue summarization: Recent advances and new frontiers, Feng et al., 2021. Link: https://arxiv.org/abs/2107.03175v1