All Projects → mgalley → Dstc7 End To End Conversation Modeling

mgalley / Dstc7 End To End Conversation Modeling

Grounded conversational dataset for end-to-end conversational AI (official DSTC7 data)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Dstc7 End To End Conversation Modeling

Dstc8 Schema Guided Dialogue
The Schema-Guided Dialogue Dataset
Stars: ✭ 277 (+96.45%)
Mutual labels:  dialogue, dataset, dialogue-systems
Unit Dmkit
Stars: ✭ 279 (+97.87%)
Mutual labels:  dialogue, dialogue-systems
Dialog Generation Paper
A list of recent papers regarding dialogue generation
Stars: ✭ 265 (+87.94%)
Mutual labels:  dialogue, dialogue-systems
Multiwoz
Source code for end-to-end dialogue model from the MultiWOZ paper (Budzianowski et al. 2018, EMNLP)
Stars: ✭ 384 (+172.34%)
Mutual labels:  dialogue, dialogue-systems
TalkerMakerDeluxe
A FOSS Branching Game Dialogue Editor
Stars: ✭ 90 (-36.17%)
Mutual labels:  dialogue, dialogue-systems
dialogue-datasets
collect the open dialog corpus and some useful data processing utils.
Stars: ✭ 24 (-82.98%)
Mutual labels:  dialogue, dialogue-systems
Meld
MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversation
Stars: ✭ 373 (+164.54%)
Mutual labels:  dialogue, dialogue-systems
Dialogue
Node based dialogue system
Stars: ✭ 207 (+46.81%)
Mutual labels:  dialogue, dialogue-systems
Letsgodataset
This repository makes the integral Let's Go dataset publicly available.
Stars: ✭ 41 (-70.92%)
Mutual labels:  dataset, dialogue-systems
Dream
DREAM: A Challenge Dataset and Models for Dialogue-Based Reading Comprehension
Stars: ✭ 60 (-57.45%)
Mutual labels:  dialogue, dataset
Dialogue Understanding
This repository contains PyTorch implementation for the baseline models from the paper Utterance-level Dialogue Understanding: An Empirical Study
Stars: ✭ 77 (-45.39%)
Mutual labels:  dialogue, dialogue-systems
DlgSystem
Dialogue Plugin System for Unreal Engine | 🪞 Mirror of https://bit.ly/DlgSource
Stars: ✭ 136 (-3.55%)
Mutual labels:  dialogue, dialogue-systems
ADEM
TOWARDS AN AUTOMATIC TURING TEST: LEARNING TO EVALUATE DIALOGUE RESPONSES
Stars: ✭ 25 (-82.27%)
Mutual labels:  dialogue, dialogue-systems
Whiskers
A Visual Dialogue Editor made using the Godot Engine
Stars: ✭ 229 (+62.41%)
Mutual labels:  dialogue, dialogue-systems
Crd3
The repo containing the Critical Role Dungeons and Dragons Dataset.
Stars: ✭ 83 (-41.13%)
Mutual labels:  dataset, dialogue-systems
Nndial
NNDial is an open source toolkit for building end-to-end trainable task-oriented dialogue models. It is released by Tsung-Hsien (Shawn) Wen from Cambridge Dialogue Systems Group under Apache License 2.0.
Stars: ✭ 332 (+135.46%)
Mutual labels:  dialogue, dialogue-systems
Anyq
FAQ-based Question Answering System
Stars: ✭ 2,336 (+1556.74%)
Mutual labels:  dialogue, dialogue-systems
Convlab 2
ConvLab-2: An Open-Source Toolkit for Building, Evaluating, and Diagnosing Dialogue Systems
Stars: ✭ 196 (+39.01%)
Mutual labels:  dialogue, dialogue-systems
Rnnlg
RNNLG is an open source benchmark toolkit for Natural Language Generation (NLG) in spoken dialogue system application domains. It is released by Tsung-Hsien (Shawn) Wen from Cambridge Dialogue Systems Group under Apache License 2.0.
Stars: ✭ 487 (+245.39%)
Mutual labels:  dialogue, dialogue-systems
Atis dataset
The ATIS (Airline Travel Information System) Dataset
Stars: ✭ 81 (-42.55%)
Mutual labels:  dataset, dialogue-systems

DSTC7: End-to-End Conversation Modeling

DSTC7 has ended on January 27, 2019. This github project is still available 'as is', but we unfortunately no longer have time to maintain the code or to provide assistance with this project.

News

  • 10/29/2018: Spreadsheet containing indivdual judgments used for human evaluation.
  • 10/23/2018 and 10/15/2018: Automatic and human evaluation results posted. The code to reproduce the automatic evaluation and get the same scores can be found here.
  • 10/8/2018: Participants submitted system outputs.
  • 9/10/2018-10/8/2018: Evaluation phase, instructions here.
  • 7/11/2018: An FAQ section has been added to the data extraction page.
  • 7/1/2018: Official training data is up.
  • 6/18/2018: Trial data is up.
  • 6/1/2018: Task description is up.
  • 6/1/2018: Registration for DSTC7 is now open.

Registration

Please register [here] Registration has now closed.

Task

This DSTC7 track presents an end-to-end conversational modeling task, in which the goal is to generate conversational responses that go beyond trivial chitchat by injecting informative responses that are grounded in external knowledge. This task is distinct from what is commonly thought of as goal-oriented, task-oriented, or task-completion dialog in that there is no specific or predefined goal (e.g., booking a flight, or reserving a table at a restaurant). Instead, it targets human-like interactions where the underlying goal is often ill-defined or not known in advance, of the kind seen, for example, in work and other productive environments (e.g.,brainstorming meetings) where people share information.

Please check this description for more details about the task, which follows our previous work "A Knowledge-Grounded Neural Conversation Model" and our original task proposal.

Data

We extend the knowledge-grounded setting, with each system input consisting of two parts:

  • Conversational data from Reddit.
  • Contextually-relevant “facts”, taken from the website that started the (Reddit) conversation.

Please check the data extraction for the input data pipeline. Note: We are providing scripts to extract the data from a Reddit dump, as we are unable to release the data directly ourselves.

Evaluation

As described in the task description (Section 4), We will evaluate response quality using both automatic and human evaluations on two criteria.

  • Appropriateness;
  • Informativeness.

We will use automatic evaluation metrics such as BLEU and METEOR to have preliminary score for each submission prior to the human evaluation. Participants can also use these metrics for their own evaluations during the development phase. We will allow participants to submit multiple system outputs with one system marked as “primary” for human evaluation. We will provide a BLEU scoring script to help participants decide which system they want to select as primary.

We will use crowdsourcing for human evaluation. For each response, we ask humans if it is an (1) appropriate and (2) informative response, on a scale from 1 to 5. The system with best average Appropriateness and Informativeness will be determined the winner.

Baseline

A standard seq2seq baseline model will be provided soon.

Timeline

Phase Dates
1. Development Phase June 1 – September 9
      1.1 Code (data extraction code, seq2seq baseline) June 1
      1.2 "Trial" data made available June 18
      1.3 Official training data made available July 1
2. Evaluation Phase September 10 – October 8
      2.1 Test data made available September 10
      2.2 Participants submit their system outputs October 8
3. Results are released October
      3.1 Automatic scores (BLEU, etc.) October 16
      3.2 Human evaluation October 23

Organizers

Reference

If you submit any system to DSTC7-Task2 or publish any other work making use of the resources provided on this project, we ask you to cite the following task description paper:

Michel Galley, Chris Brockett, Xiang Gao, Bill Dolan, Jianfeng Gao. End-to-End conversation Modeling: DSTC7 Task 2 Description. In DSTC7 workshop (forthcoming).

Contact Information

  • For questions specific to Task 2, you can contact us at [email protected]. (No longer maintained.)
  • You can get the latest updates and participate in discussions on DSTC mailing list.
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].