Skip to content

rrajasek95/DSTC9-Dialog-Evaluation-Challenge

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

404 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Cruz Control

GPT2 Models

We currently have the following training scripts for the models:

  • GPT2 Baseline Text + Fact
  • Knowledge Dependent Policy Driven Neural Response Generator using Mezza Tags

Contact

For any clarification related to the above code, please reach out to Rishi Rajasekaran (rrajasek@ucsc.edu)

DSTC9 Baseline Code (untested)

Response Generation

Scripts to train Seq2Seq and Transformer models on the Amazon Topical-Chat Corpus. This code serves as the baseline for DSTC9 Track 3.

To train: python3 train.py --use_knowledge --transformer --save_path transformer/

To test: python3 test.py --use_knowledge --transformer --save_path transformer/

To serve interactive model with TF-IDF based fact selection: python3 dynamic.py --use_knowledge --transformer --save_path transformer/

Data

The pre-processed data can be found in data.zip. If you would like to use a different pre-processing strategy, please download the original data from here.

The dataset preparation code is split between the utils.py file and the tc_dataset.py. The data loading and tokenization is done in utils.py while the data preparation to feed into the model is done in tc_dataset.py.

Contact

If you experience any issues with this code, please contact me at mehrishikib@gmail.com

Setup

  • spacy
  • python -m spacy download en_core_web_lg
  • nltk.download('punkt')

About

Project repo for the DSTC9 dialog evaluation challenge

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages