I've come up with a step-by-step guide to using the notebook in this repo. You can find it here (feel free to dismiss the subscription prompt).
Repo description 99% generated by GPT-4, and not all proofread, tell me if there are errors or GPT-isms.
I've licensed this code under MIT. I have no clue what kind of license fits the dataset itself, and I'm not going to give any legal advice on that front either.
This repository contains the code used to generate the Augmental dataset, as well as the model training code to finetune models on it. The dataset stands out due to its innovative approach of utilizing AI to enhance human-written scripts from visual novels, bridging the gap between purely synthetic data and manual data curation.
- Introduction
- Prerequisites
- Data Generation
- Model Training
- Acknowledgements
- References and Useful Links
The Augmental dataset is a novel multiturn dataset containing 7.86k replies spread across about 480 different conversations among 7 distinct characters. This dataset was crafted by refining and enhancing the script of the visual novel Steins;Gate using GPT-4. The dataset prioritizes quality, longer responses, and retaining the human-like essence of the conversation while benefitting from the capabilities of GPT-4.
The dataset is generated from the .scx.txt files of Steins;Gate. It's essential to have a legal copy of Steins;Gate to extract the required files. Please ensure you have the rights to use the text from the visual novel for your purposes.
- sc3tools for extracting
.scx.txtfiles from Steins;Gate.
- Extract the
.scx.txtfiles from Steins;Gate using sc3tools. - Merge the extracted
.scx.txtfiles into a single text file.
- Open the
processing_refactor.ipynbnotebook. - Before running the notebook, ensure you've toggled the
dataset_has_been_manually_editedflag at the top:- Set to
Trueif working with the original dataset. If this is true, it (shouldn't) make any OAI calls and will leave any gaps in the dataset alone. - Set to
Falseif you're generating new data.
- Set to
- Run the notebook to process the raw text file. The output will be the Augmental dataset, ready for model training.
The training code for finetuning models on the Augmental dataset is contained in train.py.
Use the processing_refactor notebook how you would normally use a notebook. Cells that run OpenAI will skip generations that have already been saved to files in the ./annotated_convs, ./scenarios, and ./anchors directories. !!DO NOT DELETE FILES IN THOSE DIRECTORIES UNLESS YOU WANT PAINFUL ERRORS!!
python train.pyThis dataset is an evolution of the dataset that was used to train MythoMakise, a model that achieved notable recognition. The current model, trained on the Augmental dataset, promises even higher quality interactions and versatility. See the note on the Augmental Dataset HF page for legal considerations. TLDR: if the legal holders of the Steins;Gate IP tell me to take this down I will without a second thought.
- Reddit Post introducing the dataset.
- Model Card detailing the specifics and innovations behind the model.
- Ko-fi Support Link
- Essay on Substack discussing the rationale behind the dataset. Feedback, suggestions, and contributions are always welcome!