Skip to content

Commit dc05339

Browse files
committed
add dolly as input model
1 parent fc22734 commit dc05339

File tree

2 files changed

+4
-0
lines changed

2 files changed

+4
-0
lines changed

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -90,6 +90,7 @@ Otherwise, follow the steps above. The 12B param model may not function well in
9090
- Add the `dolly` repo to Databricks (under Repos click Add Repo, enter `https://github.com/databrickslabs/dolly.git`, then click Create Repo).
9191
- Start a `12.2 LTS ML (includes Apache Spark 3.3.2, GPU, Scala 2.12)` single-node cluster with node type having 8 A100 GPUs (e.g. `Standard_ND96asr_v4` or `p4d.24xlarge`). Note that these instance types may not be available in all regions, or may be difficult to provision. In Databricks, note that you must select the GPU runtime first, and unselect "Use Photon", for these instance types to appear (where supported).
9292
- Open the `train_dolly` notebook in the Repo (which is the `train_dolly.py` file in the Github `dolly` repo), attach to your GPU cluster, and run all cells. When training finishes, the notebook will save the model under `/dbfs/dolly_training`.
93+
- You can also directly use Dolly as the input model and further train your own model using your file tuning records uploaded to the data directory. Simply upload the files and set `local_files` as the training_dataset in the `train_dolly` notebook.
9394

9495
### Training on Other Instances
9596

training/consts.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,9 @@
44
"EleutherAI/pythia-6.9b",
55
"EleutherAI/pythia-12b",
66
"EleutherAI/gpt-j-6B",
7+
"databricks/dolly-v2-3b"
8+
"databricks/dolly-v2-7b",
9+
"databricks/dolly-v2-12b",
710
]
811
DEFAULT_TRAINING_DATASET = "databricks/databricks-dolly-15k"
912
INTRO_BLURB = (

0 commit comments

Comments
 (0)