Skip to content

Commit c895055

Browse files
authored
Update llavaction_video_demo.ipynb
- changed header to note demo data (not model) - model from HuggingFace directly
1 parent 3ba3950 commit c895055

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

example/llavaction_video_demo.ipynb

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -141,7 +141,7 @@
141141
"id": "gTr6BFHuva0v"
142142
},
143143
"source": [
144-
"## Grab our model weights and demo data:\n"
144+
"## Grab our demo video (from EPIC-KITCHENs-100):\n"
145145
]
146146
},
147147
{
@@ -152,7 +152,7 @@
152152
},
153153
"outputs": [],
154154
"source": [
155-
"# currently, demo data and a 0.5B model is hosted here:\n",
155+
"# demo data is hosted here:\n",
156156
"!gdown --folder https://drive.google.com/drive/folders/1ql8MSWTK-2_uGH1EzPOrifauwUNg4E6i -O ./data"
157157
]
158158
},
@@ -190,7 +190,7 @@
190190
"source": [
191191
"## Setting Key parameters for LLaVAction:\n",
192192
"\n",
193-
"- Specify where to load the EPIC-KITCHENS-100 videos and the LLaVAction checkpoint for the inference. If you followed our steps above, these paths do not need changed for Google Colaboratory.\n",
193+
"- Specify where to load the EPIC-KITCHENS-100 video and grab the LLaVAction checkpoint from HuggingFace for the inference. If you followed our steps above, these paths do not need changed for Google Colaboratory.\n",
194194
"- ⚠️ You can adjust `n_frames` to a higher number for better performance (which we empirically observed), but note this uses more compute."
195195
]
196196
},
@@ -235,7 +235,7 @@
235235
"output_type": "stream",
236236
"text": [
237237
"overwrite_config {}\n",
238-
"Loaded LLaVA model: /content/data/checkpoint/dev_ov_0.5b_16f_top5_full\n"
238+
"Loaded LLaVA model: MLAdaptiveIntelligence/LLaVAction-0.5B\n"
239239
]
240240
},
241241
{
@@ -290,11 +290,11 @@
290290
],
291291
"source": [
292292
"data_root = '/content/data/EK100_512/EK100'\n",
293-
"checkpoint_path = 'MLAdaptiveIntelligence/LLaVAction-0.5B' # or MLAdaptiveIntelligence/LLaVAction-7B\n",
293+
"huggingface_checkpoint_path = 'MLAdaptiveIntelligence/LLaVAction-0.5B' # or MLAdaptiveIntelligence/LLaVAction-7B\n",
294294
"n_frames = 16\n",
295295
"\n",
296296
"inferencer = SelectiveInferencer(data_root,\n",
297-
" checkpoint_path,\n",
297+
" huggingface_checkpoint_path,\n",
298298
" include_time_instruction = False,\n",
299299
" n_frames = n_frames,\n",
300300
" use_flash_attention = True)"

0 commit comments

Comments
 (0)