|
171 | 171 | " untar=True, cache_dir='.',\n",
|
172 | 172 | " cache_subdir='')\n",
|
173 | 173 | "\n",
|
174 |
| - "dataset_dir = os.path.join(os.path.dirname(dataset), 'aclImdb')" |
| 174 | + "dataset_dir = os.path.join(os.path.dirname(dataset), 'aclImdb_v1')" |
175 | 175 | ]
|
176 | 176 | },
|
177 | 177 | {
|
|
193 | 193 | },
|
194 | 194 | "outputs": [],
|
195 | 195 | "source": [
|
196 |
| - "train_dir = os.path.join(dataset_dir, 'train')\n", |
| 196 | + "train_dir = os.path.join(dataset_dir, 'aclImdb', 'train')\n", |
197 | 197 | "os.listdir(train_dir)"
|
198 | 198 | ]
|
199 | 199 | },
|
|
214 | 214 | },
|
215 | 215 | "outputs": [],
|
216 | 216 | "source": [
|
217 |
| - "sample_file = os.path.join(train_dir, 'pos/1181_9.txt')\n", |
| 217 | + "sample_file = os.path.join(train_dir, 'pos', '1181_9.txt')\n", |
218 | 218 | "with open(sample_file) as f:\n",
|
219 | 219 | " print(f.read())"
|
220 | 220 | ]
|
|
286 | 286 | "seed = 42\n",
|
287 | 287 | "\n",
|
288 | 288 | "raw_train_ds = tf.keras.utils.text_dataset_from_directory(\n",
|
289 |
| - " 'aclImdb/train',\n", |
| 289 | + " train_dir,\n", |
290 | 290 | " batch_size=batch_size,\n",
|
291 | 291 | " validation_split=0.2,\n",
|
292 | 292 | " subset='training',\n",
|
|
366 | 366 | "outputs": [],
|
367 | 367 | "source": [
|
368 | 368 | "raw_val_ds = tf.keras.utils.text_dataset_from_directory(\n",
|
369 |
| - " 'aclImdb/train',\n", |
| 369 | + " train_dir,\n", |
370 | 370 | " batch_size=batch_size,\n",
|
371 | 371 | " validation_split=0.2,\n",
|
372 | 372 | " subset='validation',\n",
|
|
0 commit comments