|
22 | 22 | from omegaconf import DictConfig, OmegaConf |
23 | 23 | from torch import nn |
24 | 24 | from torchdata.stateful_dataloader import StatefulDataLoader |
25 | | -from torchtitan.components.checkpoint import ModelWrapper |
26 | 25 | from torchtitan.components.loss import LossFunction |
27 | 26 | from torchtitan.components.lr_scheduler import LRSchedulersContainer |
28 | 27 | from torchtitan.components.optimizer import OptimizersContainer |
@@ -76,22 +75,20 @@ def setup(self): |
76 | 75 | # self.train_config.packing_config, |
77 | 76 | # ) |
78 | 77 |
|
79 | | - # TODO: confirm that this is working properly |
80 | | - # Should also use load, not dcp_load |
81 | 78 | self.checkpointer.load(step=self.current_step) |
82 | 79 | # self.profiler = self.setup_profiler(self.train_config.profiler_config) |
83 | 80 | # self.logger = self.setup_logger(self.train_config.logger_config) |
84 | 81 |
|
85 | 82 | def setup_data(self): |
86 | 83 | tokenizer = HuggingFaceModelTokenizer( |
87 | 84 | tokenizer_json_path=os.path.join( |
88 | | - self.job_config.model.tokenizer_path, "tokenizer.json" |
| 85 | + self.job_config.model.hf_assets_path, "tokenizer.json" |
89 | 86 | ), |
90 | 87 | tokenizer_config_json_path=os.path.join( |
91 | | - self.job_config.model.tokenizer_path, "tokenizer_config.json" |
| 88 | + self.job_config.model.hf_assets_path, "tokenizer_config.json" |
92 | 89 | ), |
93 | 90 | generation_config_path=os.path.join( |
94 | | - self.job_config.model.tokenizer_path, "generation_config.json" |
| 91 | + self.job_config.model.hf_assets_path, "generation_config.json" |
95 | 92 | ), |
96 | 93 | ) |
97 | 94 |
|
|
0 commit comments