You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
flop_counter: bool=False# Enable Flop counter to measure model throughput, can not be used with pytorch profiler at the same time.
54
+
flop_counter_startpoint: int=3# The step to start profiling, default is 3, which means after 3 steps of warmup stage, the profiler will start to count flops.
55
+
use_profiler: bool=False# Enable pytorch profiler, can not be used with flop counter at the same time.
56
+
profiler_dir: str="PATH/to/save/profiler/results"# will be used if using profiler
52
57
```
53
58
54
59
*[Datasets config file](../../src/llama_recipes/configs/datasets.py) provides the available options for datasets.
@@ -84,7 +89,13 @@ You can enable [W&B](https://wandb.ai/) experiment tracking by using `use_wandb`
To help with benchmarking effort, we are adding the support for counting the flops during the fine-tuning process. You can achieve this by setting `--flop_counter` when launching your single/multi GPU fine-tuning. Use `--flop_counter_startpoint` to choose which step to count the flops. It is recommended to allow a warmup stage before using the flop counter.
100
+
101
+
Similarly, you can set `--use_profiler` flag and pass a profiling output path using `--profiler_dir` to capture the profile traces of your model using [PyTorch profiler](https://pytorch.org/tutorials/intermediate/tensorboard_profiler_tutorial.html). This would be helpful for debugging purposes. However, the `--flop_counter` and `--use_profiler` can not be used in the same time to ensure the measurement accuarcy.
0 commit comments