@@ -46,7 +46,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
4646Uses the normalized float 4 (nf4) data type. This is recommended over "fp4" based on the paper's experimental results and theoretical analysis.
4747
4848``` bash
49- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
49+ pip install bitsandbytes
5050
5151litgpt generate base --quantize bnb.nf4 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
5252...
@@ -62,7 +62,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
6262In average, this amounts to about 0.37 bits per parameter (approximately 3 GB for a 65B model).
6363
6464``` bash
65- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
65+ pip install bitsandbytes
6666
6767litgpt generate base --quantize bnb.nf4-dq --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
6868...
@@ -77,7 +77,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
7777Uses pure FP4 quantization.
7878
7979``` bash
80- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
80+ pip install bitsandbytes
8181
8282litgpt generate base --quantize bnb.fp4 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
8383...
@@ -93,7 +93,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
9393In average, this amounts to about 0.37 bits per parameter (approximately 3 GB for a 65B model).
9494
9595``` bash
96- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
96+ pip install bitsandbytes
9797
9898litgpt generate base --quantize bnb.fp4-dq --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
9999...
@@ -106,7 +106,7 @@ Memory used: 5.38 GB
106106Enabled with [ bitsandbytes] ( https://github.com/TimDettmers/bitsandbytes ) . Check out the [ paper] ( https://arxiv.org/abs/2110.02861 ) to learn more about how it works.
107107
108108``` bash
109- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
109+ pip install bitsandbytes
110110
111111litgpt generate base --quantize bnb.int8 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision 16-true --max_new_tokens 256
112112...
0 commit comments