You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CHANGELOG.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -49,7 +49,7 @@ Features:
49
49
Bug fixes:
50
50
- Fixed a bug where weight decay was incorrectly applied to 32-bit Adam. #13
51
51
- Fixed an unsafe use of eval. #8
52
-
- Fixed a bug where the StableEmbedding layer 32-bit optimizer override would not work without registering the whole model first (`bnb.optim.GlobalOptimManager.get_instance().register_parameters(model.parameters())`). #13#15
52
+
- Fixed a bug where the StableEmbedding layer 32-bit optimizer override would not work without registering the whole model first (`bnb.optim.GlobalOptimManager.get_instance().register_parameters(model.parameters())`). #13#15
53
53
54
54
Docs:
55
55
- Added instructions how to solve "\_\_fatbinwrap_" errors.
Copy file name to clipboardExpand all lines: README.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# bitsandbytes
2
2
3
-
The bitsandbytes is a lightweight wrapper around CUDA custom functions, in particular 8-bit optimizers, matrix multiplication (LLM.int8()), and quantization functions.
3
+
The bitsandbytes is a lightweight wrapper around CUDA custom functions, in particular 8-bit optimizers, matrix multiplication (LLM.int8()), and quantization functions.
4
4
5
5
6
6
@@ -48,7 +48,7 @@ out = linear(x.to(torch.float16))
48
48
49
49
Requirements: anaconda, cudatoolkit, pytorch
50
50
51
-
Hardware requirements:
51
+
Hardware requirements:
52
52
- LLM.int8(): NVIDIA Turing (RTX 20xx; T4) or Ampere GPU (RTX 30xx; A4-A100); (a GPU from 2018 or older).
53
53
- 8-bit optimizers and quantization: NVIDIA Maxwell GPU or newer (>=GTX 9XX).
54
54
@@ -87,7 +87,7 @@ Note that by default all parameter tensors with less than 4096 elements are kept
87
87
```
88
88
# parameter tensors with less than 16384 values are optimized in 32-bit
89
89
# it is recommended to use multiplies of 4096
90
-
adam = bnb.optim.Adam8bit(model.parameters(), min_8bit_size=16384)
90
+
adam = bnb.optim.Adam8bit(model.parameters(), min_8bit_size=16384)
91
91
```
92
92
93
93
### Change Bits and other Hyperparameters for Individual Parameters
Copy file name to clipboardExpand all lines: compile_from_source.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ Basic steps.
4
4
1.`make [target]` where `[target]` is among `cuda92, cuda10x, cuda110, cuda11x, cpuonly`
5
5
2.`CUDA_VERSION=XXX python setup.py install`
6
6
7
-
To run these steps you will need to have the nvcc compiler installed that comes with a CUDA installation. If you use anaconda (recommended) then you can figure out which version of CUDA you are using with PyTorch via the command `conda list | grep cudatoolkit`. Then you can install the nvcc compiler by downloading and installing the same CUDA version from the [CUDA toolkit archive](https://developer.nvidia.com/cuda-toolkit-archive).
7
+
To run these steps you will need to have the nvcc compiler installed that comes with a CUDA installation. If you use anaconda (recommended) then you can figure out which version of CUDA you are using with PyTorch via the command `conda list | grep cudatoolkit`. Then you can install the nvcc compiler by downloading and installing the same CUDA version from the [CUDA toolkit archive](https://developer.nvidia.com/cuda-toolkit-archive).
8
8
9
9
For your convenience, there is an installation script in the root directory that installs CUDA 11.1 locally and configures it automatically. After installing you should add the `bin` sub-directory to the `$PATH` variable to make the compiler visible to your system. To do this you can add this to your `.bashrc` by executing these commands:
By default, the Makefile will look at your `CUDA_HOME` environmental variable to find your CUDA version for compiling the library. If this path is not set it is inferred from the path of your `nvcc` compiler.
16
+
By default, the Makefile will look at your `CUDA_HOME` environmental variable to find your CUDA version for compiling the library. If this path is not set it is inferred from the path of your `nvcc` compiler.
17
17
18
18
Either `nvcc` needs to be in path for the `CUDA_HOME` variable needs to be set to the CUDA directory root (e.g. `/usr/local/cuda`) in order for compilation to succeed
0 commit comments