Update dependency bitsandbytes to ^0.46.0 #133
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
^0.42.0
->^0.46.0
Release Notes
bitsandbytes-foundation/bitsandbytes (bitsandbytes)
v0.46.1
Compare Source
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.46.0...0.46.1
v0.46.0
: : torch.compile() support; custom ops refactor; Linux aarch64 wheelsCompare Source
Highlights
torch.compile
without graph breaks for LLM.int8().torch.compile
without graph breaks for 4bit.fullgraph=False
.fullgraph=True
.torch.library
and custom ops APIs. This helps enable ourtorch.compile
and additional hardware compatibility efforts.bitsandbytes
.Compatability Changes
manylinux_2_24
(previouslymanylinux_2_34
).What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.4...0.46.0
v0.45.5
Compare Source
This is a minor release that affects CPU-only usage of bitsandbytes. The CPU build of the library was inadvertently omitted from the v0.45.4 wheels.
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.4...0.45.5
v0.45.4
Compare Source
This is a minor release that affects CPU-only usage of bitsandbytes. There is one bugfix and improved system compatibility on Linux.
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.3...0.45.4
v0.45.3
Compare Source
Overview
This is a small patch release containing a few bug fixes.
Additionally, this release contains a CUDA 12.8 build which adds the sm100 and sm120 targets for NVIDIA Blackwell GPUs.
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.2...0.45.3
v0.45.2
Compare Source
This patch release fixes a compatibility issue with Triton 3.2 in PyTorch 2.6. When importing
bitsandbytes
without any GPUs visible in an environment with Triton installed, a RuntimeError may be raised:RuntimeError: 0 active drivers ([]). There should only be one.
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.1...0.45.2
v0.45.1
Compare Source
Improvements:
triton>=3.2.0
pyproject.toml
Bug Fixes:
Dependencies:
v0.45.0
Compare Source
This is a significant release, bringing support for LLM.int8() to NVIDIA Hopper GPUs such as the H100.
As part of the compatibility enhancements, we've rebuilt much of the LLM.int8() code in order to simplify for future compatibility and maintenance. We no longer use the col32 or architecture-specific tensor layout formats while maintaining backwards compatibility. We additionally bring performance improvements targeted for inference scenarios.
Performance Improvements
This release includes broad performance improvements for a wide variety of inference scenarios. See this X thread for a detailed explanation.
Breaking Changes
🤗PEFT users wishing to merge adapters with 8-bit weights will need to upgrade to
peft>=0.14.0
.Packaging Improvements
Deprecations
FutureWarning
when used. These functions will become unavailable in future releases. This should have minimal impact on most end-users.block_wise=False
is not recommended and support will be removed in a future release.Full Changelog
v0.44.1
Compare Source
Bug fixes:
v0.44.0
Compare Source
New: AdEMAMix Optimizer
The AdEMAMix optimizer is a modification to AdamW which proposes tracking two EMAs to better leverage past gradients. This allows for faster convergence with less training data and improved resistance to forgetting.
We've implemented 8bit and paged variations:
AdEMAMix
,AdEMAMix8bit
,PagedAdEMAMix
, andPagedAdEMAMix8bit
. These can be used with a similar API to existing optimizers.Improvements:
Full Changelog:
v0.43.3
Compare Source
Improvements:
Params4bit.__new__
post PR #970. It supports models exported with non-defaultquant_storage
, such as this NF4 model with BF16 storage.v0.43.2
Compare Source
This release is quite significant as the QLoRA bug fix big implications for higher
seqlen
and batch sizes.For each sequence (i.e. batch size increase of one) we expect memory savings of:
seqlen=1024
, and 4888GB forseqlen=128,00
seqlen=1024
and 1258GB forseqlen=128,00
This was due to activations being unnecessary for frozen parameters, yet the memory for them was still erroneously allocated due to the now fixed bug.
Improvements:
Bug Fixes
str2optimizer32bit
(#1222, thanks @EtienneDosSantos)v0.43.1
Compare Source
Improvements:
Bug Fixes
Internal Improvements:
v0.43.0
Compare Source
Improvements and New Features:
Bug Fixes:
Backwards Compatibility
v0.42
tov0.43
, when using 4bit quantization, models may generate slightly different outputs (approximately up to the 2nd decimal place) due to a fix in the code. For anyone interested in the details, see this comment.Internal and Build System Enhancements:
Contributors:
This release is made possible thanks to the many active contributors that submitted PRs and many others who contributed to discussions, reviews, and testing. Your efforts greatly enhance the library's quality and user experience. It's truly inspiring to work with such a dedicated and competent group of volunteers and professionals!
We give a special thanks to @TimDettmers for managing to find a little bit of time for valuable consultations on critical topics, despite preparing for and touring the states applying for professor positions. We wish him the utmost success!
We also extend our gratitude to the broader community for your continued support, feedback, and engagement, which play a crucial role in driving the library's development forward.
Configuration
📅 Schedule: Branch creation - "after 5am on saturday" (UTC), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
To execute skipped test pipelines write comment
/ok-to-test
.This PR has been generated by MintMaker (powered by Renovate Bot).