Update dependency bitsandbytes to ^0.47.0#133
Update dependency bitsandbytes to ^0.47.0#133red-hat-konflux[bot] wants to merge 1 commit intokonflux-pocfrom
Conversation
|
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: red-hat-konflux[bot] The full list of commands accepted by this bot can be found here. DetailsNeeds approval from an approver in each of these files:Approvers can indicate their approval by writing |
|
Hi @red-hat-konflux[bot]. Thanks for your PR. I'm waiting for a opendatahub-io member to verify that this patch is reasonable to test. If it is, they should reply with Once the patch is verified, the new status will be reflected by the I understand the commands that are listed here. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
3686a0a to
69440dc
Compare
Signed-off-by: red-hat-konflux <126015336+red-hat-konflux[bot]@users.noreply.github.com>
69440dc to
e74dca9
Compare
|
Important Review skippedBot user detected. To trigger a single review, invoke the You can disable this status message by setting the 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Join our Discord community for assistance with any issues or questions. CodeRabbit Commands (Invoked using PR/Issue comments)Type Other keywords and placeholders
CodeRabbit Configuration File (
|
This PR contains the following updates:
^0.42.0->^0.47.0Release Notes
bitsandbytes-foundation/bitsandbytes (bitsandbytes)
v0.47.0Compare Source
Highlights:
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.46.0...0.47.0
v0.46.1Compare Source
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.46.0...0.46.1
v0.46.0: : torch.compile() support; custom ops refactor; Linux aarch64 wheelsCompare Source
Highlights
torch.compilewithout graph breaks for LLM.int8().torch.compilewithout graph breaks for 4bit.fullgraph=False.fullgraph=True.torch.libraryand custom ops APIs. This helps enable ourtorch.compileand additional hardware compatibility efforts.bitsandbytes.Compatability Changes
manylinux_2_24(previouslymanylinux_2_34).What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.4...0.46.0
v0.45.5Compare Source
This is a minor release that affects CPU-only usage of bitsandbytes. The CPU build of the library was inadvertently omitted from the v0.45.4 wheels.
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.4...0.45.5
v0.45.4Compare Source
This is a minor release that affects CPU-only usage of bitsandbytes. There is one bugfix and improved system compatibility on Linux.
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.3...0.45.4
v0.45.3Compare Source
Overview
This is a small patch release containing a few bug fixes.
Additionally, this release contains a CUDA 12.8 build which adds the sm100 and sm120 targets for NVIDIA Blackwell GPUs.
What's Changed
New Contributors
Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.2...0.45.3
v0.45.2Compare Source
This patch release fixes a compatibility issue with Triton 3.2 in PyTorch 2.6. When importing
bitsandbyteswithout any GPUs visible in an environment with Triton installed, a RuntimeError may be raised:RuntimeError: 0 active drivers ([]). There should only be one.Full Changelog: bitsandbytes-foundation/bitsandbytes@0.45.1...0.45.2
v0.45.1Compare Source
Improvements:
triton>=3.2.0pyproject.tomlBug Fixes:
Dependencies:
v0.45.0Compare Source
This is a significant release, bringing support for LLM.int8() to NVIDIA Hopper GPUs such as the H100.
As part of the compatibility enhancements, we've rebuilt much of the LLM.int8() code in order to simplify for future compatibility and maintenance. We no longer use the col32 or architecture-specific tensor layout formats while maintaining backwards compatibility. We additionally bring performance improvements targeted for inference scenarios.
Performance Improvements
This release includes broad performance improvements for a wide variety of inference scenarios. See this X thread for a detailed explanation.
Breaking Changes
🤗PEFT users wishing to merge adapters with 8-bit weights will need to upgrade to
peft>=0.14.0.Packaging Improvements
Deprecations
FutureWarningwhen used. These functions will become unavailable in future releases. This should have minimal impact on most end-users.block_wise=Falseis not recommended and support will be removed in a future release.Full Changelog
v0.44.1Compare Source
Bug fixes:
v0.44.0Compare Source
New: AdEMAMix Optimizer
The AdEMAMix optimizer is a modification to AdamW which proposes tracking two EMAs to better leverage past gradients. This allows for faster convergence with less training data and improved resistance to forgetting.
We've implemented 8bit and paged variations:
AdEMAMix,AdEMAMix8bit,PagedAdEMAMix, andPagedAdEMAMix8bit. These can be used with a similar API to existing optimizers.Improvements:
Full Changelog:
v0.43.3Compare Source
Improvements:
Params4bit.__new__post PR #970. It supports models exported with non-defaultquant_storage, such as this NF4 model with BF16 storage.v0.43.2Compare Source
This release is quite significant as the QLoRA bug fix big implications for higher
seqlenand batch sizes.For each sequence (i.e. batch size increase of one) we expect memory savings of:
seqlen=1024, and 4888GB forseqlen=128,00seqlen=1024and 1258GB forseqlen=128,00This was due to activations being unnecessary for frozen parameters, yet the memory for them was still erroneously allocated due to the now fixed bug.
Improvements:
Bug Fixes
str2optimizer32bit(#1222, thanks @EtienneDosSantos)v0.43.1Compare Source
Improvements:
Bug Fixes
Internal Improvements:
v0.43.0Compare Source
Improvements and New Features:
Bug Fixes:
Backwards Compatibility
v0.42tov0.43, when using 4bit quantization, models may generate slightly different outputs (approximately up to the 2nd decimal place) due to a fix in the code. For anyone interested in the details, see this comment.Internal and Build System Enhancements:
Contributors:
This release is made possible thanks to the many active contributors that submitted PRs and many others who contributed to discussions, reviews, and testing. Your efforts greatly enhance the library's quality and user experience. It's truly inspiring to work with such a dedicated and competent group of volunteers and professionals!
We give a special thanks to @TimDettmers for managing to find a little bit of time for valuable consultations on critical topics, despite preparing for and touring the states applying for professor positions. We wish him the utmost success!
We also extend our gratitude to the broader community for your continued support, feedback, and engagement, which play a crucial role in driving the library's development forward.
Configuration
📅 Schedule: Branch creation - "after 5am on saturday" (UTC), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
To execute skipped test pipelines write comment
/ok-to-test.This PR has been generated by MintMaker (powered by Renovate Bot).