fix(deps): update dependency bitsandbytes to ^0.45.0 #169
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
^0.43.1
->^0.45.0
Release Notes
bitsandbytes-foundation/bitsandbytes (bitsandbytes)
v0.45.1
Compare Source
Improvements:
triton>=3.2.0
pyproject.toml
Bug Fixes:
Dependencies:
v0.45.0
Compare Source
This is a significant release, bringing support for LLM.int8() to NVIDIA Hopper GPUs such as the H100.
As part of the compatibility enhancements, we've rebuilt much of the LLM.int8() code in order to simplify for future compatibility and maintenance. We no longer use the col32 or architecture-specific tensor layout formats while maintaining backwards compatibility. We additionally bring performance improvements targeted for inference scenarios.
Performance Improvements
This release includes broad performance improvements for a wide variety of inference scenarios. See this X thread for a detailed explanation.
Breaking Changes
🤗PEFT users wishing to merge adapters with 8-bit weights will need to upgrade to
peft>=0.14.0
.Packaging Improvements
Deprecations
FutureWarning
when used. These functions will become unavailable in future releases. This should have minimal impact on most end-users.block_wise=False
is not recommended and support will be removed in a future release.Full Changelog
v0.44.1
Compare Source
Bug fixes:
v0.44.0
Compare Source
New: AdEMAMix Optimizer
The AdEMAMix optimizer is a modification to AdamW which proposes tracking two EMAs to better leverage past gradients. This allows for faster convergence with less training data and improved resistance to forgetting.
We've implemented 8bit and paged variations:
AdEMAMix
,AdEMAMix8bit
,PagedAdEMAMix
, andPagedAdEMAMix8bit
. These can be used with a similar API to existing optimizers.Improvements:
Full Changelog:
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR was generated by Mend Renovate. View the repository job log.