Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

fix(deps): update dependency bitsandbytes to ^0.45.0 #169

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

renovate[bot]
Copy link
Contributor

@renovate renovate bot commented Oct 30, 2024

This PR contains the following updates:

Package Change Age Adoption Passing Confidence
bitsandbytes (changelog) ^0.43.1 -> ^0.45.0 age adoption passing confidence

Release Notes

bitsandbytes-foundation/bitsandbytes (bitsandbytes)

v0.45.1

Compare Source

Improvements:
  • Compatibility for triton>=3.2.0
  • Moved package configuration to pyproject.toml
  • Build system: initial support for NVIDIA Blackwell B100 GPUs, RTX 50 Blackwell series GPUs and Jetson Thor Blackwell.
    • Note: Binaries built for these platforms are not included in this release. They will be included in future releases upon the availability of the upcoming CUDA Toolkit 12.7 and 12.8.
Bug Fixes:
  • Packaging: wheels will no longer include unit tests. (#​1478)
Dependencies:
  • Sets the minimum PyTorch version to 2.0.0.

v0.45.0

Compare Source

This is a significant release, bringing support for LLM.int8() to NVIDIA Hopper GPUs such as the H100.

As part of the compatibility enhancements, we've rebuilt much of the LLM.int8() code in order to simplify for future compatibility and maintenance. We no longer use the col32 or architecture-specific tensor layout formats while maintaining backwards compatibility. We additionally bring performance improvements targeted for inference scenarios.

Performance Improvements

This release includes broad performance improvements for a wide variety of inference scenarios. See this X thread for a detailed explanation.

Breaking Changes

🤗PEFT users wishing to merge adapters with 8-bit weights will need to upgrade to peft>=0.14.0.

Packaging Improvements
  • The size of our wheel has been reduced by ~43.5% from 122.4 MB to 69.1 MB! This results in an on-disk size decrease from ~396MB to ~224MB.
  • Binaries built with CUDA Toolkit 12.6.2 are now included in the PyPI distribution.
  • The CUDA 12.5.0 build has been updated to CUDA Toolkit 12.5.1.
Deprecations
  • A number of public API functions have been marked for deprecation and will emit FutureWarning when used. These functions will become unavailable in future releases. This should have minimal impact on most end-users.
  • The k-bit quantization features are deprecated in favor of blockwise quantization. For all optimizers, using block_wise=False is not recommended and support will be removed in a future release.
  • As part of the refactoring process, we've implemented many new 8bit operations. These operations no longer use specialized data layouts.
Full Changelog

v0.44.1

Compare Source

Bug fixes:

v0.44.0

Compare Source

New: AdEMAMix Optimizer

The AdEMAMix optimizer is a modification to AdamW which proposes tracking two EMAs to better leverage past gradients. This allows for faster convergence with less training data and improved resistance to forgetting.

We've implemented 8bit and paged variations: AdEMAMix, AdEMAMix8bit, PagedAdEMAMix, and PagedAdEMAMix8bit. These can be used with a similar API to existing optimizers.

Improvements:
  • 8-bit Optimizers: The block size for all 8-bit optimizers has been reduced from 2048 to 256 in this release. This is a change from the original implementation proposed in the paper which improves accuracy.
  • CUDA Graphs support: A fix to enable CUDA Graphs capture of kernel functions was made in #​1330. This allows for performance improvements with inference frameworks like vLLM. Thanks @​jeejeelee!
Full Changelog:

Configuration

📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).

🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.

Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.

🔕 Ignore: Close this PR and you won't be reminded about this update again.


  • If you want to rebase/retry this PR, check this box

This PR was generated by Mend Renovate. View the repository job log.

@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from c5ef113 to 02ba822 Compare November 6, 2024 09:18
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from 02ba822 to 6665494 Compare November 19, 2024 16:05
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from 6665494 to 4bdda35 Compare December 5, 2024 17:21
@renovate renovate bot changed the title fix(deps): update dependency bitsandbytes to ^0.44.0 fix(deps): update dependency bitsandbytes to ^0.45.0 Dec 5, 2024
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from 4bdda35 to c65c3ba Compare December 17, 2024 09:13
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from c65c3ba to fc52255 Compare December 28, 2024 13:55
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch 8 times, most recently from 8df7d79 to dbcafd0 Compare January 28, 2025 09:49
@renovate renovate bot force-pushed the renovate/bitsandbytes-0.x branch from dbcafd0 to 41aad91 Compare January 30, 2025 09:56
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

0 participants