Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

feat:support any num_heads for get_alibi_slope #200

Merged
merged 1 commit into from
Apr 11, 2024

Conversation

yz-tang
Copy link
Contributor

@yz-tang yz-tang commented Apr 9, 2024

When I was using flashinfer, I encountered that the heads of some models were not powers of 2. I refer to flashinfer/python/tests/alibi_reference.py, modifies this part of the C++ code.

Copy link
Collaborator

@yzh119 yzh119 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@yz-tang thank you for the improvement!

@yzh119 yzh119 merged commit b217a6f into flashinfer-ai:main Apr 11, 2024
yzh119 added a commit that referenced this pull request May 2, 2024
🤖 I have created a release *beep* *boop*
---


##
[0.0.4](v0.0.3...v0.0.4)
(2024-05-01)


### Features

* pytorch 2.3 support
* more gqa group sizes
* add mma instructions for fp8
([#179](#179))
([d305798](d305798))
* mma rowsum for fp8
([#180](#180))
([5af935c](5af935c))
* support any num_heads for get_alibi_slope
([#200](#200))
([b217a6f](b217a6f))


### Bug Fixes

* fix python package dispatch error message
([#182](#182))
([8eed01c](8eed01c))

---
This PR was generated with [Release
Please](https://github.com/googleapis/release-please). See
[documentation](https://github.com/googleapis/release-please#release-please).

---------

Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: Zihao Ye <expye@outlook.com>
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants