Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Port SparseML Remote Code Fix #21

Merged
merged 1 commit into from
Jul 12, 2024
Merged

Port SparseML Remote Code Fix #21

merged 1 commit into from
Jul 12, 2024

Conversation

Satrat
Copy link
Contributor

@Satrat Satrat commented Jul 11, 2024

SUMMARY:
Porting this SparseML PR for adding remote code support: neuralmagic/sparseml#2349

TEST PLAN:
manually tested loading microsoft/Phi-3-mini-128k-instruct

@Satrat Satrat merged commit 94f6a7b into main Jul 12, 2024
8 of 12 checks passed
@Satrat Satrat deleted the fix/remotecode branch July 12, 2024 03:18
markmc pushed a commit to markmc/llm-compressor that referenced this pull request Nov 13, 2024
* test forward (vllm-project#16)

* test frozen (vllm-project#17)

* test frozen

* rename

* lifecycle conftest (vllm-project#21)

* test initalize (vllm-project#18)

* test initalize

* newline

* parametrize weights and inp_act

* remove dup

* test lifecycle (vllm-project#19)

* test lifecycle

* comments

* comments

* add quantization test

* Lifecycle/min max obs (vllm-project#20)

* min max test

* add minmax obs

* test scale range and min_max update

* rebase

* rebase

* fix

* fix
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants