Skip to content

For Mistral 7B - Generate Text using Input Embeddings + Add no_repeat_ngram_size Support #86

New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Open
wants to merge 6 commits into
base: main
Choose a base branch
from

Conversation

davidshtian
Copy link

@davidshtian davidshtian commented Apr 18, 2024

Description of changes:

Currently, the Mistral 7B model does not take input_embeds (skip the embedding layer) as accepted input, but only input_ids. While lots of multimodal models are using Mistral as base LLM, and will send the input_embeds directly to LLM, it would be great and general to support this feature. And also add no_repeat_ngram_size sampling parameter support. Hope it helps~ Thanks~

By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.

@davidshtian davidshtian changed the title Generate Text from Mistral 7B using Embeddings For Mistral 7B - Generate Text using Input Embeddings + Add no_repeat_ngram_size Support May 17, 2024
Copy link

@jianyinglangaws jianyinglangaws left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me!

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants