Multi Modality #2453
aamir-gmail
started this conversation in
General
Multi Modality
#2453
Replies: 2 comments 1 reply
-
It supports image to text, but it does not support text to image. |
Beta Was this translation helpful? Give feedback.
0 replies
-
when you say " It supports image to text", you mean multimodal models like
LLava
https://llava-vl.github.io/
…On Mon, Sep 16, 2024 at 6:22 PM Kevin Hu ***@***.***> wrote:
It supports image to text, but it does not support text to image.
—
Reply to this email directly, view it on GitHub
<#2453 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AJA2ECPGGASMCNVEURD6ZYTZW2IMRAVCNFSM6AAAAABOIUOIJSVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTANRVGY3TEMY>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
--
Kind Regards
Aamir Mirza
|
Beta Was this translation helpful? Give feedback.
1 reply
# for free
to join this conversation on GitHub.
Already have an account?
# to comment
-
I see there i support for image2text in the supported model in ollama , I would resume they relate to llava:13b . I tried using them however the image I uploaded in the chat ended up in the knowledge base. My question is whether there is support for using Multi-modal models like llava 7 or 13 b
Beta Was this translation helpful? Give feedback.
All reactions