Using Llava or other vision-language multi-models with Ollama component? #3095
Replies: 1 comment 1 reply
-
I'm back, still hoping to find answers on how to use vision-language multi-models in langchain using Ollama. As of v1.0 langachian seems to have made note that it is able to input images. Maybe this is true for openAI? I don't know because I don't use it. But it doesn't seem to work with Ollama. I will say when I try now, my error is different then what I reported above before. Now it says "Error building Component Ollama: not enough values to unpack (expected 2, got 1)". I can confirm that it works fine only inputting text. I am using the 'chat input' component as I'm under the impression that's the only way to supply the image and prompt at the moment. Regardless, I am no programmer, I'd take a crack at building a custom component for this task if I had a better understanding of what needed to be done. But apparently no one is using this kind of vision feature in langchain, or at least isn't releasing components for the function on the 'store' or talking about this functionality, like, anywhere at all, because I have been looking. And I know its only been about 2 months, so I'm not trying to sound impatient or anything, but the documentation doesn't really mention this newly added functionality at all, other then the fact it supposedly exists and that more stuff related is planned, as of 1.0. Is ANYONE using these vision capabilities? With OpenAi or otherwise? If a single person has gotten this working with Ollma I'd like to hear about and let me know if I'm just doing something wrong, or if its just not really 'usable' atm but that its in progress so I know if I should wait around for this functionality in Langchain to be more established or if I should look for other options. Literally any info at all would be nice. |
Beta Was this translation helpful? Give feedback.
-
Is it possible to use Llava models with images with the Langflow ollama component?
It tosses an error if I try and send an image to it. It will run without an image but that defeats the purpose of llava. Other models also can't receive images with the ollama component, which is fair enough as I don't think any other models can do anything with images anyway. But it would be nice if the ollama component could receive images?
Unless I am missing something.
Beta Was this translation helpful? Give feedback.
All reactions