Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Mistral's new visual model: Pixtral-12b-240910 #6748

Open
awaescher opened this issue Sep 11, 2024 · 36 comments
Open

Support Mistral's new visual model: Pixtral-12b-240910 #6748

awaescher opened this issue Sep 11, 2024 · 36 comments
Labels
model request Model requests

Comments

@awaescher
Copy link
Contributor

Mistral AI just dropped Pixtral, their 12b model with vision support.

@awaescher awaescher added the model request Model requests label Sep 11, 2024
@Qualzz
Copy link

Qualzz commented Sep 11, 2024

Gosh I love them.

@Kerimcanak
Copy link

Support.
I want to see this model to be usable on Ollama.

@tuanlda78202
Copy link

Hope too ~

@briansan
Copy link

+1

@awaescher
Copy link
Contributor Author

Kudos to the Ollama team ❤️

image

(source)

@abenmrad
Copy link

Related to Pixtral, but more generally regarding Multimodal support in Ollama:

From my experiments today, Ollama is still supporting multi-modal chat with LLaVa (retried today with v0.3.10).

There were indeed some changes in the Llama.cpp server a while back.

I was genuinely interested to understand how Ollama can still handle it while Llama.cpp reportedly cannot anymore. Was Ollama relying on llama-cli, not llama-server?

Turns out it's not relying on either. Ollama is integrating directly with the llama.cpp code base. As for LLaVa support, they lifted the LLaVa support directly from the Llama.cpp server codebase and have been maintaining that in addition to everything else since then...

Ollama team are truly unsung heroes in this technological revolution.

@BennisonDevadoss
Copy link

Is it available at Ollama now?

@webboty
Copy link

webboty commented Sep 19, 2024

Is it available at Ollama now?

ollama pull pixtral
pulling manifest
Error: pull model manifest: file does not exist

Not yet.

@dropout-developer
Copy link

Is it available at Ollama now?

ollama pull pixtral pulling manifest Error: pull model manifest: file does not exist

Not yet.

Same for me

@CRCODE22
Copy link

Tried to convert and add this with ollama https://huggingface.co/DewEfresh/pixtral-12b-8bit/tree/main but it seems the architecture is not supported by Ollama (yet).

K:\AI\DewEfresh\pixtral-12b-8bit>ollama create Pixtral-12B-2409
transferring model data 100%
converting model
Error: unsupported architecture

K:\AI\DewEfresh\pixtral-12b-8bit>ollama create --quantize q8_K_M Pixtral-12B-2409
transferring model data 100%
converting model
Error: unsupported architecture

@celobusana
Copy link

celobusana commented Sep 19, 2024

I tried to run the Pixtral with python code using an RTX 4060 with 16GB, but it was not possible :(. Perhaps it would work with a 4090 with 24GB.

torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 60.00 MiB. GPU 0 has a total capacity of 15.70 GiB of which 34.69 MiB is free. Including non-PyTorch memory, this process has 15.64 GiB memory in use. Of the allocated memory 15.50 GiB is allocated by PyTorch, and 10.20 MiB is reserved by PyTorch but unallocated.

@thecr7guy2
Copy link

Cant wait to try the gguf version of Pixtral man
Ollama gonna deliver a christmas present early for us

@ddpasa
Copy link

ddpasa commented Sep 22, 2024

I tried to run the Pixtral with python code using an RTX 4060 with 16GB, but it was not possible :(. Perhaps it would work with a 4090 with 24GB.

torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 60.00 MiB. GPU 0 has a total capacity of 15.70 GiB of which 34.69 MiB is free. Including non-PyTorch memory, this process has 15.64 GiB memory in use. Of the allocated memory 15.50 GiB is allocated by PyTorch, and 10.20 MiB is reserved by PyTorch but unallocated.

You need to quantize to run an 12b model on 16GB hardware.

@Qualzz
Copy link

Qualzz commented Sep 25, 2024

any news ?

@jasondalycan
Copy link

+1 for this feature

@A-REDA-EL
Copy link

any news ?

@frostworx
Copy link

please stop spamming here.

multiple people are subscribed to this issue and patiently wait until it is done.
all of them receive an email for every "+1" - it is annoying and for sure doesn't speed up anything.
thanks for your understanding.

(sorry for another mail, subscribers...)

@robert-mcdermott
Copy link

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

@Qualzz
Copy link

Qualzz commented Oct 1, 2024 via email

@sr-tream
Copy link

sr-tream commented Oct 2, 2024

in the near future, and many users may switch to alternatives that offer this functionality.

Not with local models, I guess.

@Rakhsan
Copy link

Rakhsan commented Oct 3, 2024

How much long any hint pls?

@wijjj
Copy link

wijjj commented Oct 10, 2024

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

@nuryslyrt
Copy link

+1 to this feature :)

@Rakhsan
Copy link

Rakhsan commented Oct 11, 2024

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

there is no Llama3.2 9B THERE ARE 1, 3, 11, 90

@wijjj
Copy link

wijjj commented Oct 15, 2024

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

there is no Llama3.2 9B THERE ARE 1, 3, 11, 90

yeah. Thanks. You know what we mean. Anyways: waiting patiently.

@semidark
Copy link

please stop spamming here.

multiple people are subscribed to this issue and patiently wait until it is done.
all of them receive an email for every "+1" - it is annoying and for sure doesn't speed up anything.
thanks for your understanding.

(sorry for another mail, subscribers...)

Please do your Daily checks for "the drop" without spamming. Holy flipping cow.

@pbasov
Copy link

pbasov commented Oct 22, 2024

Just to point out an alternative for now - LM Studio just released with Pixtral support:
https://lmstudio.ai/blog/lmstudio-v0.3.5

@oderwat
Copy link

oderwat commented Oct 22, 2024

@pbasov Do I understand it right that this is Apple-Silicone only?

@pbasov
Copy link

pbasov commented Oct 23, 2024

@oderwat I believe so, yes, since it's enabled by the MLX engine and llama.cpp still doesn't support it.
ggerganov/llama.cpp#9440

But I'm sure ollama is going to get Pixtral support very soon, seeing that llama3.2 vision support is being rolled in 0.4

@Rakhsan
Copy link

Rakhsan commented Oct 27, 2024

guys I ma save you some time. I learned that pixtral will COME by end of december. top secret. now stop wasting your time checking every few days

@nonetrix
Copy link

@oderwat I believe so, yes, since it's enabled by the MLX engine and llama.cpp still doesn't support it. ggerganov/llama.cpp#9440

But I'm sure ollama is going to get Pixtral support very soon, seeing that llama3.2 vision support is being rolled in 0.4

Seems they are switching to their own inference engine as well, or at least for vision models. Honestly just wished they made some kind of patch set for llama cpp and maintained it, probably more productive imo if llama cpp doesn't want to add them currently

This was referenced Nov 13, 2024
@Leeaandrob
Copy link

👍

@Tianoklein
Copy link

PLEASE!
Allow us to compare 2 or more images via ollama API!

@nonetrix
Copy link

nonetrix commented Dec 5, 2024

PLEASE! Allow us to compare 2 or more images via ollama API!

I'm pretty sure this is a thing that the LLM itself has to support correct me if I'm wrong

@Leeaandrob
Copy link

Leeaandrob commented Dec 5, 2024 via email

@frostworx
Copy link

oh my... I suggest delaying pixtral support by one week for every useless comment here...

(sorry again, subscribers)

unsubscribing

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
model request Model requests
Projects
None yet
Development

No branches or pull requests