-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Support for gemma3 from google #12963
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Need to upgrade to Ollama v0.6, this should add support for gemma3 |
@puffer-duck But v0.6 not support intel card accelerate, am i right? |
"error": { |
Yes we need this. Can Intel publish patches to Ollama so that we can compile ourselves? Or setup automatic nightly builds that follow the latest version of Ollama. |
Note this was brought up here: #12950 That is generally about the version disparity versus just |
Please! |
Hi All, Gemma3 is now supported in ipex-llm llamacpp! (Ollama support is in progress—we'll provide updates once it's ready.) Important Notes: The 27B Gemma3 q4_k_m model requires >16GB VMem.
Get Started:Please follow the following steps to try it out: 1. Download the latest ipex-llm llamacpp portable zip:2. Get mmproj.gguf & gemma3 gguf model filesPlease download the pre-quantized version from HF: https://huggingface.co/collections/ggml-org/gemma-3-67d126315ac810df1ad9e913 (You must download both the text model and the mmproj file) Note: Vision capability is available on these model sizes: 3. Run gemma33.1 Linuxngl=99 3.1.1 Text only
3.1.2 Single turn (Vision)
3.2.2 Chat mode (Vision)
3.2 WIN3.2.1 Text only
3.2.2 Single turn (Vision)
3.2.3 Chat mode (Vision)
|
Thank you for your efforts. I see that the portable zip of ollama in the pre-release has been updated to 20250313, but it seems that the ollama portable zip still cannot run gemma3 properly. |
will it be possible to run on server or python? many thanks |
We will release the Ollama portable zip with gemma3 support soon. |
Hi All, you may install our latest version of ipex-llm ollama via
You may see ipex-llm ollama quickstart for more details. |
@sgwhat No Ollama Portable Zip ? |
I've tried the official gemma3 models in 4b and 12b, as well as the q4_K_M versions from ollama, and then also |
@ExplodingDragon @yizhangliu Releasing. You may see #12963 (comment) to run it first. |
@sgwhat It looks good. Are there any plans to submit the Ollama patch to the upstream? OneAPI already offers out-of-the-box support on certain systems like ArchLinux. Could you consider providing a statically linked Ollama or similar package? |
Thanks. But it's not easy to do "pip install --pre --upgrae ipex-llm[cpp]". |
Hi All, The Ollama portable zip is now available! Please follow the instructions link to download. Note 1: For now, you need to either use ModelScope as the model source (see details here: link), or run a local GGUF model downloaded from HuggingFace (see details here: link) for Gemma3 Note 2: The text input support for Gemma3 is ready, while the image input support is still WIP for Ollama |
After deployment, I asked a few questions about pictures, but the answers were incorrect. |
Hi @cunkai, currently ipex-llm ollama Gemma3 does not have good support for the image part; we have only fully supported the text part. We will add full support in a future ipex-llm ollama 0.6.x release. |
For now, you may also run a local GGUF model downloaded from HF; see #12963 (comment) |
Edit: It works with a model file from HF. Unfortunately I can't get it to work:
I'll try to download from HF instead. |
Tried
|
See #12963 (comment) |
Yesterday I tried to run it directly through lama-сpp - and it worked (on 2-3x A770). On the model gemma3 27B Q8 (modelscope.cn/lmstudio-community/gemma-3-27b-it-GGUF:Q8_0) this command:
|
Yes, the support is llama.cpp is complete (see #12963 (comment)); the image support in Ollama is still in progress |
It's OK. |
Using the GGUF version and the instructions from #12963 (comment) along with the portable version made it work. I'm getting some strange results from the GGUF version, but I'm seeing those strange results on my AMD-based machine too, so that seems unrelated. Thanks for the help y'all! |
@jason-dai no luck with that either
on the
|
I got the same error, got it working by downloading models from HF instead. |
Same i am facing the same issue just like everyone
|
Today I tried again (
|
Hi all, we are working on upgrading ipex-llm ollama version to re-support gemma3. Before that, you may manage to run |
currently not working... |
Has this been fixed in the latest IPEX-LLM v2.2.0 release? |
Not yet, we could support 0.6.2 within this week. |
I can't pull from docker image from "intelanalytics/ipex-llm-inference-cpp-xpu:latest". |
ollama just update to 0.6.5, can this fix the gemma3 gguf model file problem? with: |
Hi @rafasaurus @zimoai @DocMAX, I will release an initial version to support gemma3, maybe next Monday or Tuesday. |
Really looking forward to the latest IPEX Ollama for Gemma 3! The interleaved sliding window attention will be extremely useful for long context windows! Thanks again for all of the work you're doing here!!!! |
Does the 2.3.0-nightly build add support for Gemma3? |
Not sure, i am running docker version |
Same here:
|
Currently we have supported gemma3:fp16. |
When will there be support for Q4? |
Maybe this week. |
请更新ollama,已支持gemma3
Error: llama runner process has terminated: this model is not supported by your version of Ollama. You may need to upgrade
The text was updated successfully, but these errors were encountered: