Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: Fietje 2 model won't download & run #3015

Closed
jasperslot opened this issue Jun 9, 2024 · 5 comments
Closed

bug: Fietje 2 model won't download & run #3015

jasperslot opened this issue Jun 9, 2024 · 5 comments
Labels
P1: important Important feature / fix type: bug Something isn't working

Comments

@jasperslot
Copy link

Describe the bug
I have tried to use the Fietje 2 model (https://huggingface.co/BramVanroy/fietje-2-chat-gguf) to run, but it doesn't run. It does work in Ollama (https://ollama.com/bramvanroy/fietje-2b-chat:Q8_0) without issues. There is no clear message why it doesn't start. I have tried by adding the model directly by copy & paste the huggingface url en choose the Q8 model, but it doesn't download (stuck at 0%). So, I downloaded the GGUF manually and imported into Jan. This works, but when I try to start the model, it doesn't start, but I also don't get a clear error message.

Expected behavior
I expect the model to run, just like with Ollama.

Environment details

  • Operating System: MacOS Sonoma 14.4.1,
  • Jan Version: 0.5.0
  • Processor: Apple M3 Max
  • RAM: 64GB

Logs
app.log

@jasperslot jasperslot added the type: bug Something isn't working label Jun 9, 2024
@Van-QA Van-QA added the P1: important Important feature / fix label Jun 10, 2024
@namchuai
Copy link
Contributor

Thanks @jasperslot ! I checked this and got those below error.

jan:dev: 2024-06-10T03:48:39.878Z [CORTEX]::Error: llama_model_loader: - kv  16:                      tokenizer.ggml.merges arr[str,50000]   = ["Ġ t", "Ġ a", "h e", "i n", "r e",...
jan:dev: llama_model_loader: - kv  17:                tokenizer.ggml.bos_token_id u32              = 50295
jan:dev: llama_model_loader: - kv  18:                tokenizer.ggml.eos_token_id u32              = 50296
jan:dev: llama_model_loader: - kv  19:            tokenizer.ggml.unknown_token_id u32              = 50256
jan:dev: llama_model_loader: - kv  20:            tokenizer.ggml.padding_token_id u32              = 50296
jan:dev: llama_model_loader: - kv  21:                    tokenizer.chat_template str              = {% for message in messages %}{{'<|im_...
jan:dev: llama_model_loader: - kv  22:               general.quantization_version u32              = 2
jan:dev: llama_model_loader: - type  f32:  259 tensors
jan:dev: llama_model_loader: - type q8_0:  194 tensors
jan:dev: 
jan:dev: 2024-06-10T03:48:39.910Z [CORTEX]::Error: llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'phi-2'
jan:dev: llama_load_model_from_file: failed to load model
jan:dev: 
jan:dev: 2024-06-10T03:48:39.912Z [CORTEX]::Error: llama_init_from_gpt_params: error: failed to load model '/Users/_/jan/models/Q8_0/fietje-2b-chat-Q8_0.gguf/fietje-2b-chat-Q8_0.gguf'
jan:dev: 
jan:dev: 2024-06-10T03:48:39.912Z [CORTEX]::Debug: {"timestamp":1717991319,"level":"ERROR","function":"LoadModel","line":168,"message":"llama.cpp unable to load model","model":"/Users/_/jan/models/Q8_0/fietje-2b-chat-Q8_0.gguf/fietje-2b-chat-Q8_0.gguf"}
jan:dev: 20240610 03:48:39.912616 UTC 3911969 ERROR Error loading the model - llama_engine.cc:385

@Van-QA Van-QA added this to the v.0.5.1 🍖 Kebap milestone Jun 10, 2024
@Van-QA
Copy link
Contributor

Van-QA commented Jun 12, 2024

Adding related ticket: ggerganov/llama.cpp#7219

@Van-QA Van-QA modified the milestones: v.0.6.0, v0.6.1 Aug 5, 2024
@Van-QA Van-QA moved this to Icebox in Jan & Cortex Aug 5, 2024
@imtuyethan imtuyethan moved this from Icebox to Need Investigation in Jan & Cortex Sep 2, 2024
@imtuyethan imtuyethan moved this from Need Investigation to Icebox in Jan & Cortex Sep 2, 2024
@imtuyethan
Copy link
Contributor

What are the latest updates on this? Cannot be solved? @Van-QA @namchuai

@0xSage
Copy link
Contributor

0xSage commented Sep 5, 2024

@jasperslot Seems fixed by llamacpp upstream.
Closing
Here's my screenshot:
image

@0xSage 0xSage closed this as completed Sep 5, 2024
@github-project-automation github-project-automation bot moved this from Icebox to Completed in Jan & Cortex Sep 5, 2024
@jasperslot
Copy link
Author

Hi, can confirm its working in Jan 0.5.3 now 🎉

@imtuyethan imtuyethan added this to the v0.5.4 milestone Sep 11, 2024
@imtuyethan imtuyethan changed the title Fietje 2 model won't download & run bug: Fietje 2 model won't download & run Sep 11, 2024
@imtuyethan imtuyethan removed this from the v0.5.4 milestone Sep 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
P1: important Important feature / fix type: bug Something isn't working
Projects
Archived in project
Development

No branches or pull requests

5 participants