-
Notifications
You must be signed in to change notification settings - Fork 7.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DeepSeek-R1 GGUF can't be loaded #3404
Comments
I am having the same issue. Version 3.6.1, Windows 10 Pro. |
I am having the same issue. Version 3.7.0, Windows 11 Pro. I tried many models of several people to run DeepSeek-R1, but without success so far. |
I suspect the chat template needs modification to accommodate the new output format. |
Same here. W11H, v3.7.0 |
Good morning all, Just to confirm, I tried some of the lower parameter versions of bartowski's Deepseek models (7B and 14B) but was receiving an error when trying to load the model into gpt4all. However, I tried loading in the larger model Hope this helps. |
Same here on Bazzite Linux |
I wonder if there is a compatible chat template generated for the 70B model which could be used for 14B or 32B even. |
I wonder if the Llama 8B distill works as well and it's only the QWEN models experiencing the issue. |
I have the same problem. The software is unable to directly open the downloaded gguf file. |
Same here, v3.7.0, W11H. Can't load 14B and 7B model:
|
Also possible solution here: |
the solution did not work for me |
Same problem here. v3.7.0, Win11. Can't load 7B, 8B, 14B DeepSeek-R1 Llama and Qwen models. |
Same problem, Win 11 |
PLEASE MAKE DEEPSEEK compatible to GPT4All, As I had tried all GGFUF of Deepseek, but none are working. LMStudio is working super fast and completely fine with Deepseek, so make it compatible too. |
Same issue here Windows 11 Pro 23H2 |
This is not an issue, GPT4ALL docs clearly say custom models are not supported OOTB and it is up to user to additional configuration to work. This issue should close. |
A pr was merged into the repo a few hours ago that fixes this, but if you want to get it working right now use this chat template which is based on that pr: {%- if not add_generation_prompt is defined %}
{%- set add_generation_prompt = false %}
{%- endif %}
{%- if messages[0]['role'] == 'system' %}
{{- messages[0]['content'] }}
{%- endif %}
{%- for message in messages %}
{%- if message['role'] == 'user' %}
{{- '<|User|>' + message['content'] }}
{%- endif %}
{%- if message['role'] == 'assistant' %}
{%- set content = message['content'].split('</think>', 1) | last %}
{{- '<|Assistant|>' + content + '<|end▁of▁sentence|>' }}
{%- endif %}
{%- endfor -%}
{%- if add_generation_prompt %}
{{- '<|Assistant|>' }}
{%- endif %} |
Thanks you so much for this, it works now. But can it be possible or any plan of UI changes for Thinking Part, as it looks very good in DeepSeek online or LM Studio, the THINKING-REASONING PART should look beautifully in small box with small font with brain ICON blinking. |
Bug Report
GPT4All downloads the DeepSeek-R1 library, but receive an error attempting to load the model for chat.
Steps to Reproduce
Expected Behavior
GGuf file should load as with all of the others. Ollama loads the same file fine.
Your Environment
I notice in the Chat Template settings an error is present:
Syntax error: 1:267: error: This feature has not been supported yet
The text was updated successfully, but these errors were encountered: