-
Notifications
You must be signed in to change notification settings - Fork 518
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to load model from file: /data/DeepSeek R1-Distill-Qwen-32B-Q4 K M.gguf #2830
Comments
llama-cpp-python 升级到 0.3.7 试下。 |
|
尝试容器内升级了一下llama-cpp-python,gguf可以成功加载了,但是因为我升级的方式不对,gguf是按照CPU加载的
我尝试过:
|
我也是一样的,一直编译失败 |
这个llama_cpp_python,我感觉是我安装包里面难安装的,可能还是没摸清门道,按照官网的介绍每次编译不成功 刚刚从dockerfile里面翻了一下: 也还没有造好的轮子,可能得git源码包进行编译 |
llama-cpp-python 跟不上 llama.cpp 的发展了。我们计划自己来做 llama.cpp 的封装。cc @codingl2k1 |
System Info / 系統信息
xinference v1.2.1
Failed to load model from file: /
data/DeepSeek R1-Distill-Qwen-32B-Q4 K M.gguf
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
xinference v1.2.1
The command used to start Xinference / 用以启动 xinference 的命令
xinference v1.2.1
Reproduction / 复现过程
Expected behavior / 期待表现
成功运行DeepSeek R1-Distill-Qwen-32B-Q4 K M.gguf
The text was updated successfully, but these errors were encountered: