We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
centos7,docker:26.0.0
xprobe/xinference:v0.14.0.post1
docker run -itd --name="xjs-inference" -v ./:/app -P -p 9995-9999:9995-9999 --gpus '"device=0,1"' xprobe/xinference:v0.14.0.post1 xinference-local -H 0.0.0.0 --port 9997 --log-level debug
xinference register --model-type embedding --file gte-Qwen2-7B-instruct.json --persist xinference launch --model-name gte-Qwen2-7B-instruct --model-type embedding
能有一个装有nvcc和flash_attn 2.5.6版本以上的docker容器
The text was updated successfully, but these errors were encountered:
同求
Sorry, something went wrong.
xinference0.15.2可用的flash-attn安装包链接:https://pan.baidu.com/s/1OTOKLzKcSukjvDqQ-F6XVQ?pwd=1111 提取码:1111
No branches or pull requests
System Info / 系統信息
centos7,docker:26.0.0
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
xprobe/xinference:v0.14.0.post1
The command used to start Xinference / 用以启动 xinference 的命令
docker run -itd --name="xjs-inference"
-v ./:/app
-P -p 9995-9999:9995-9999
--gpus '"device=0,1"'
xprobe/xinference:v0.14.0.post1
xinference-local -H 0.0.0.0 --port 9997 --log-level debug
Reproduction / 复现过程
xinference register --model-type embedding --file gte-Qwen2-7B-instruct.json --persist
xinference launch
--model-name gte-Qwen2-7B-instruct
--model-type embedding
Expected behavior / 期待表现
能有一个装有nvcc和flash_attn 2.5.6版本以上的docker容器
The text was updated successfully, but these errors were encountered: