Does vllm support CPU? #999
-
Can we use vllm only on CPU without GPU machine? |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 1 reply
-
I think the short answer is no, as vLLM's engine relies on custom kernels written in CUDA. |
Beta Was this translation helpful? Give feedback.
-
You can try ctranslate2 or llama.cpp. |
Beta Was this translation helpful? Give feedback.
-
This was not clear to me either - any way to highlight this in bold somewhere on main docs? Sorry if I overlooked. I am trying to do some local testing - that's my use case |
Beta Was this translation helpful? Give feedback.
-
Intel CPU is supported from my own experience. |
Beta Was this translation helpful? Give feedback.
Intel CPU is supported from my own experience.