Skip to content

该模型能否在llama.cpp上运行 #42

Answered by OleehyO
KagaJiankui asked this question in Q&A
Discussion options

You must be logged in to vote

TexTeller的架构是自定义的,所以llama.cpp/fastllm.cpp应该是没有兼容的,如果想转到纯C++框架可以考虑把模型转ONNX然后使用TensorRT进行推理。

另外,这一版的TexTeller在某些场景下还不够好,所以还没有考虑q4/q8,但是下一版应该会有。

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by OleehyO
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #42 on June 07, 2024 06:10.