v0.1.20
🎉 Introduction to new functions of GPTCache
- support the
temperature
param, like openai
A non-negative number of sampling temperature, defaults to 0.
A higher temperature makes the output more random.
A lower temperature means a more deterministic and confident output.
- Add llama adapter
from gptcache.adapter.llama_cpp import Llama
llm = Llama('./models/7B/ggml-model.bin')
answer = llm(prompt=question)
Full Changelog: 0.1.19...0.1.20