-
Notifications
You must be signed in to change notification settings - Fork 529
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The status of this repository? #660
Comments
Are you having any problems or have any other ideas? Let's discuss it together. |
OK. Is GPTCache integrated into Langchain? Readme mentions this but the link is dead.
|
because the langchain doc link has changed, you can visit the langchain repo to find the doc. I think it's https://python.langchain.com/api_reference/community/cache/langchain_community.cache.GPTCache.html#gptcache |
Thanks for the link. I wonder is there other way to support locally deployed huggingface model (e.g. llama2) , rather than using API? Could you please suggest? |
Currently, GPTCache supports the APIs of OpenAI and Cohere. You can try to package your embedding service into an API that is compatible with them. Similar to how most LLM models are now compatible with OpenAI’s SDK. |
Great. Thanks for the advice. |
Sorry I may not have thoroughly undertstood your meaning. What is the "embedding service"? Is it the "embedding generator" in your paper? |
Thanks. I wonder how can I manage the cache such as delete some entry in the cache. I notice that there are no such APIs in api.py. From the doc the way to achieve seperation is via Session. But I want to know can I delete some entry in the cache? Thanks! |
You can try to learn about eviction manager., link: https://github.com/zilliztech/GPTCache/blob/main/gptcache/manager/eviction_manager.py |
Hi. GPTCache is cool, but there has been no commits for 5 months. Howeve readme mentions that "repository is still under heavy development".
Can anyone explain what is the situation now for the repo?
The text was updated successfully, but these errors were encountered: