-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added support for Groq #11
base: main
Are you sure you want to change the base?
Conversation
You can actually use the current OpenAI implementation :)
Taken from here: https://github.com/openai/openai-python?tab=readme-ov-file#module-level-client |
There isn't a single instance of the term "groq" in the link you attached: And Google doesn't return any decent results for "https://api.groq.com/openai/v1/chat/completions" So, no you cannot use the current OpenAI implementation. Not to mention that OpenAI itself recommends not using the module level client you link in application code Thanks ;) |
Oh, looks like I needed to update myself a little bit on the lib and such, here is the thing: Here (https://console.groq.com/docs/quickstart) you can find the endpoint I mentioned: And here is how to use it: Note: Groq is OpenAI compatible. Means that you can use any implementation of that API.
Here's the output:
Please let me know if it works for you! |
there has been a rate limit on groq api, does anyone know how to bypass it & use it in free tier? |
Groq is delivering the world's fastest LLM inference with very generous free API credits. The default usage is with llama3-70b as it is currently Groq's most advanced free model, though paying users have preview access to llama3.1-405B. To use llama3.1-405B, replace all instances of
llama3-70b-8192
byllama-3.1-405b-reasoning