Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adjust LLM parameters to speed things up #361

Merged
merged 4 commits into from
Feb 8, 2025
Merged

Adjust LLM parameters to speed things up #361

merged 4 commits into from
Feb 8, 2025

Conversation

MrOrz
Copy link
Member

@MrOrz MrOrz commented Feb 7, 2025

The most recent model, gemini-2.0-flash-001, suffers from slow response issue, compared to its predecessor, the experimental gemini-2.0-flash-exp.
圖片

On the other hand, I found that gemini-1.5-pro-002 in asia-east1 (Taiwan) and asia-northeast1 (Tokyo) is actually as fast as gemini-2.0-flash-exp on us-cental1, and its quality is also pretty good.
圖片

This PR sets up the LLM model settings for LLM transcript to prioritize gemini-1.5-pro-002 in Asia, then gemini-2.0-flash-exp. For now, we ignore the latest gemini-2.0-flash-001 until it gets a significant speed boost.

@coveralls
Copy link

Coverage Status

coverage: 83.059% (+0.006%) from 83.053%
when pulling 784b3f5 on llm-adjustment
into e188e81 on master.

@MrOrz MrOrz self-assigned this Feb 8, 2025
@MrOrz MrOrz marked this pull request as ready for review February 8, 2025 07:09
@MrOrz MrOrz merged commit 51658aa into master Feb 8, 2025
4 checks passed
@MrOrz MrOrz deleted the llm-adjustment branch February 8, 2025 07:43
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants