Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: add optional non-breaking ViT-L/14@336px support #33

Open
yurijmikhalevich opened this issue Jan 24, 2023 · 2 comments
Open

feat: add optional non-breaking ViT-L/14@336px support #33

yurijmikhalevich opened this issue Jan 24, 2023 · 2 comments
Labels
feat New feature or request priority:low Low priority issues

Comments

@yurijmikhalevich
Copy link
Owner

yurijmikhalevich commented Jan 24, 2023

For those who wants extra search quality and ready to sacrifice speed for it:

Stats from a dataset of 269 photos on a much slower NAS (using the v1.2.3 and v2.0.0-pre AppImage distributions):

  • ingestion: 3m56.626s vs 125m0.507s (+3070%)
  • search: 0m18.064s vs 3m19.742s (+1006%)

More details are here.

Like and comment if you want to see the largest CLIP model in rclip.

@yurijmikhalevich yurijmikhalevich added feat New feature or request priority:low Low priority issues labels Jan 24, 2023
@unanthropomorph
Copy link

I am a bit late to this but since you're already using openclip I'd recommend using laion/CLIP-ViT-B-16-DataComp.XL-s13B-b90K or laion/CLIP-ViT-B-32-256x256-DataComp-s34B-b86K as they yield only a slightly lower performance than ViT-L/14@336 with roughly 1/3rd of the parameters.
Even better, give the user the option to pick an openclip model from huggingface.

@yurijmikhalevich
Copy link
Owner Author

@unanthropomorph, thank you for flagging this! Yes, this is what I was considering, too.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat New feature or request priority:low Low priority issues
Projects
None yet
Development

No branches or pull requests

2 participants