Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

High codebook utilization of the mimi model #187

Open
1 task done
ciqiangchen opened this issue Jan 18, 2025 · 0 comments
Open
1 task done

High codebook utilization of the mimi model #187

ciqiangchen opened this issue Jan 18, 2025 · 0 comments
Labels
question Further information is requested

Comments

@ciqiangchen
Copy link

Due diligence

  • I have done my due diligence in trying to find the answer myself.

Topic

The PyTorch implementation

Question

When I attempted to reproduce the MIMI codec, I found that my codebook utilization was very low in all layers except for the first distillation layer. Meanwhile, I noticed that the codebook utilization of all RVQ layers in the official checkpoint was very high.

Is there any technique to improve the utilization of these codebooks, especially for the RVQ layers other than the distillation layer? I’m looking forward to your reply. Thank you!

@ciqiangchen ciqiangchen added the question Further information is requested label Jan 18, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant