Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Try RWKV (better than LSTM for Language Modeling) #58

Open
BlinkDL opened this issue Aug 20, 2023 · 3 comments
Open

Try RWKV (better than LSTM for Language Modeling) #58

BlinkDL opened this issue Aug 20, 2023 · 3 comments

Comments

@BlinkDL
Copy link

BlinkDL commented Aug 20, 2023

Hi Byron,

Would you like to try RWKV for better compression?

Some experiments by Fabrice:
https://bellard.org/ts_server/ts_zip.html

You can do online training of RWKV and it learns fast.

Bo

@byronknoll
Copy link
Owner

That is a nice idea. If I have time, I might try it out in tensorflow-compress: https://github.com/byronknoll/tensorflow-compress

@xXWarMachineRoXx
Copy link

Isn't the compression ratio 7.382 for enwiki9 with RWKV?

@byronknoll
Copy link
Owner

Fabrice Bellard tried out a pre-trained RWKV model on enwik9: https://bellard.org/ts_zip/

For usage in tensorflow-compress, the model would not be pre-trained (i.e. it would be worse compression rate, but smaller decompressor size).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants