Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I noticed that you didn't do data augmentation like in Kdeep original post. #2

Open
AberChiong opened this issue Sep 5, 2023 · 0 comments

Comments

@AberChiong
Copy link

pred[i] = _pred.mean().cpu().numpy()

Hello!

I'm refactoring your code as an important part of my master's project.

After I completed the entire process of pytorch-based programming (referring to your code), I immediately conducted a performance evaluation test of the model. The result is this:
{'RMSE': 1.5596018, 'Spearman Correlation': 0.6496260994698196, 'Pearson Correlation': 0.6565482435856405}

This is extremely inconsistent with the performance evaluation results in the original text of Kdeep, and then I carefully read the original text and your code. I found the following issues:

original text: We augment our data by rotating each subgrid 90 deg, providing 24 times our initial training size. This augmenting methodology is also used at prediction time and then averaged out, reducing variance.
but you: I didn't find any code for data augmentation, which also appeared in the validation of the model (It may also be because my programming ability is too weak or this part of the code has not been seriously discovered).

So I would like to ask you, do you think data augmentation should be done in this model? What is the performance of the model trained in your environment?

Bast wish to u, Im waiting for your reply. :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant