Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MNIST classification #1

Open
janisgp opened this issue Jun 23, 2019 · 1 comment
Open

MNIST classification #1

janisgp opened this issue Jun 23, 2019 · 1 comment

Comments

@janisgp
Copy link

janisgp commented Jun 23, 2019

I am curious how the classification setting works. You mention in your paper that you use the cross entropy loss.

Do you use as final layer a softmax? How do you propagate the variance through the softmax?

@js05212
Copy link
Owner

js05212 commented Jul 18, 2019

Hi,

Thanks for the interest and good question! For MNIST classification, we use elementwise sigmoid followed by cross entropy. The output mean of the sigmoid will take the mean and variance from the previous layer (pre-activation linear layer) as input. This is how both the mean and variance can affect the final prediction.

There has also been follow-ups of NPN (e.g., work from ICLR 2018 if I remember correctly) trying to extend it with softmax layer.

Hao

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants