-
Notifications
You must be signed in to change notification settings - Fork 213
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Scaling of Data #34
Comments
The student T likelihood scales with the squared mean distance, which is non-linear w.r.t. to data scaling. bayesian_changepoint_detection/bayesian_changepoint_detection/offline_likelihoods.py Line 138 in 2dd95f5
Also intuitively that makes sense, because the difference of your generative models is now different, and thus the probability of them being the same/different should be different |
Thanks for quick reply. Confusion for me is that often the scale is arbitrary, such as if there are multiple ways to make some data dimensionless, yet that could yield vastly different results; my assumption before was that I should always just always normalize over the entire time series. Is there some prior used here in calculating the student T likelihood that I should keep in mind with how I scale my data, or any other way to decide the scale? |
Good question. I would believe that mean centering your data is probably a good idea. But w.r.t. scaling I have to think a bit more. It has probably to do with an implicit prior somewhere, but I cannot pinpoint it right now. |
Hi,
I've noticed is the scaling of the data can have an effect on the result, but I am not sure why it would and can't find any reason for it in the code or references. Below I have the CP probabilities for the same data with or without a constant factor, which are somewhat different.
Are there some assumptions about the input data I am missing?
Thanks
The text was updated successfully, but these errors were encountered: