Sure. But for practical purposes (e.g. trading), it’s ok to assume coin flips are I.i.d. For practical purposes we also want to model rare, one-off events; but how should we go about that? Using the same tools we have to model coins seems inappropriate here
-
-
Replying to @ericjang11 @sandersted
You can think of this in the language of scoring rules or decision theory where the probability is a probabilistic forecast. Essentially, the [0, 1) interval is your forecaster’s action space.
1 reply 0 retweets 4 likes -
Then, you can ask the same forecaster to give you probabilistic forecasts of multiple events, and then, say, measure it’s track record over multiple forecasts using a scoring rule S(q_i, x_i) where q_i is the forecast (probability) and x_i is the observed outcome
1 reply 0 retweets 1 like -
The scoring rule S tells you: how much penalty do I get if I published a forecast 0.3 and I then actually observe a positive outcome. The log loss is a common choice.
1 reply 0 retweets 1 like -
In the i.i.d. Case, this gives you your empirical risk, or likelihood. But you can use this to compare the track record of two forecasters over multiple forecasted events.
1 reply 0 retweets 1 like -
Giving a probabilistic forecast makes sense even without assuming that the world is indeed random. You can say I’m giving you a 0.3 probability of rain today rather than tell you it won’t be raining just in case it eventually is raining, and I don’t want to appear very stupid.
1 reply 0 retweets 0 likes -
So I’d suggest reading up on scoring rules (I know) only because it gives you a language to talk/think about this. Also, you can always invoke
@maosbot to give you the full Bayesian dogmatic answer.2 replies 0 retweets 1 like -
Thanks
@fhuszar ! I think your comments assume that the forecasters can be “calibrated” a la scoring rule by making predictions on multiple events. That assumes that there is some mutual information to be had between likelihoods of past and future events?1 reply 0 retweets 0 likes -
Replying to @ericjang11 @fhuszar and
I.e I could have a great “track record” forecasting horse races, but it’s not clear that the statistical model I calibrated on horse betting is useful for 2016 election
3 replies 0 retweets 0 likes -
Replying to @ericjang11 @fhuszar and
There are some bounds you can get on the generalization error of your model when doing domain adaptation -
@AnimaAnandkumar had a paper about it very recently. In your particular example, I suspect most bounds are going to be pretty useless in practice :)1 reply 0 retweets 2 likes
Prof. Anima Anandkumar Retweeted Prof. Anima Anandkumar
This is paper on domain adaptation with shifts in label distribution. Generalization bounds derived were used to regularize estimation of importance weights. These had a significant impact in practice. So it was useful in practicehttps://twitter.com/AnimaAnandkumar/status/1078992995901456385 …
Prof. Anima Anandkumar added,
-
-
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.