This post is so ratio'd.It’s a simple least squares model based on what he’s explained to me. When I’ve called him out on it he says it ‘does what it is supposed to do’ and then references least squared error. That’s when I went after him on his choice of loss functions.
Anyhow, at a minimum, for early season game predictions to be of any value, he should be considering a Bayesian model where the prior is based on things like returning players, their stats, attendance at road games and other similar things. Picking the prior should not be overly difficult as one in that position has ample historical data to work with and can use a multitude of features from a prior year to predict the following year.
It used to be that things like a Gibbs Sampling / Markov Chain Monte Carlo model was a LOT of work to pull off. In ‘95 I was involved in writing a Bayes Multinomial Probit in a matrix algebra language (SAS/IML) and it sucked... there was no direct function for a Kronecker product if I recall correctly among other things... but now, with the general availability of every model under the sun with Python and R, there’s no excuse for shit models. If the results are not good, it just means the modeler is uninformed or fucking lazy. In this case, I think it may be a combination of the two.
I mean hell, he probably doesn’t even need to go that complex. I imagine that an XGBoost model that included some features to represent past year’s history and some type of feature to represent time (number of game played in current season) would create interactions that would fairly naturally weight early season predictions to the historical data and create a more accurate prediction in weeks 1-5.
Now tell us the Super Lotto numbers and maybe you'll have some credibility around here....
Upvote
0