Minimum Message Length (MML) |
|
Minimum message length (MML) inference was devised by Chris Wallace and David Boulton c1968 [WB68] and developed by Chris Wallace [Wal05] and many colleagues. MML is a Bayesian method of inference:
for hypothesis H, data D, event E. MML is a practical realisation [All05, All18] of Ockham's razor. Some have assumed that MML is the same as maximum aposterior inference (MAP) but in general it is not. And unlike the later minimum description length (MDL) principle, MML favours explicit priors and fully parameterized models. Key points are that every continuous (real, floating point) variable has some limited measurement accuracy and that every continuous parameter has some optimal limited precision to which it should be inferred and stated. A consequence is that even continuous data and continuous parameters have non-zero probabilities (and hence finite message lengths), not just probability densities, and therefore Bayes's theorem still applies as is. Interestingly, there are many cases where even a discrete parameter must be estimated to less precision than its discreteness would seem to allow. Some statistical models that have been MML-ed include:
MML has theoretical support in the form of Kolmogorov complexity. Strict MML (SMML) is a sort of MML "gold standard". Unfortunately SMML is computationally intractible for all but simple problems but, happily, accurate and feasible approximations to SMML exist [WF87]. References
|
|
↑ © L. Allison, www.allisons.org/ll/ (or as otherwise indicated). Created with "vi (Linux)", charset=iso-8859-1, fetched Monday, 02-Oct-2023 21:29:54 UTC. Free: Linux, Ubuntu operating-sys, OpenOffice office-suite, The GIMP ~photoshop, Firefox web-browser, FlashBlock flash on/off. |