The value must be chosen carefully
Probability formalizes uncertainty using set theory from [unresolved:ch-sets]. Events live in a σ-algebra (see [unresolved:def-sigma-algebra]), and measures quantify likelihoods.
# | Statement |
---|---|
1 | Nonnegativity: P(E) ≥ 0. |
2 | Normalization: P(Ω) = 1. |
3 | Countable additivity for disjoint events. |
Conditioning refines uncertainty by restricting to a subset event; its algebra mirrors set-intersection (see [unresolved:def-intersection]).

Random variables map outcomes to numbers; expectations average values with respect to probabilities. Linearity of expectation connects to series appearing in trigonometry (see [unresolved:def-taylor-sine]).
For historical foundations of probability and classical reasoning about inference, see (Kolmogorov, Foundations of the Theory of Probability) and the lucid, modern exposition of Bayesian ideas in [2]. For practical introductions used widely in undergraduate courses, consult [3] and [4].
Methodological developments in hypothesis testing and estimation are surveyed by [5] and summarized in standard texts such as [6, pp. 45-48] and [7, ch. 2]. Modern connections to machine learning appear in [8] and [9].
Applied topics such as density estimation and asymptotics are covered by [10] and [11]. For readers wanting a concise compendium, (Wasserman, All of Statistics: A Concise Course in Statistical Inference, p. 101) is useful. Edge cases: an unknown key will be left as a literal when unresolved — e.g., [does-not-exist ??] — and numeric-looking postnotes such as [14, p. 20] are interpreted as page fragments (rendered as p. 20).
For algorithmic perspectives applied to probability, see [15] and [16, ch. 5]; a combined citation demonstrating multiple keys and per-key extras is: (Knuth, The TeXbook, p. 20; Cormen et al., Introduction to Algorithms, ch. 3; Fisher, Statistical Methods for Research Workers).
Sorting demos for multi-key citations:
- Appearance order (default): (Knuth, The TeXbook; Cormen et al., Introduction to Algorithms; Fisher, Statistical Methods for Research Workers)
- Author alphabetical order: (Cormen et al., Introduction to Algorithms; Fisher, Statistical Methods for Research Workers; Knuth, The TeXbook)
- Author order: [15, 17, 14]
- Alphabetic (title/alpha) order: [15, 17, 14]
References
- [13]does-not-exist
- [03]Sheldon M. Ross. “A First Course in Probability.” Prentice Hall, 2002.
- [12]Larry Wasserman. “All of Statistics: A Concise Course in Statistical Inference.” Springer, 2004.
- [11]Aad van der Vaart. “Asymptotic Statistics.” Cambridge University Press, 1998.
- [10]B. W. Silverman. “Density Estimation for Statistics and Data Analysis.” Chapman and Hall, 1986.
- [01]Andrey N. Kolmogorov. “Foundations of the Theory of Probability.” Chelsea Publishing, 1933.
- [15]Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, Clifford Stein. “Introduction to Algorithms.” MIT Press, 2009.
- [05]Jerzy Neyman, Egon S. Pearson. “On the Problem of the Most Efficient Tests of Statistical Hypotheses.” Philosophical Transactions of the Royal Society A, vol. 231, pp. 289337, 1933.
- [08]Christopher M. Bishop. “Pattern Recognition and Machine Learning.” Springer, 2006.
- [04]Geoffrey Grimmett, David Stirzaker. “Probability and Random Processes.” Oxford University Press, 1992.
- [02]E. T. Jaynes. “Probability Theory: The Logic of Science.” Cambridge University Press, 2003.
- [16]A. Papoulis, S. U. Pillai. “Probability, Random Variables, and Stochastic Processes.” McGraw-Hill, 2002.
- [06]George Casella, Roger L. Berger. “Statistical Inference.” Duxbury, 2002.
- [17]R. A. Fisher. “Statistical Methods for Research Workers.” Oliver and Boyd, 1925.
- [07]E. L. Lehmann, Joseph P. Romano. “Testing Statistical Hypotheses.” Springer, 1998.
- [09]Trevor Hastie, Robert Tibshirani, Jerome Friedman. “The Elements of Statistical Learning: Data Mining, Inference, and Prediction.” Springer, 2009.
- [14]Donald E. Knuth. “The TeXbook.” Addison-Wesley, 1984.