Sufficient of Statistics
Statistics is a problem-solving procedure:
flowchart LR subgraph AA["Data"] direction TB A1@{shape: braces, label: "distribution<br>dependence<br>..."} end subgraph BB["Method"] direction TB B1@{shape: braces, label: "algorithms<br>statistics<br>..."} end subgraph CC["Problem"] direction TB C1@{shape: braces, label: "output<br>metric<br>..."} end AA --- BB --- CC
Map of Contents
Basic Concepts
- Setup
- Estimation
- Point Estimation
- Interval Estimation: Confidence Interval
- Distributional Estimation: Bayesian Inference
- Hypothesis Testing
- Regression
- Prediction
Advanced Topics
- Advanced concepts
- Inference for CDFs
- Expectation Maximization
- More on hypothesis testing
- Hypothesis testing and other statistical procedures
Interactive Notebooks
- Tail Bounds and Concentration Inequalities
- Constructing Confidence Interval
- Best Estimator for Uniform Distribution Parameter
Statistics and Machine Learning
Statistics | Machine Learning |
---|---|
Estimation | learning |
confidence | uncertainty |
Regression | Supervised Learning |
Clustering | Unsupervised Learning |
hypothesis | Classification |
covariates | features |
coefficient | weight |
predictor | input |
response | output |
intercept | bias |
derived predictor | Hidden Units |
penalty function | weight decay |
Probability and Statistics
graph LR A(Data generating process) --Probability--> B(Observed data) B --Statistical inference--> A
Probability Theory answers the following question:
Given a data generating process, what are the properties of the outcomes?
While statistical inference answers the inverse problem:
Given the outcomes, what can we say about the process that generated the data?
References
- Textbooks
- Larry Wasserman, All of Statistics
- van der Vaart, Asymptotic Statistics
- Sheldon Ross, Introduction to Probability and Statistics for Engineers and Scientists
- Courses
- Columbia STAT 5701, 5703
- MIT 6.S951, 18.6501