Probability
Distributions
| | CONTINUOUS DISTRIBUTIONS : | |
| | DISCRETE DISTRIBUTIONS : | |
Linear Regression
| | SIMPLE LINEAR REGRESSION | |
| | MULTIPLE LINEAR REGRESSION | |
Estimation
| | | Confidence intervals | | | Confidence intervals for means of normal distributions | One sample confidence intervals. Two samples confidence intervals : paired samples, independent samples (variances known, unknown but equal, unknown and not equal). | | Approximate confidence intervals on means | Asymptotic interval (no demonstration). Welch's approximation. | | | | | Mean Square Error (MSE) | | | Mean Square Error (MSE) Minimum Mean Square Error (MMSE) estimators | MSE of a parameter estimator. Best estimate of a random variable X. Best estimate of X when a second r.v. Y is available. Properties of Minimum Mean Square Error estimators. | | | | | Sufficient statistic | | | First examples of sufficient statistics | Sufficient statistics for : * The Bernoulli distribution b(p), * The uniform distribution U[0, q], * The Poisson distribution P(l), from the definition only. | | The factorization theorem and applications | A necessary and sufficient condition for a statistic to be sufficient. Examples : Bernoulli, uniform, Poisson, normal (two methods), Gamma, exponential. | |
Tests
| | | ANOVA (One way) | | | Overview of ANOVA | General principle of ANOVA | | Variance decomposition | Total Sum of Squares, Factorial and Residual sums of Squares. A purely geometrical step. | | Distributions of the Sums of Squares | Sums of Squares as random variables. Distributions, independence. Properties as estimators of the common variance. | | ANOVA's F test | ANOVA is a F test. | | | | | Dunnett's test | | | Dunnett's test | Comparing group means to the mean of a reference group. | | | | | t-test | | | What are t-tests ? | Is a sample average trustworthy ? | | One-sample t-test | Is the sample mean significantly different from expected ? | | Student's t | Distribution of the mean when the variance is unknown | | "Two dependent samples" t-test | Are the means of 2 dependent samples equal ? | | "Two Independent samples" t-test | Are the means of 2 independent samples equal ? | | t-test results | How do I read software results of t-tests ? | | | | | Chi-square tests | | | The basic Chi-square test | Does a sample match a multinomial distribution ? | | Continuous reference distribution | Adapting the test to a continuous variable | | Estimated parameters | If some parameters of the reference distribution are unknown | | The Chi-square test of equality | Do several samples originate from the same distribution ? | | The Chi-square test of independence | Are two categorical variables independent ? | | Complements on the Chi-square of independence | Largest value, contributions, alternate coefficients. | | | | | The Fisher-Irwin test | | | The Fisher-Irwin test | Are these two coins identically biased ? | | | | | The Kolmogorov-Smirnov test | | | The Kolmogorov statistic | Its definition, distribution function, and the ensuing test. | | Complements on the Kolmogorov test | Very short on : K-test or Chi-2 test ? Estimated parameters. Normality test. | | | | | The Mann-Whitney test | | | The Mann-Whitney statistic | Its definition, distribution function, and the ensuing test. | | Complements on the Mann-Whitney test | Very short on : Why ranks ? Location-shift test. | | | | | Newman-Keuls test | | | The Newman-Keuls test | Pairwise comparisons of group means that avoid "paradoxical" conclusions. | |
Classification
| | | Fisher's linear discriminant | | | Fisher's criterion and Fisher's vector | Definition and justification of Fisher's criterion. Maximizing Fisher's criterion (2 classes). Fisher's discriminant. | | Maximizing the generalized Fisher's criterion | Maximizing the ratio of two quadratic forms. Maximizing the generalized Fisher's criterion. | | | | | Discriminant Analysis | | | What is Discriminant Analysis ? | The most basic classification technique. | | Discriminant Function Analysis | Finding new variables that are good at separating classes. | | Building a classifier | Creating linear or quadratic Classification Functions. | | Complements on DA | Just a little bit of maths. | | | | | Logistic Regression | | | What is Logistic Regression ? | LR is a powerful generalization of Discriminant Analysis. | | What is the "logit" ? | The information needed to build a score. | | Linear logit beyond DA | Getting rid of the normality assumption. | | Estimating the coefficients of the model | Likelihood, and how it is maximized. | | | | | Decision Trees | | | What are Decision Trees ? | Heuristic, yet powerful classifiers. Can do Regression too. | | Growing a Tree | Node splitting, Tree growth and Tree use. | | Three types of predictors | Handling categorical, ordinal and numerical predictors | | Splitting a node | Misclassification, Gini index, Entropy, Chi-square, Twoing | | Priors and costs | Weighting the observations to favorably bias the Tree. | | Stopping rules and Pruning | Getting the right size Tree to avoid overfitting | | | | |
Exploratory Data Analysis
| | | Principal Components Analysis (PCA) | | | What is PCA ? | An optimal way to display data on a plane, and more. | | What are Principal Components | The most efficient synthetic variables for representing data. | | Finding the Principal Components | Maximizing the inertia of projected observations. | | Projection of the observations | The best projection of data on a plane. | | Projection of the variables | Visualizing correlation between variables. | | Interpreting PCA results | Interpreting the Principal Components and data distribution. | | Other applications of PCA | Data Compression and Dimensionality reduction. | | | | | Correspondence Analysis | | | Overview of Correspondence Analysis | Visualizing the interaction of two categorical variables. | | Reformating data | Contengency tables, frequencies, profiles. | | The Chi-square distance | ...is more appropriate than euclidian distance. | | The two PCAs | How many dimensions, barycenters, total inertia. | | General principles of interpretation of CA | Factors, weights, inertias, plots, quality of representation. | | Complete treatment of a real case | A simple example from A to Z | | Complete treatment of a real case (1) | Interpreting the inertia, the Chi-square, the factors. | | Complete treatment of a real case (2) | Interpreting the plot of modalities for each variable. | | Complete treatment of a real case (3) | Interpreting the combined plot of modalities. | | Complements on CA | Supplementary variables, ordinal variables, Guttman effect. | |
No comments:
Post a Comment