|Unit level:||Level 6|
|Teaching period(s):||Semester 1|
|Offered by||School of Mathematics|
|Available as a free choice unit?:||N
Students are not permitted to take more than one of MATH38001 or MATH48001 for credit in the same or different undergraduate year. Students are not permitted to take MATH48001 and MATH68001 for credit in an undergraduate programme and then a postgraduate programme.
This course unit aims to introduce students to the principles of efficient estimation and hypothesis testing and acquaint them with the more successful methods of estimation and of constructing test procedures.
Statistical Inference is the body of principles and methods underlying the statistical analysis of data. In this course we introduce desirable properties that good estimators and hypothesis tests should enjoy and use them as criteria in the development of optimal estimators and test procedures. This is done both from the Classical/Frequentist as well as from the Bayesian point of view.
On successful completion of this module students will be able
- to determine how good an estimator or test procedure is on a number of criteria;
- to construct estimators and test procedures based both on the maximum likelihood principle and on Bayesian principles.
- Other - 20%
- Written exam - 80%
Assessment Further Information
- Coursework: weighting 20%
- End of semester examination: three hours, weighting 80%
- Estimation: point estimation, unbiasedness, mean square error, consistency, sufficiency, factorization theorem, Cramer-Rao inequality, the score function, Fisher information; efficiency: most efficient estimators, minimal sufficiency, Rao Blackwell theorem and its use in improving an estimator. 
- Methods of estimation: maximum likelihood estimators (m.l.e) and their asymptotic properties, asymptotic distribution of the score function. Confidence intervals based on the m.l.e and on the score function (multivariate case included). Restricted m.l.e and their asymptotic properties. 
- Hypothesis testing: Neyman-Pearson criteria, size and power function. Simple null vs simple alternative hypothesis and the Neyman-Pearson lemma. Hypothesis tests based on (i) m.l.e's; (ii) score function; (iii) the generalised likelihood ratio, profile log-likelihood and its use in interval estimation. The Deviance function and graphical methods in obtaining confidence regions for parameters. 
- Bayesian inference: introduction, priors, posteriors, conjugate prior, non-informative priors, Jeffrey's non informative prior, Bayesian estimation, predictive distributions, accuracy of an estimate, loss functions and expected posterior loss, optimal decisions with respect to a loss function, credibility intervals, highest posterior density credible intervals, hypothesis tests, large sample Bayesian approximation. 
- Beaumont, G. P., Intermediate Mathematical Statistics. Chapman & Hall 1980.
- Cox, D. R. and Hinkley, D. V., Theoretical Statistics. , Chapman & Hall 1974.
- Lindgren, B. W. Statistical Theory, 4th edition, Chapman & Hall 1993.
- Mood, A. M., Graybill, F. A. and Boes, D. C., Introduction to the Theory of Statistics, 3rd edition, McGraw-Hill 1974.
- Silvey, S. D., Statistical Inference, Chapman & Hall 1075.
Feedback tutorials will provide an opportunity for students' work to be discussed and provide feedback on their understanding. Coursework or in-class tests (where applicable) also provide an opportunity for students to receive feedback. Students can also get feedback on their understanding directly from the lecturer, for example during the lecturer's office hour.
- Lectures - 33 hours
- Tutorials - 11 hours
- Independent study hours - 106 hours