LATEST SALES & OFFERS: PROMOTIONS

Close Notification

Your cart does not contain any items

A Modern Introduction to Probability and Statistics

Understanding Statistical Principles in the Age of the Computer

Graham Upton

$287.95

Hardback

Forthcoming
Pre-Order now

QTY:

English
Oxford University Press
01 July 2025
Probability and statistics are subjects fundamental to data analysis, making them essential for efficient artificial intelligence. Although the foundational concepts of probability and statistics remain constant, what needs to be taught is constantly evolving.

The first half of the book introduces probability, conditional probability and the standard probability distributions in the traditional way. The second half considers the power of the modern computer and our reliance on technology to do the calculations for us. Offering a fresh presentation that builds on the author's previous book, Understanding Statistics, this book includes exercises (with solutions at the rear of the book) and worked examples. Chapters close with a brief mention of the relevant R commands and summary of the content. Increasingly difficult mathematical sections are clearly indicated, and these can be omitted without affecting the understanding of the remaining material.

Aimed at first year graduates, this book is also suitable for readers familiar with mathematical notation.
By:  
Imprint:   Oxford University Press
Country of Publication:   United Kingdom
Dimensions:   Height: 246mm,  Width: 189mm, 
ISBN:   9780198943129
ISBN 10:   0198943121
Pages:   384
Publication Date:  
Audience:   College/higher education ,  Primary
Format:   Hardback
Publisher's Status:   Forthcoming
1 Probability 1: Probability 1.1 Relative Frequency 1.2 Preliminary definitions 1.3 The probability scale 1.4 Probability with equally likely outcomes 1.5 The complementary event E' 1.6 Venn diagrams 1.7 Unions and intersections of events 1.8 Mutually exclusive events 1.9 Exhaustive events 1.10 Probability trees 1.11 Sample proportions and probability 1.12 Unequally likely possibilities 1.13 Physical independence 1.14 Orderings 1.15 Permutations and combinations 1.16 Sampling without replacement 1.17 Sampling without replacement 2: Conditional Probability 2.1 Notation 2.2 Statistical independence 2.3 Mutual and pairwise independence 2.4 The total probability theorem (The partition theorem) 2.5 Bayes>' theorem 2.6 *The Monty Hall problem 3: Probability distributions 3.1 Notation 3.2 Probability distributions 3.3 The discrete uniform distribution 3.4 The Bernoulli distribution 3.5 The Binomial Distribution 3.6 Notation 3.7 's inequality 9.2 Chebyshev>'s inequality 9.3 The weak law of large numbers 9.4 The strong law of large numbers 10: Joint Distributions 10.1 Joint probability mass function 10.2 Marginal distributions 10.3 Conditional distributions 10.4 2 Statistics 11: Data sources 11.1 Data collection by observation 11.2 National Consuses 11.3 Sampling 11.4 Questionnaires 11.5 Questionnaire Design 12: Summarising data 12.1 A single variable 12.2 Two variables 12.3 More than two variables 12.4 Choosing which display to use 12.5 Dirty Data 13: General Summary Statistics 13.1 Measure of location: The mode 13.2 Measure of location: The mean 13.3 Measure of location: The mean of a frequency distribution 13.4 Measure of location: The mean of grouped data 13.5 Simplifying calculations 13.6 Measure of location: The median 13.7 Quantiles 13.8 Measures of spread: The Range and Inter-quartile Range 13.9 Boxplot 13.10 Deviations from the mean 13.11 The mean deviation 13.12 Measure of spread: The variance 13.13 Calculating the variance by hand 13.14 Measure of spread: The standard deviation 13.15 Variance and standard deviation for frequency distributions 13.16 Symmetric and skewed data 13.17 Standardising to a prescribed mean and standard deviation 13.8 *Calculating the combined mean and variance of several samples 13.19 Combining proportions 14: Point and interval estimation 14.1 Point estimates 14.2 Estimation methods 14.3 Confidence intervals 14.4 Confidence intervals with discrete distributions 14.5 One-sided confidence intervals 14.6 Confidence intervals for a variance 15: Single-sampled hypothesis tests 15.1 The null and alternative hypothesis 15.2 Critical regions and significance levels 15.3 The test procedure 15.4 Identifying two hypotheses 15.5 Tail probabilities: the p value approach 15.6 Hypothesis tests and confidence intervals 15.7 Hypothesis tests for a mean 15.8 Testing for normality 15.9 Hypothesis test for the variance of a normal distribution 5.10 Hypothesis tests with discrete distributions 5.11 Type I and Type II errors 5.12 Hypothesis tests for a proportion based on a small sample 15.13 Hypothesis tests for a Poisson mean based on a small sample 16: Two samples & paired samples 16.1 The comparison of two means 16.2 Confidence interval for the difference between two normal means 16.3 Paired samples 16.4 The comparison of the variances of two normal distributions 16.5 Confidence interval for a variance ratio 17: Goodness of fit 17.1 The chi-squared test 17.2 Small expected frequencies 17.3 Goodness of fit to prescribed distribution type 17.4 Comparing distribution functions 17.5 The dispersion test 17.6 Contingency tables 17.7 The 2 x 2 table: the comparison of two proportions 17.8 *Multi-way contingency tables 18: Correlation 18.1 The product-moment correlation coefficent 18.2 Nonsense correlation: storks and goosebury bushes 18.3 The ecological fallacy: Immigration and illiteracy 18.4 Simpson's paradox: Amputation or injection? 18.5 Rank correlation 19: Regression 19.1 The equation of a straight line 19.2 Why 'regression'? 19.3 The method of least squares 19.4 Transformations, extrapolation and outliers 19.5 Properties of the estimators 19.6 Analysis of Variance (ANOVA) 19.7 Multiple Regression 20 *The Bayesian approach

Graham Upton is a retired Professor of Applied Statistics, formerly of the Department of Mathematical sciences at The University of Essex. He has published numerous books, including The Oxford Dictionary of Statistics, Data Analysis: A Gentle Introduction for Future Data Scientists, and Understanding Statistics with OUP.

See Also