Skip to Content
University of Brighton

Library Catalogue

An adventure in statistics: the reality enigma

An adventure in statistics: the reality enigma

Field, Andy P., author

Andy Field's approach gently introduces students across the social sciences to the importance and relevance of statistics. By weaving in a compelling narrative, he takes students on an exciting journey through introductory level statistics overcoming potential anxiety around the subject and providing a vibrant alternative to the dullness of many typical offerings

Book. English.
Published Los Angeles: SAGE, 2016
Rate this

Available on the shelf at Falmer and Hastings.

  • Falmer – One available in 519.5/FIE

    Barcode Shelfmark Loan type Status
    08761574 519.5/FIE Long loan Available
    08761582 519.5/FIE Long loan Due back 26th May
    08761566 519.5/FIE Long loan Due back 4th June
  • Hastings – One available in 519.5/FIE

    Barcode Shelfmark Loan type Status
    08146489 519.5/FIE Long loan Available


Statement of responsibility: Andy Field
ISBN: 1446210456, 9781446210451
Intended audience: Specialized.
Note: Includes bibliographical references and index.
Physical Description: xvi, 746 pages : illustrations (colour) ; 25 cm
Subject: Mathematics.; Statistics.


  1. Prologue
  2. The Dying Stars1 Why You Need Science
  3. The Beginning and The End 1.1. Will you love me now? 1.2. How science works 1.2.1. The research process 1.2.2. Science as a life skill 1.3. Research methods 1.3.1. Correlational research methods 1.3.2. Experimental research methods 1.3.3. Practice, order and randomization 1.4. Why we need science2 Reporting Research, Variables and Measurement
  4. Breaking the Law 2.1. Writing up research 2.2. Maths and statistical notation 2.3. Variables and measurement 2.3.1. The conspiracy unfolds 2.3.2. Qualitative and quantitative data 2.3.3. Levels of measurement 2.3.4. Measurement error 2.3.5. Validity and reliability3 Summarizing Data
  5. She Loves Me Not? 3.1. Frequency distributions 3.1.1. Tabulated frequency distributions 3.1.2. Grouped frequency distributions 3.1.3. Graphical frequency distributions 3.1.4. Idealized distributions 3.1.5. Histograms for nominal and ordinal data 3.2. Throwing Shapes4 Fitting Models (Central Tendency)
  6. Somewhere In The Middle 4.1. Statistical Models 4.1.1. From the dead 4.1.2. Why do we need statistical models? 4.1.3. Sample size 4.1.4. The one and only statistical model 4.2. Central Tendency 4.2.1. The mode 4.2.2. The median 4.2.3. The mean 4.3. The 'fit' of the mean: variance 4.3.1. The fit of the mean 4.3.2. Estimating the fit of the mean from a sample 4.3.3. Outliers and variance 4..4. Dispersion 4.4.1. The standard deviation as an indication of dispersion 4.4.2. The range and interquartile range5 Presenting Data
  7. Aggressive Perfector 5.1. Types of graphs 5.2. Another perfect day 5.3. The art of presenting data 5.3.1. What makes a good graph? 5.3.2. Bar graphs 5.3.3. Line graphs 5.3.4. Boxplots (box-whisker diagrams) 5.3.5. Graphing relationships: the scatterplot 5.3.6. Pie charts6 Z-Scores
  8. The wolf is loose 6.1. Interpreting raw scores 6.2. Standardizing a score 6.3. Using z-scores to compare distributions 6.4. Using z-scores to compare scores 6.5. Z-scores for samples7 Probability
  9. The Bridge of Death 7.1. Probability 7.1.1. Classical probability 7.1.2. Empirical probability 7.2. Probability and frequency distributions 7.2.1. The discs of death 7.2.2. Probability density functions 7.2.3. Probability and the normal distribution 7.2.4. The probability of a score greater than x 7.2.5. The probability of a score less than x
  10. The tunnels of death 7.2.6. The probability of a score between two values
  11. The catapults of death 7.3. Conditional probability
  12. DeathscotchInferential Statistics
  13. Going Beyond the Data 8.1. Estimating parameters 8.2. How well does a sample represent the population? 8.2.1. Sampling distributions 8.2.2. The standard error 8.2.3. The central limit theorem 8.3. Confidence Intervals 8.3.1. Calculating confidence intervals 8.3.2. Calculating other confidence intervals 8.3.3. Confidence intervals in small samples 8.4. Inferential statistics9 Robust Estimation
  14. Man Without Faith or Trust 9.1. Sources of bias 9.1.1. Extreme scores and non-normal distributions 9.1.2. The mixed normal distribution 9.2. A great mistake 9.3. Reducing bias 9.3.1. Transforming data 9.3.2. Trimming data 9.3.3. M-estimators 9.3.4. Winsorizing 9.3.5. The bootstrap 9.4. A final point about extreme scores10 Hypothesis Testing
  15. In Reality All is Void 10.1. Null hypothesis significance testing 10.1.1. Types of hypothesis 10.1.2. Fisher's p-value 10.1.3. The principles of NHST 10.1.4. Test statistics 10.1.5. One- and two-tailed tests 10.1.6. Type I and Type II errors 10.1.7. Inflated error rates 10.1.8. Statistical power 10.1.9. Confidence intervals and statistical significance 10.1.10. Sample size and statistical significance11 Modern Approaches to Theory Testing
  16. A Careworn Heart 11.1. Problems with NHST 11.1.1. What can you conclude from a 'significance' test? 11.1.2. All-or-nothing thinking 11.1.3. NHST is influenced by the intentions of the scientist 11.2. Effect sizes 11.2.1. Cohen's d 11.2.2. Pearson's correlation coefficient,r 11.2.3. The odds ratio 11.3. Meta-analysis 11.4. Bayesian approaches 11.4.1. Asking a different question 11.4.2. Bayes' theorem revisited 11.4.3. Comparing hypothesis 11.4.4. Benefits of bayesian approaches12 Assumptions
  17. Starblind 12.1. Fitting models: bringing it all together 12.2. Assumptions 12.2.1. Additivity and linearity 12.2.2. Independent errors 12.2.3. Homoscedasticity/ homogeneity of variance 12.2.4. Normally distributed something or other 12.2.5. External variables 12.2.6. Variable types 12.2.7. Multicollinearity 12.2.8. Non-zero variance 12.3. Turning ever towards the sun13 Relationships
  18. A Stranger's Grave 13.1. Finding relationships in categorical data 13.1.1. Pearson's chi-square test 13.1.2. Assumptions 13.1.3. Fisher's exact test 13.1.4. Yates's correction 13.1.5. The likelihood ratio (G-test) 13.1.6. Standardized residuals 13.1.7. Calculating an effect size 13.1.8. Using a computer 13.1.9. Bayes factors for contingency tables 13.1.10. Summary 13.2. What evil lay dormant 13.3. Modelling relationships 13.3.1. Covariance 13.3.2. Pearson's correlation coefficient 13.3.3. The significance of the correlation coefficient 13.3.4. Confidence intervals for r 13.3.5. Using a computer 13.3.6. Robust estimation of the correlation 13.3.7. Bayesian approaches to relationships between two variables 13.3.8. Correlation and causation 13.3.9. Calculating the effect size 13.4. Silent sorrow in empty boats14 The General Linear Model
  19. Red Fire Coming Out From His Gills 14.1. The linear model with one predictor 14.1.1. Estimating parameters 14.1.2. Interpreting regression coefficients 14.1.3. Standardized regression coefficients 14.1.4. The standard error of b 14.1.5. Confidence intervals for b 14.1.6. Test statistic for b 14.1.7. Assessing the goodness of fit 14.1.8. Fitting a linear model using a computer 14.1.9. When this fails 14.2. Bias in the linear model 14.3. A general procedure for fitting linear models 14.4. Models with several predictors 14.4.1. The expanded linear model 14.4.2. Methods for entering predictors 14.4.3. Estimating parameters 14.4.4. Using a computer to build more complex models 14.5. Robust regression 14.5.1. Bayes factors for linear models15 Comparing Two Means
  20. Rock or Bust 15.1. Testing differences between means
  21. The rationale 15.2. Means and the linear model 15.2.1. Estimating the model parameters 15.2.2. How the model works 15.2.3. Testing the model parameters 15.2.4. The independent t-test on a computer 15.2.5. Assumptions of the model 15.3. Everything you believe is wrong 15.4. The paired-samples t-test 15.4.1. The paired-samples t-test on a computer 15.5. Alternative approaches 15.5.1. Effect sizes 15.5.2. Robust tests of two means 15.5.3. Bayes factors for comparing two means16 Comparing Several Means
  22. Faith in Others 16.1. General procedure for comparing means 16.2. Comparing several means with the linear model 16.2.1. Dummy coding 16.2.2. The F-ratio as a test of means 16.2.3. The total sum of squares (SSt) 16.2.4. The model sum of squares (SSm) 16.2.5. The residual sum of squares (SSr) 16.2.6. Partitioning variance 16.2.7. Mean squares 16.2.8. The F-ratio 16.2.9. Comparing several means using a computer 16.3. Contrast coding 16.3.1. Generating contrasts 16.3.2. Devising weights 16.3.3. Contrasts and the linear model 16.3.4. Post hoc procedures 16.3.5. Contrasts and post hoc tests using a computer 16.4. Storm of memories 16.5. Repeated-measures designs 16.5.1. The total sum of squares, SSt 16.5.2. The within-participant variance, SSw 16.5.3. The model sum of squares, SSm 16.5.4. The residual sum of squares, SSr 16.5.5. Mean squares and the F-ratio 16.5.6. Repeated-measures designs using a computer 16.6. Alternative approaches 16.6.1. Effect sizes 16.6.2. Robust tests of several means 16.6.3. Bayesian analysis of several means 16.7. The invisible manFactorial Designs 17.1. Factorial designs 17.2. General procedure and assumptions 17.3. Analysing factorial designs 17.3.1. Factorial designs and the linear model 17.3.2. The fit of the model 17.3.3. Factorial designs on a computer 17.4. From the pinnacle to the pit 17.5. Alternative approaches 17.5.1. Calculating effect sizes 17.5.2. Robust analysis of factorial designs 17.5.3. Bayes factors for factorial designs 17.6. Interpreting interaction effectsEpilogue
  23. The Genial Night
  24. SI Momentum Requiris, Circumspice


See what students are saying!

“PERFECT FOR EVERYONE that finds stats difficult, pointless or boring, this book proves them wrong! The way the concepts are explained is so easy to grasp and it makes it even entertaining to learn! I wish I had discovered this sooner!”

||“A MUST-HAVE FOR ANY STATISTICS STUDENT looking for a thorough understanding of statistical terminology and concepts. Students will come for the statistics and stay for the narrative.”||“A UNIQUE, ENGAGING EXPERIENCE…The narrative helps to build a context for introducing concepts and allows for the characters to explain the concepts through their successive parts. This is a must-have for those studying statistics.”||“Funny and engaging book, YOU CANNOT STOP READING it because you cannot wait to know what happens to Zach and Alice.”||“This book is EXTREMELY HELPFUL IN UNDERSTANDING THE BASICS behind some of the more complex statistical procedures gone over in lectures.”