LATEST DISCOUNTS & SALES: PROMOTIONS

Close Notification

Your cart does not contain any items

Statistical Tools for the Comprehensive Practice of Industrial Hygiene and Environmental Health Sciences

David L. Johnson

$213.95

Hardback

Not in-store but you can order this
How long will it take?

QTY:

English
John Wiley & Sons Inc
06 January 2017
Reviews and reinforces concepts and techniques typical of a first statistics course with additional techniques useful to the IH/EHS practitioner.

Includes both parametric and non-parametric techniques described and illustrated in a worker health and environmental protection practice context Illustrated through numerous examples presented in the context of IH/EHS field practice and research, using the statistical analysis tools available in Excel® wherever possible Emphasizes the application of statistical tools to IH/EHS-type data in order to answer IH/EHS-relevant questions Includes an instructor’s manual that follows in parallel with the textbook, including PowerPoints to help prepare lectures and answers in the text as for the Exercises section of each chapter.

By:  
Imprint:   John Wiley & Sons Inc
Country of Publication:   United States
Dimensions:   Height: 257mm,  Width: 183mm,  Spine: 25mm
Weight:   839g
ISBN:   9781119143017
ISBN 10:   1119143012
Pages:   400
Publication Date:  
Audience:   Professional and scholarly ,  Undergraduate
Format:   Hardback
Publisher's Status:   Active
Preface xv Acknowledgments xvii About the Author xix About the Companion Website xxi 1 Some Basic Concepts 1 1.1 Introduction 1 1.2 Physical versus Statistical Sampling 2 1.3 Representative Measures 3 1.4 Strategies for Representative Sampling 3 1.5 Measurement Precision 4 1.6 Probability Concepts 6 1.6.1 The Relative Frequency Approach 7 1.6.2 The Classical Approach – Probability Based on Deductive Reasoning 7 1.6.3 Subjective Probability 7 1.6.4 Complement of a Probability 7 1.6.5 Mutually Exclusive Events 8 1.6.6 Independent Events 8 1.6.7 Events that Are Not Mutually Exclusive 9 1.6.8 Marginal and Conditional Probabilities 9 1.6.9 Testing for Independence 11 1.7 Permutations and Combinations 12 1.7.1 Permutations for Sampling without Replacement 12 1.7.2 Permutations for Sampling with Replacement 13 1.7.3 Combinations 13 1.8 Introduction to Frequency Distributions 14 1.8.1 The Binomial Distribution 14 1.8.2 The Normal Distribution 16 1.8.3 The Chi-Square Distribution 20 1.9 Confidence Intervals and Hypothesis Testing 22 1.10 Summary 23 1.11 Addendum: Glossary of Some Useful Excel Functions 23 1.12 Exercises 26 References 28 2 Descriptive Statistics and Methods of Presenting Data 29 2.1 Introduction 29 2.2 Quantitative Descriptors of Data and Data Distributions 29 2.3 Displaying Data with Frequency Tables 33 2.4 Displaying Data with Histograms and Frequency Polygons 34 2.5 Displaying Data Frequency Distributions with Cumulative Probability Plots 35 2.6 Displaying Data with NED and Q–Q Plots 38 2.7 Displaying Data with Box-and-Whisker Plots 41 2.8 Data Transformations to Achieve Normality 42 2.9 Identifying Outliers 43 2.10 What to Do with Censored Values? 45 2.11 Summary 45 2.12 Exercises 46 References 48 3 Analysis of Frequency Data 49 3.1 Introduction 49 3.2 Tests for Association and Goodness-of-Fit 50 3.2.1 r × c Contingency Tables and the Chi-Square Test 50 3.2.2 Fisher’s Exact Test 54 3.3 Binomial Proportions 55 3.4 Rare Events and the Poisson Distribution 57 3.4.1 Poisson Probabilities 57 3.4.2 Confidence Interval on a Poisson Count 60 3.4.3 Testing for Fit with the Poisson Distribution 61 3.4.4 Comparing Two Poisson Rates 62 3.4.5 Type I Error, Type II Error, and Power 64 3.4.6 Power and Sample Size in Comparing Two Poisson Rates 64 3.5 Summary 65 3.6 Exercises 66 References 69 4 Comparing Two Conditions 71 4.1 Introduction 71 4.2 Standard Error of the Mean 71 4.3 Confidence Interval on a Mean 72 4.4 The t-Distribution 73 4.5 Parametric One-Sample Test – Student’s t-Test 74 4.6 Two-Tailed versus One-Tailed Hypothesis Tests 76 4.7 Confidence Interval on a Variance 77 4.8 Other Applications of the Confidence Interval Concept in IH/EHS Work 79 4.8.1 OSHA Compliance Determinations 79 4.8.2 Laboratory Analyses – LOB, LOD, and LOQ 80 4.9 Precision, Power, and Sample Size for One Mean 81 4.9.1 Sample Size Required to Estimate a Mean with a Stated Precision 81 4.9.2 Sample Size Required to Detect a Specified Difference in Student’s t-Test 81 4.10 Iterative Solutions Using the Excel Goal Seek Utility 82 4.11 Parametric Two-Sample Tests 83 4.11.1 Confidence Interval for a Difference in Means: The Two-Sample t-Test 83 4.11.2 Two-Sample t-Test When Variances Are Equal 84 4.11.3 Verifying the Assumptions of the Two-Sample t-Test 85 4.11.3.1 Lilliefors Test for Normality 86 4.11.3.2 Shapiro–Wilk W-Test for Normality 87 4.11.3.3 Testing for Homogeneity of Variance 91 4.11.3.4 Transformations to Stabilize Variance 93 4.11.4 Two-Sample t-Test with Unequal Variances – Welch’s Test 93 4.11.5 Paired Sample t-Test 95 4.11.6 Precision, Power, and Sample Size for Comparing Two Means 96 4.12 Testing for Difference in Two Binomial Proportions 99 4.12.1 Testing a Binomial Proportion for Difference from a Known Value 100 4.12.2 Testing Two Binomial Proportions for Difference 100 4.13 Nonparametric Two-Sample Tests 102 4.13.1 Mann–Whitney U Test 102 4.13.2 Wilcoxon Matched Pairs Test 104 4.13.3 McNemar and Binomial Tests for Paired Nominal Data 105 4.14 Summary 107 4.15 Exercises 107 References 111 5 Characterizing the Upper Tail of the Exposure Distribution 113 5.1 Introduction 113 5.2 Upper Tolerance Limits 113 5.3 Exceedance Fractions 115 5.4 Distribution Free Tolerance Limits 117 5.5 Summary 119 5.6 Exercises 119 References 121 6 One-Way Analysis of Variance 123 6.1 Introduction 123 6.2 Parametric One-Way ANOVA 123 6.2.1 How the Parametric ANOVA Works – Sums of Squares and the F-Test 124 6.2.2 Post hoc Multiple Pairwise Comparisons in Parametric ANOVA 127 6.2.2.1 Tukey’s Test 127 6.2.2.2 Tukey–Kramer Test 128 6.2.2.3 Dunnett’s Test for Comparing Means to a Control Mean 130 6.2.2.4 Planned Contrasts Using the Scheffé S Test 132 6.2.3 Checking the ANOVA Model Assumptions – NED Plots and Variance Tests 134 6.2.3.1 Levene’s Test 134 6.2.3.2 Bartlett’s Test 135 6.3 Nonparametric Analysis of Variance 136 6.3.1 Kruskal–Wallis Nonparametric One-Way ANOVA 137 6.3.2 Post hoc Multiple Pairwise Comparisons in Nonparametric ANOVA 139 6.3.2.1 Nemenyi’s Test 139 6.3.2.2 Bonferroni–Dunn Test 140 6.4 ANOVA Disconnects 142 6.5 Summary 144 6.6 Exercises 145 References 149 7 Two-Way Analysis of Variance 151 7.1 Introduction 151 7.2 Parametric Two-Way ANOVA 151 7.2.1 Two-Way ANOVA without Interaction 154 7.2.2 Checking for Homogeneity of Variance 154 7.2.3 Multiple Pairwise Comparisons When There Is No Interaction Term 154 7.2.4 Two-Way ANOVA with Interaction 156 7.2.5 Multiple Pairwise Comparisons with Interaction 158 7.2.6 Two-Way ANOVA without Replication 160 7.2.7 Repeated-Measures ANOVA 160 7.2.8 Two-Way ANOVA with Unequal Sample Sizes 162 7.3 Nonparametric Two-Way ANOVA 162 7.3.1 Rank Tests 162 7.3.1.1 The Rank Test 162 7.3.1.2 The Rank Transform Test 166 7.3.1.3 Other Options – Aligned Rank Tests 166 7.3.2 Repeated-Measures Nonparametric ANOVA – Friedman’s Test 166 7.3.2.1 Friedman’s Test without Replication 167 7.3.2.2 Multiple Comparisons for Friedman’s Test without Replication 169 7.3.2.3 Friedman’s Test with Replication 170 7.3.2.4 Multiple Comparisons for Friedman’s Test with Replication 172 7.4 More Powerful Non-ANOVA Approaches: Linear Modeling 172 7.5 Summary 172 7.6 Exercises 172 References 178 8 Correlation Analysis 181 8.1 Introduction 181 8.2 Simple Parametric Correlation Analysis 181 8.2.1 Testing the Correlation Coefficient for Significance 184 8.2.1.1 t-Test for Significance 185 8.2.1.2 F-Test for Significance 186 8.2.2 Confidence Limits on the Correlation Coefficient 186 8.2.3 Power in Simple Correlation Analysis 187 8.2.4 Comparing Two Correlation Coefficients for Difference 188 8.2.5 Comparing More Than Two Correlation Coefficients for Difference 189 8.2.6 Multiple Pairwise Comparisons of Correlation Coefficients 190 8.3 Simple Nonparametric Correlation Analysis 190 8.3.1 Spearman Rank Correlation Coefficient 190 8.3.2 Testing Spearman’s Rank Correlation Coefficient for Statistical Significance 191 8.3.3 Correction to Spearman’s Rank Correlation Coefficient When There Are Tied Ranks 193 8.4 Multiple Correlation Analysis 195 8.4.1 Parametric Multiple Correlation 195 8.4.2 Nonparametric Multiple Correlation: Kendall’s Coefficient of Concordance 195 8.5 Determining Causation 198 8.6 Summary 198 8.7 Exercises 198 References 204 9 Regression Analysis 205 9.1 Introduction 205 9.2 Linear Regression 205 9.2.1 Simple Linear Regression 207 9.2.2 Nonconstant Variance – Transformations and Weighted Least Squares Regression 209 9.2.3 Multiple Linear Regression 213 9.2.3.1 Multiple Regression in Excel 215 9.2.3.2 Multiple Regression Using the Excel Solver Utility 218 9.2.3.3 Multiple Regression Using Advanced Software Packages 221 9.2.4 Using Regression for Factorial ANOVA with Unequal Sample Sizes 222 9.2.5 Multiple Correlation Analysis Using Multiple Regression 227 9.2.5.1 Assumptions of Parametric Multiple Correlation 233 9.2.5.2 Options When Collinearity Is a Problem 233 9.2.6 Polynomial Regression 234 9.2.7 Interpreting Linear Regression Results 234 9.2.8 Linear Regression versus ANOVA 235 9.3 Logistic Regression 235 9.3.1 Odds and Odds Ratios 236 9.3.2 The Logit Transformation 238 9.3.3 The Likelihood Function 240 9.3.4 Logistic Regression in Excel 240 9.3.5 Likelihood Ratio Test for Significance of MLE Coefficients 241 9.3.6 Odds Ratio Confidence Limits in Multivariate Models 243 9.4 Poisson Regression 243 9.4.1 Poisson Regression Model 243 9.4.2 Poisson Regression in Excel 244 9.5 Regression with Excel Add-ons 245 9.6 Summary 246 9.7 Exercises 246 References 252 10 Analysis of Covariance 253 10.1 Introduction 253 10.2 The Simple ANCOVA Model and Its Assumptions 253 10.2.1 Required Regressions 255 10.2.2 Checking the ANCOVA Assumptions 258 10.2.2.1 Linearity, Independence, and Normality 258 10.2.2.2 Similar Variances 258 10.2.2.3 Equal Regression Slopes 258 10.2.3 Testing and Estimating the Treatment Effects 259 10.3 The Two-Factor Covariance Model 261 10.4 Summary 261 10.5 Exercises 261 Reference 263 11 Experimental Design 265 11.1 Introduction 265 11.2 Randomization 266 11.3 Simple Randomized Experiments 266 11.4 Experimental Designs Blocking on Categorical Factors 267 11.5 Randomized Full Factorial Experimental Design 270 11.6 Randomized Full Factorial Design with Blocking 271 11.7 Split Plot Experimental Designs 272 11.8 Balanced Experimental Designs – Latin Square 273 11.9 Two-Level Factorial Experimental Designs with Quantitative Factors 274 11.9.1 Two-Level Factorial Designs for Exploratory Studies 274 11.9.2 The Standard Order 275 11.9.3 Calculating Main Effects 276 11.9.4 Calculating Interactions 278 11.9.5 Estimating Standard Errors 278 11.9.6 Estimating Effects with REGRESSION in Excel 279 11.9.7 Interpretation 280 11.9.8 Cube, Surface, and NED Plots as an Aid to Interpretation 280 11.9.9 Fractional Factorial Two-Level Experiments 282 11.10 Summary 282 11.11 Exercises 283 References 284 12 Uncertainty and Sensitivity Analysis 285 12.1 Introduction 285 12.2 Simulation Modeling 285 12.2.1 Propagation of Errors 286 12.2.2 Simple Bounding 287 12.2.2.1 Sums and Differences 287 12.2.2.2 Products and Ratios 287 12.2.2.3 Powers 289 12.2.3 Addition in Quadrature 289 12.2.3.1 Sums and Differences 289 12.2.3.2 Products and Ratios 290 12.2.3.3 Powers 292 12.2.4 LOD and LOQ Revisited – Dust Sample Gravimetric Analysis 292 12.3 Uncertainty Analysis 295 12.4 Sensitivity Analysis 296 12.4.1 One-at-a-Time (OAT) Analysis 296 12.4.2 Variance-Based Analysis 297 12.5 Further Reading on Uncertainty and Sensitivity Analysis 297 12.6 Monte Carlo Simulation 297 12.7 Monte Carlo Simulation in Excel 298 12.7.1 Generating Random Numbers in Excel 298 12.7.2 The Populated Spreadsheet Approach 299 12.7.3 Monte Carlo Simulation Using VBA Macros 299 12.8 Summary 303 12.9 Exercises 303 References 307 13 Bayes’ Theorem and Bayesian Decision Analysis 309 13.1 Introduction 309 13.2 Bayes’ Theorem 310 13.3 Sensitivity, Specificity, and Positive and Negative Predictive Value in Screening Tests 310 13.4 Bayesian Decision Analysis in Exposure Control Banding 312 13.4.1 Introduction to BDA 312 13.4.2 The Prior Distribution and the Parameter Space 314 13.4.3 The Posterior Distribution and Likelihood Function 314 13.4.4 Relative Influences of the Prior and the Data 315 13.4.5 Frequentist versus Bayesian Perspectives 316 13.5 Exercises 316 References 318 A z-Tables of the Standard Normal Distribution 321 B Critical Values of the Chi-Square Distribution 327 C Critical Values for the t-Distribution 329 D Critical Values for Lilliefors Test 331 Reference 332 E Shapiro–Wilk W Test 𝜶 Coefficients and Critical Values 333 Reference 336 F Critical Values of the F Distribution for 𝜶 = 0.05 337 G Critical U Values for the Mann–Whitney U Test 341 Reference 342 H Critical Wilcoxon Matched Pairs Test t Values 343 Reference 344 I K Values for Upper Tolerance Limits 345 Reference 346 J Exceedance Fraction 95% Lower Confidence Limit versus Z 347 Reference 347 K q Values for Tukey’s, Tukey–Kramer, and Nemenyi’s MSD Tests 349 L q′ Values for Dunnett’s Test 351 Reference 353 M Q Values for the Bonferroni–Dunn MSD Test 355 N Critical Spearman Rank Correlation Test Values 357 O Critical Values of Kendall’s W 359 Reference 361 Index 363

David L. Johnson has over 40 years of experience in environmental engineering and occupational safety and health practice, research, and teaching. Dr. Johnson was a practicing environmental engineer and industrial hygienist with the United States Army for 20 years, serving in a variety of positions in the United States, Europe, and the Middle East. He joined the faculty of the University of Oklahoma’s College of Public Health, Department of Occupational and Environmental Health in 1991.

See Also