‏757.00 ₪

Applied Univariate, Bivariate, and Multivariate Statistics

‏757.00 ₪
ISBN13
9781118632338
יצא לאור ב
New York
זמן אספקה
21 ימי עסקים
עמודים
760
פורמט
Hardback
תאריך יציאה לאור
5 בינו׳ 2016
A clear and efficient balance between theory and application of statistical modeling techniques in the social and behavioral sciences Written as a general and accessible introduction, Applied Univariate, Bivariate, and Multivariate Statistics provides an overview of statistical modeling techniques used in fields in the social and behavioral sciences. Blending statistical theory and methodology, the book surveys both the technical and theoretical aspects of good data analysis. Featuring applied resources at various levels, the book includes statistical techniques such as t-tests and correlation as well as more advanced procedures such as MANOVA, factor analysis, and structural equation modeling. To promote a more in-depth interpretation of statistical techniques across the sciences, the book surveys some of the technical arguments underlying formulas and equations. Applied Univariate, Bivariate, and Multivariate Statistics also features * Demonstrations of statistical techniques using software packages such as R and SPSS(R) * Examples of hypothetical and real data with subsequent statistical analyses * Historical and philosophical insights into many of the techniques used in modern social science * A companion website that includes further instructional details, additional data sets, solutions to selected exercises, and multiple programming options An ideal textbook for courses in statistics and methodology at the upper- undergraduate and graduate-levels in psychology, political science, biology, sociology, education, economics, communications, law, and survey research, Applied Univariate, Bivariate, and Multivariate Statistics is also a useful reference for practitioners and researchers in their field of application. DANIEL J. DENIS, PhD, is Associate Professor of Quantitative Psychology at the University of Montana where he teaches courses in univariate and multivariate statistics. He has published a number of articles in peer-reviewed journals and has served as consultant to researchers and practitioners in a variety of fields.
מידע נוסף
עמודים 760
פורמט Hardback
ISBN10 1118632338
יצא לאור ב New York
תאריך יציאה לאור 5 בינו׳ 2016
תוכן עניינים Preface xix About the Companion Website xxxiii 1 Preliminary Considerations 1 1.1 The Philosophical Bases of Knowledge: Rationalistic versus Empiricist Pursuits 1 1.2 What is a Model ? 4 1.3 Social Sciences versus Hard Sciences 6 1.4 Is Complexity a Good Depiction of Reality? Are Multivariate Methods Useful? 8 1.5 Causality 9 1.6 The Nature of Mathematics: Mathematics as a Representation of Concepts 10 1.7 As a Social Scientist How Much Mathematics Do You Need to Know? 11 1.8 Statistics and Relativity 12 1.9 Experimental versus Statistical Control 13 1.10 Statistical versus Physical Effects 14 1.11 Understanding What Applied Statistics Means 15 Review Exercises 15 2 Mathematics and Probability Theory 18 2.1 Set Theory 20 2.2 Cartesian Product A x B 24 2.3 Sets of Numbers 26 2.4 Set Theory Into Practice: Samples Populations and Probability 27 2.5 Probability 28 2.6 Interpretations of Probability: Frequentist versus Subjective 35 2.7 Bayes Theorem: Inverting Conditional Probabilities 39 2.8 Statistical Inference 44 2.9 Essential Mathematics: Precalculus Calculus and Algebra 48 2.10 Chapter Summary and Highlights 72 Review Exercises 74 3 Introductory Statistics 78 3.1 Densities and Distributions 79 3.2 Chi-Square Distributions and Goodness-of-Fit Test 91 3.3 Sensitivity and Specificity 98 3.4 Scales of Measurement: Nominal Ordinal and Interval Ratio 98 3.5 Mathematical Variables versus Random Variables 101 3.6 Moments and Expectations 103 3.7 Estimation and Estimators 106 3.8 Variance 108 3.9 Degrees of Freedom 110 3.10 Skewness and Kurtosis 111 3.11 Sampling Distributions 113 3.12 Central Limit Theorem 116 3.13 Confidence Intervals 117 3.14 Bootstrap and Resampling Techniques 119 3.15 Likelihood Ratio Tests and Penalized Log-Likelihood Statistics 121 3.16 Akaike s Information Criteria 122 3.17 Covariance and Correlation 123 3.18 Other Correlation Coefficients 128 3.19 Student s t Distribution 131 3.20 Statistical Power 139 3.21 Paired Samples t-Test: Statistical Test for Matched Pairs (Elementary Blocking) Designs 146 3.22 Blocking with Several Conditions 149 3.23 Composite Variables: Linear Combinations 149 3.24 Models in Matrix Form 151 3.25 Graphical Approaches 152 3.26 What Makes a p-Value Small? A Critical Overview and Simple Demonstration of Null Hypothesis Significance Testing 155 3.27 Chapter Summary and Highlights 164 Review Exercises 167 4 Analysis of Variance: Fixed Effects Models 173 4.1 What is Analysis of Variance? Fixed versus Random Effects 174 4.2 How Analysis of Variance Works: A Big Picture Overview 178 4.3 Logic and Theory of ANOVA: A Deeper Look 180 4.4 From Sums of Squares to Unbiased Variance Estimators: Dividing by Degrees of Freedom 189 4.5 Expected Mean Squares for One-Way Fixed Effects Model: Deriving the F-Ratio 190 4.6 The Null Hypothesis in ANOVA 196 4.7 Fixed Effects ANOVA: Model Assumptions 198 4.8 A Word on Experimental Design and Randomization 201 4.9 A Preview of the Concept of Nesting 201 4.10 Balanced versus Unbalanced Data in ANOVA Models 202 4.11 Measures of Association and Effect Size in ANOVA: Measures of Variance Explained 202 4.12 The F-Test and the Independent Samples t-Test 205 4.13 Contrasts and Post-Hocs 205 4.14 Post-Hoc Tests 212 4.15 Sample Size and Power for ANOVA: Estimation with R and G Power 218 4.16 Fixed Effects One-Way Analysis of Variance in R: Mathematics Achievement as a Function of Teacher 222 4.17 Analysis of Variance Via R s lm 226 4.18 Kruskal Wallis Test in R 227 4.19 ANOVA in SPSS: Achievement as a Function of Teacher 228 4.20 Chapter Summary and Highlights 230 Review Exercises 232 5 Factorial Analysis of Variance: Modeling Interactions 237 5.1 What is Factorial Analysis of Variance? 238 5.2 Theory of Factorial ANOVA: A Deeper Look 239 5.3 Comparing One-Way ANOVA to Two-Way ANOVA: Cell Effects in Factorial ANOVA versus Sample Effects in One-Way ANOVA 245 5.4 Partitioning the Sums of Squares for Factorial ANOVA: The Case of Two Factors 246 5.5 Interpreting Main Effects in the Presence of Interactions 253 5.6 Effect Size Measures 253 5.7 Three-Way Four-Way and Higher-Order Models 254 5.8 Simple Main Effects 254 5.9 Nested Designs 256 5.9.1 Varieties of Nesting: Nesting of Levels versus Subjects 257 5.10 Achievement as a Function of Teacher and Textbook: Example of Factorial ANOVA in R 258 5.11 Interaction Contrasts 266 5.12 Chapter Summary and Highlights 267 Review Exercises 268 6 Introduction to Random Effects and Mixed Models 270 6.1 What is Random Effects Analysis of Variance? 271 6.2 Theory of Random Effects Models 272 6.3 Estimation in Random Effects Models 273 6.4 Defining Null Hypotheses in Random Effects Models 276 6.5 Comparing Null Hypotheses in Fixed versus Random Effects Models: The Importance of Assumptions 278 6.6 Estimating Variance Components in Random Effects Models: ANOVA ML REML Estimators 279 6.7 Is Achievement a Function of Teacher? One-Way Random Effects Model in R 282 6.8 R Analysis Using REML 285 6.9 Analysis in SPSS: Obtaining Variance Components 286 6.10 Factorial Random Effects: A Two-Way Model 287 6.11 Fixed Effects versus Random Effects: A Way of Conceptualizing Their Differences 289 6.12 Conceptualizing the Two-Way Random Effects Model: The Makeup of a Randomly Chosen Observation 289 6.13 Sums of Squares and Expected Mean Squares for Random Effects: The Contaminating Influence of Interaction Effects 291 6.14 You Get What You Go in with: The Importance of Model Assumptions and Model Selection 293 6.15 Mixed Model Analysis of Variance: Incorporating Fixed and Random Effects 294 6.16 Mixed Models in Matrices 298 6.17 Multilevel Modeling as a Special Case of the Mixed Model: Incorporating Nesting and Clustering 299 6.18 Chapter Summary and Highlights 300 Review Exercises 301 7 Randomized Blocks and Repeated Measures 303 7.1 What Is a Randomized Block Design? 304 7.2 Randomized Block Designs: Subjects Nested Within Blocks 304 7.3 Theory of Randomized Block Designs 306 7.4 Tukey Test for Nonadditivity 311 7.5 Assumptions for the Variance Covariance Matrix 311 7.6 Intraclass Correlation 313 7.7 Repeated Measures Models: A Special Case of Randomized Block Designs 314 7.8 Independent versus Paired Samples t-Test 315 7.9 The Subject Factor: Fixed or Random Effect? 316 7.10 Model for One-Way Repeated Measures Design 317 7.11 Analysis Using R: One-Way Repeated Measures: Learning as a Function of Trial 318 7.12 Analysis Using SPSS: One-Way Repeated Measures: Learning as a Function of Trial 322 7.13 SPSS: Two-Way Repeated Measures Analysis of Variance: Mixed Design: One Between Factor One Within Factor 326 7.14 Chapter Summary and Highlights 330 Review Exercises 331 8 Linear Regression 333 8.1 Brief History of Regression 334 8.2 Regression Analysis and Science: Experimental versus Correlational Distinctions 336 8.3 A Motivating Example: Can Offspring Height Be Predicted? 337 8.4 Theory of Regression Analysis: A Deeper Look 339 8.5 Multilevel Yearnings 342 8.6 The Least-Squares Line 342 8.7 Making Predictions Without Regression 343 8.8 More About i 345 8.9 Model Assumptions for Linear Regression 346 8.10 Estimation of Model Parameters in Regression 349 8.11 Null Hypotheses for Regression 351 8.12 Significance Tests and Confidence Intervals for Model Parameters 353 8.13 Other Formulations of the Regression Model 355 8.14 The Regression Model in Matrices: Allowing for More Complex Multivariable Models 356 8.15 Ordinary Least-Squares in Matrices 359 8.16 Analysis of Variance for Regression 360 8.17 Measures of Model Fit for Regression: How Well Does the Linear Equation Fit? 363 8.18 Adjusted R2 364 8.19 What Explained Variance Means: And More Importantly What It Does Not Mean 364 8.20 Values Fit by Regression 365 8.21 Least-Squares Regression in R: Using Matrix Operations 365 8.22 Linear Regression Using R 368 8.23 Regression Diagnostics: A Check on Model Assumptions 370 8.24 Regression in SPSS: Predicting Quantitative from Verbal 379 8.25 Power Analysis for Linear Regression in R 383 8.26 Chapter Summary and Highlights 384 Review Exercises 385 9 Multiple Linear Regression 389 9.1 Theory of Partial Correlation and Multiple Regression 390 9.2 Semipartial Correlations 392 9.3 Multiple Regression 393 9.4 Some Perspective on Regression Coefficients: Experimental Coefficients ? 394 9.5 Multiple Regression Model in Matrices 395 9.6 Estimation of Parameters 396 9.7 Conceptualizing Multiple R 396 9.8 Interpreting Regression Coefficients: The Case of Uncorrelated Predictors 397 9.9 Anderson s IRIS Data: Predicting Sepal Length from Petal Length and Petal Width 397 9.10 Fitting Other Functional Forms: A Brief Look at Polynomial Regression 402 9.11 Measures of Collinearity in Regression: Variance Inflation Factor and Tolerance 403 9.12 R-Squared as a Function of Partial and Semipartial Correlations: The Stepping Stones to Forward and Stepwise Regression 405 9.13 Model-Building Strategies: Simultaneous Hierarchichal Forward and Stepwise 406 9.14 Power Analysis for Multiple Regression 410 9.15 Introduction to Statistical Mediation: Concepts and Controversy 411 9.16 Chapter Summary and Highlights 414 Review Exercises 415 10 Interactions in Multiple Linear Regression: Dichotomous Polytomous and Continuous Moderators 418 10.1 The Additive Regression Model with Two Predictors 420 10.2 Why the Interaction is the Product Term xizi: Drawing an Analogy to Factorial ANOVA 420 10.3 A Motivating Example of Interaction in Regression: Crossing a Continuous Predictor with a Dichotomous Predictor 421 10.4 Theory of Interactions in Regression 424 10.5 Simple Slopes for Continuous Moderators 427 10.6 A Simple Numerical Example: How Slopes Can Change as a Function of the Moderator 428 10.7 Calculating Simple Slopes: A Useful Algebraic Derivation 430 10.8 Summing Up the Idea of Interactions in Regression 432 10.9 Do Moderators Really Moderate Anything? Some Philosophical Considerations 432 10.10 Interpreting Model Coefficients in the Context of Moderators 433 10.11 Mean-Centering Predictors: Improving the Interpretability of Simple Slopes 434 10.12 The Issue of Multicollinearity: A Second Reason to Like Mean-Centering 435 10.13 Interaction of Continuous and Polytomous Predictors in R 436 10.14 Multilevel Regression: Another Special Case of the Mixed Model 440 10.15 Chapter Summary and Highlights 441 Review Exercises 441 11 Logistic Regression and the Generalized Linear Model 443 11.1 Nonlinear Models 445 11.2 Generalized Linear Models 447 11.3 Canonical Links 450 11.4 Distributions and Generalized Linear Models 451 11.5 Dispersion Parameters and Deviance 453 11.6 Logistic Regression: A Generalized Linear Model for Binary Responses 454 11.7 Exponential and Logarithmic Functions 456 11.8 Odds Odds Ratio and the Logit 461 11.9 Putting It All Together: The Logistic Regression Model 462 11.10 Logistic Regression in R: Challenger O-Ring Data 466 11.11 Challenger Analysis in SPSS 469 11.12 Sample Size Effect Size and Power 473 11.13 Further Directions 474 11.14 Chapter Summary and Highlights 475 Review Exercises 476 12 Multivariate Analysis of Variance 479 12.1 A Motivating Example: Quantitative and Verbal Ability as a Variate 480 12.2 Constructing the Composite 482 12.3 Theory of MANOVA 482 12.4 Is the Linear Combination Meaningful? 483 12.5 Multivariate Hypotheses 487 12.6 Assumptions of MANOVA 488 12.7 Hotelling s T2: The Case of Generalizing from Univariate to Multivariate 489 12.8 The Variance Covariance Matrix S 492 12.9 From Sums of Squares and Cross-Products to Variances and Covariances 494 12.10 Hypothesis and Error Matrices of MANOVA 495 12.11 Multivariate Test Statistics 495 12.12 Equality of Variance Covariance Matrices 500 12.13 Multivariate Contrasts 501 12.14 MANOVA in R and SPSS 502 12.15 MANOVA of Fisher s Iris Data 508 12.16 Power Analysis and Sample Size for MANOVA 509 12.17 Multivariate Analysis of Covariance and Multivariate Models: A Bird s Eye View of Linear Models 511 12.18 Chapter Summary and Highlights 512 Review Exercises 513 13 Discriminant Analysis 517 13.1 What is Discriminant Analysis? The Big Picture on the Iris Data 518 13.2 Theory of Discriminant Analysis 520 13.3 LDA in R and SPSS 523 13.4 Discriminant Analysis for Several Populations 529 13.5 Discriminating Species of Iris: Discriminant Analyses for Three Populations 532 13.6 A Note on Classification and Error Rates 535 13.7 Discriminant Analysis and Beyond 537 13.8 Canonical Correlation 538 13.9 Motivating Example for Canonical Correlation: Hotelling s 1936 Data 539 13.10 Canonical Correlation as a General Linear Model 540 13.11 Theory of Canonical Correlation 541 13.12 Canonical Correlation of Hotelling s Data 544 13.13 Canonical Correlation on the Iris Data: Extracting Canonical Correlation from Regression MANOVA LDA 546 13.14 Chapter Summary and Highlights 547 Review Exercises 548 14 Principal Components Analysis 551 14.1 History of Principal Components Analysis 552 14.2 Hotelling 1933 555 14.3 Theory of Principal Components Analysis 556 14.4 Eigenvalues as Variance 557 14.5 Principal Components as Linear Combinations 558 14.6 Extracting the First Component 558 14.7 Extracting the Second Component 560 14.8 Extracting Third and Remaining Components 561 14.9 The Eigenvalue as the Variance of a Linear Combination Relative to Its Length 561 14.10 Demonstrating Principal Components Analysis: Pearson s 1901 Illustration 562 14.11 Scree Plots 566 14.12 Principal Components versus Least-Squares Regression Lines 569 14.13 Covariance versus Correlation Matrices: Principal Components and Scaling 570 14.14 Principal Components Analysis Using SPSS 570 14.15 Chapter Summary and Highlights 575 Review Exercises 576 15 Factor Analysis 579 15.1 History of Factor Analysis 580 15.2 Factor Analysis: At a Glance 580 15.3 Exploratory versus Confirmatory Factor Analysis 581 15.4 Theory of Factor Analysis: The Exploratory Factor-Analytic Model 582 15.5 The Common Factor-Analytic Model 583 15.6 Assumptions of the Factor-Analytic Model 585 15.7 Why Model Assumptions Are Important 587 15.8 The Factor Model as an Implication for the Covariance Matrix 587 15.9 Again Why is ' so Important a Result? 589 15.10 The Major Critique Against Factor Analysis: Indeterminacy and the Nonuniqueness of Solutions 589 15.11 Has Your Factor Analysis Been Successful? 591 15.12 Estimation of Parameters in Exploratory Factor Analysis 592 15.13 Estimation of Factor Scores 593 15.14 Principal Factor 593 15.15 Maximum Likelihood 595 15.16 The Concepts (and Criticisms) of Factor Rotation 596 15.17 Varimax and Quartimax Rotation 599 15.18 Should Factors Be Rotated? Is That Not Cheating? 600 15.19 Sample Size for Factor Analysis 601 15.20 Principal Components Analysis versus Factor Analysis: Two Key Differences 602 15.21 Principal Factor in SPSS: Principal Axis Factoring 604 15.22 Bartlett Test of Sphericity and Kaiser Meyer Olkin Measure of Sampling Adequacy (MSA) 612 15.23 Factor Analysis in R: Holzinger and Swineford (1939) 613 15.24 Cluster Analysis 616 15.25 What Is Cluster Analysis? The Big Picture 617 15.26 Measuring Proximity 619 15.27 Hierarchical Clustering Approaches 623 15.28 Nonhierarchical Clustering Approaches 625 15.29 K-Means Cluster Analysis in R 626 15.30 Guidelines and Warnings About Cluster Analysis 630 15.31 Chapter Summary and Highlights 630 Review Exercises 632 16 Path Analysis and Structural Equation Modeling 636 16.1 Path Analysis: A Motivating Example Predicting IQ Across Generations 637 16.2 Path Analysis and Causal Modeling 639 16.3 Early Post-Wright Path Analysis: Predicting Child s IQ (Burks 1928) 641 16.4 Decomposing Path Coefficients 642 16.5 Path Coefficients and Wright s Contribution 644 16.6 Path Analysis in R: A Quick Overview Modeling Galton s Data 644 16.7 Confirmatory Factor Analysis: The Measurement Model 648 16.8 Structural Equation Models 650 16.9 Direct Indirect and Total Effects 652 16.10 Theory of Statistical Modeling: A Deeper Look into Covariance Structures and General Modeling 653 16.11 Other Discrepancy Functions 655 16.12 The Discrepancy Function and Chi-Square 656 16.13 Identification 657 16.14 Disturbance Variables 659 16.15 Measures and Indicators of Model Fit 660 16.16 Overall Measures of Model Fit 660 16.17 Model Comparison Measures: Incremental Fit Indices 662 16.18 Which Indicator of Model Fit Is Best? 665 16.19 Structural Equation Model in R 666 16.20 How All Variables Are Latent: A Suggestion for Resolving the Manifest Latent Distinction 668 16.21 The Structural Equation Model as a General Model: Some Concluding Thoughts on Statistics and Science 669 16.22 Chapter Summary and Highlights 670 Review Exercises 671 Appendix A: Matrix Algebra 675 References 705 Index 721
זמן אספקה 21 ימי עסקים