| |

| |

Preface | |

| |

| |

Introduction | |

| |

| |

The "Image" of Statistics | |

| |

| |

Descriptive Statistics | |

| |

| |

Inferential Statistics | |

| |

| |

Statistics and Mathematics | |

| |

| |

Case Method | |

| |

| |

Our Targets | |

| |

| |

Measurement, Variables, and Scales | |

| |

| |

Variables and their Measurement | |

| |

| |

Measurement: The Observation of Variables | |

| |

| |

Measurement Scales; Nominal Measurement | |

| |

| |

Ordinal Measurement | |

| |

| |

Interval Measurement | |

| |

| |

Ratio Measurement | |

| |

| |

Interrelationships among Measurement Scales | |

| |

| |

Continuous and Discrete Variables | |

| |

| |

Frequency Distributions and Visual Displays of Data | |

| |

| |

Tabulating Data | |

| |

| |

Grouped Frequency Distributions | |

| |

| |

Grouping and Loss of Information | |

| |

| |

Graphing a Frequency Distribution: The Histogram | |

| |

| |

Frequency and Percentage Polygons | |

| |

| |

Type of Distribution | |

| |

| |

Cumulative Distributions and the Ogive Curve | |

| |

| |

Percentiles | |

| |

| |

Box-and-Whisker Plots | |

| |

| |

Stem-and-Leaf Displays | |

| |

| |

Time-Series Graphs | |

| |

| |

Misleading Graphs-How to Lie with Statistics | |

| |

| |

Measures of Central Tendency | |

| |

| |

The Mode | |

| |

| |

The Median | |

| |

| |

Summation Notation | |

| |

| |

The Mean | |

| |

| |

More Summation Notation | |

| |

| |

Adding or Subtracting a Constant | |

| |

| |

Multiplying or Dividing by a Constant | |

| |

| |

Sum of Deviations | |

| |

| |

Sum of Squared Deviations | |

| |

| |

The Mean of the Sum of Two or More Scores | |

| |

| |

The Mean of a Difference | |

| |

| |

Mean, Median, and Mode of Two or More Groups | |

| |

| |

Interpretation of Mode, Median, and Mean | |

| |

| |

Central Tendency and Skewness | |

| |

| |

Measures of Central Tendency as Inferential Statistics | |

| |

| |

Which Measure is Best? | |

| |

| |

Measures of Variability | |

| |

| |

The Range | |

| |

| |

H-Spread and the Interquartile Range | |

| |

| |

Deviation Scores | |

| |

| |

Sum of Squares | |

| |

| |

More about the Summation Operator, <F128>- | |

| |

| |

The Variance of a Population | |

| |

| |

The Variance Estimated From a Sample | |

| |

| |

The Standard Deviation | |

| |

| |

The Effect of Adding or Subtracting a Constant on Measures of Variability | |

| |

| |

The Effect of Multiplying or Dividing a Constant on Measures of Variability | |

| |

| |

Variance of a Combined Distribution | |

| |

| |

Inferential Properties of the Range, s2, and s | |

| |

| |

The Normal Distribution and Standard Scores | |

| |

| |

The Importance of the Normal Distribution | |

| |

| |

God Loves the Normal Curve | |

| |

| |

The Standard Normal Distribution as a Standard Reference Distribution: z-Scores | |

| |

| |

Ordinates of the Normal Distribution | |

| |

| |

Areas Under the Normal Curve | |

| |

| |

Other Standard Scores | |

| |

| |

T-Scores | |

| |

| |

Areas Under the Normal Curve in Samples | |

| |

| |

Skewness | |

| |

| |

Kurtosis | |

| |

| |

Transformations | |

| |

| |

Normalized Scores | |

| |

| |

Correlation: Measures of Relationship Between Two Variables | |

| |

| |

The Concept of Correlation | |

| |

| |

Scatterplots | |

| |

| |

The Measurement of Correlation | |

| |

| |

The Use of Correlation Coefficients | |

| |

| |

Interpreting r as a Percent | |

| |

| |

Linear and Curvilinear Relationships | |

| |

| |

Calculating the Pearson Product-Moment Correlation Coefficient, r | |

| |

| |

Scatterplots | |

| |

| |

Correlation Expressed in Terms of z-Scores | |

| |

| |

Linear Transformations and Correlation | |

| |

| |

The Bivariate Normal Distribution | |

| |

| |

Effects of Variability on Correlation | |

| |

| |

Correcting for Restricted Variability | |

| |

| |

Effect of Measurement Error on r and the Correction for Attenuation | |

| |

| |

The Pearson r and Marginal Distributions | |

| |

| |

The Effect of the Unit of Analysis on Correlation: Ecological Correlations | |

| |

| |

The Variance of a Sum | |

| |

| |

The Variance of a Difference | |

| |

| |

Additional Measures of Relationship: The Spearman Rank Correlation | |

| |

| |

The Phi Coefficient: Both X and Y are Dichotomies | |

| |

| |

The Point Biserial Coefficient | |

| |

| |

The Biserial Correlation | |

| |

| |

Biserial versus Point-Biserial Correlation Coefficients | |

| |

| |

The Tetrachoric Coefficient | |

| |

| |

Causation and Correlation | |

| |

| |

Regression and Prediction | |

| |

| |

Purposes of Regression Analysis | |

| |

| |

The Regression Effect | |

| |

| |

The Regression Equation Expressed in Standard z-Scores | |

| |

| |

Use of Regression Equations | |

| |

| |

Cartesian Coordinates | |

| |

| |

Estimating Y from X: The Raw-score Regression Equation | |

| |

| |

Error of Estimate | |

| |

| |

Proportion of Predictable Variance | |

| |

| |

Least-squares Criterion | |

| |

| |

Homoscedasticity and the Standard Error of Estimate | |

| |

| |

Regression and Pretest-Posttest Gains | |

| |

| |

Part Correlation | |

| |

| |

Partial Correlation | |

| |

| |

Second-Order Partial Correlation | |

| |

| |

Multiple Regression and Multiple Correlation | |

| |

| |

The Standardized Regression Equation | |

| |

| |

The Raw-Score Regression Equation | |

| |

| |

Multiple Correlation | |

| |

| |

Multiple Regression Equation with Three or More Independent Variables | |

| |

| |

Stepwise Multiple Regression | |

| |

| |

Illustration of Stepwise Multiple Regression | |

| |

| |

Dichotomous and Categorical Variables as Predictors | |

| |

| |

The Standard Error of Estimate in Multiple Regression | |

| |

| |

The Multiple Correlation as an Inferential Statistic: Correction for Bias | |

| |

| |

Assumptions | |

| |

| |

Curvilinear Regression and Correlation | |

| |

| |

Measuring Non-linear Relationships between Two Variables | |

| |

| |

Transforming Non-linear Relationships into Linear Relationships | |

| |

| |

Dichotomous Dependent Variables: Logistic Regression | |

| |

| |

Categorical Dependent Variables more than Two Categories: Discriminant Analysis | |

| |

| |

Probability | |

| |

| |

Probability as a Mathematical System | |

| |

| |

First Addition Rule of Probabilities | |

| |

| |

Second Addition Rule of Probabilities | |

| |

| |

Multiplication Rule of Probabilities | |

| |

| |

Conditional Probability | |

| |

| |

Bayes's Theorem | |

| |

| |

Permutations | |

| |

| |

Combinations | |

| |

| |

Binomial Probabilities | |

| |

| |

The Binomial and Sign Test | |

| |

| |

Intuition and Probability | |

| |

| |

Probability as an Area | |

| |

| |

Combining Probabilities | |

| |

| |

Expectations and Moments | |

| |

| |

Statistical Inference: Sampling and Interval Estimation | |

| |

| |

Overview | |

| |

| |

Populations and Samples: Parameters and Statistics | |

| |

| |

Infinite versus Finite Populations | |

| |

| |

Randomness and Random Sampling | |

| |

| |

Accidental or Convenience Samples | |

| |

| |

Random Samples | |

| |

| |

Independence | |

| |

| |

Systematic Sampling | |

| |

| |

Point and Interval Estimates | |

| |

| |

Sampling Distributions | |

| |

| |

The Standard Error of the Mean | |

| |

| |

Relationship of sx to n | |

| |

| |

Confidence Intervals | |

| |

| |

Confidence Intervals when s is Known: An Example | |

| |

| |

Central Limit Theorem: A Demonstration | |

| |

| |

The Use of Sampling Distributions | |

| |

| |

Proof that s2 = s2/n | |

| |

| |

Properties of Estimators | |

| |

| |

Unbiasedness | |

| |

| |

Consistency | |

| |

| |

Relative Efficiency | |

| |

| |

Introduction to Hypothesis Testing | |

| |

| |

Statistical Hypotheses and Explanations | |

| |

| |

Statistical versus Scientific Hypotheses | |

| |

| |

Testing Hypotheses about ï¿½ï¿½ | |

| |

| |

Testing H0: ï¿½ï¿½ = K, a One-Sample z-Test | |

| |

| |

Two Types of Errors in Hypothesis Testing | |

| |

| |

Hypothesis Testing and Confidence Intervals | |

| |

| |

Type-II Error, b, and Power | |

| |

| |

Power | |

| |

| |

Effect of a on Power | |

| |

| |

Power and the Value Hypothesized in the Alternative Hypothesis | |

| |

| |

Methods of Increasing Power | |

| |

| |

Non-Directional and Directional Alternatives: Two-Tailed versus One- Tailed Tests | |

| |

| |

Statistical Significance versus Practical Significance | |

| |

| |

Confidence Limits for the Population Median | |

| |

| |

Inference Regarding ï¿½ï¿½ when s is not Known: t versus z | |

| |

| |

The t-Distribution | |

| |

| |

Confidence Intervals Using the t-Distribution | |

| |

| |

Accuracy of Confidence Intervals when Sampling Non-Normal Distributions | |

| |

| |

Inferences about the Difference Between Two Means | |

| |

| |

Testing Statistical Hypotheses Involving Two Means | |

| |

| |

The Null Hypotheses | |

| |

| |

The t-Test for Comparing Two Independent Means | |

| |

| |

Computing sx1-x2 | |

| |

| |

An Illustration | |

| |

| |

Confidence Intervals about Mean Differences | |

| |

| |

Effect Size | |

| |

| |

t-Test Assumptions and Robustness | |

| |

| |

Homogeneity of Variance | |

| |

| |

What if Sample Sizes Are Unequal and Variances Are Heterogeneous: The Welch t' Test | |

| |

| |

Independence of Observations | |

| |

| |

Testing H0: ï¿½ï¿½1 = ï¿½ï¿½2 with Paired Observations | |

| |

| |

Direct Difference for the t-Test for Paired Observations | |

| |

| |

Cautions Regarding the Matched-Pairs Designs in Research | |

| |

| |

Power when Comparing Means | |

| |

| |

Non-Parametric Alternatives: The Mann-Whitney Test and the Wilcoxon Signed-Rank Test | |

| |

| |

Statistics for Categorical Dependent Variables: Inferences about Proportions | |

| |

| |

Overview | |

| |

| |

The Proportion as a Mean | |

| |

| |

The Variance of a Proportion | |

| |

| |

The Sampling distribution of a Proportion: The Standard Error of p | |

| |

| |

The Influence of n on sp | |

| |

| |

Influence of the Sampling Fraction on sp | |

| |

| |

The Influence of P on sp | |

| |

| |

Confidence Intervals for P | |

| |

| |

Quick Confidence Intervals for P | |

| |

| |

Testing H0: P = K | |

| |

| |

Testing Empirical versus Theoretical Distributions: Chi-Square Goodness of Fit Test | |

| |

| |

Testing Differences among Proportions: The Chi-Square Test of Association | |

| |

| |

Other Formulas for the Chi-Square Test of Association | |

| |

| |

The C2 Median Test | |

| |

| |

Chi-Square and the Phi Coefficient | |

| |

| |

Independence of Observations | |

| |

| |

Inferences about H0: P1 = P2 when Observations are Paired: McNemar's Test for Correlated Proportions | |

| |

| |

Inferences about Correlation Coefficient | |

| |

| |

Testing Statistical Hypotheses Regarding r | |

| |

| |

Testing H0: r = 0 Using the t-Test | |

| |

| |

Directional Alternatives: "Two-Tailed" vs. "One- Tailed" Tests | |

| |

| |

Sampling Distribution of r | |

| |

| |

The Fisher Z-Transformation | |

| |

| |

Setting Confidence Intervals for r | |

| |

| |

Determining Confidence Intervals Graphically | |

| |

| |

Testing the Difference between Independent Correlation Coefficients: H0: r1 = e2 = ...ej | |

| |

| |

Averaging r's | |

| |

| |

Testing Differences between Two Dependent Correlation Coefficients: H0: e31 = r32 | |

| |

| |

Inferences about Other Correlation Coefficients | |

| |

| |

The Point-Biserial Correlation Coefficient rpr | |

| |

| |

Spearman's Rank Correlation: H0: ranks = 0 | |

| |

| |

Partial Correlation: H0: r12.3 = 0 | |

| |

| |

Significance of a Multiple Correlation Coefficient | |

| |

| |

Statistical Significance in Stepwise Multiple Regression | |

| |

| |

Significance of the Biserial Correlation Coefficient rbis | |

| |

| |

Significance of the Tetrachoric Correlation Coefficient rtet | |

| |

| |

Significance of the Correlation Ratio Eta | |

| |

| |

Testing for Non-linearity of Regression | |

| |

| |

One-Factor Analysis of Variance | |

| |

| |

Why Not Several t-Tests? | |

| |

| |

ANOVA Nomenclature | |

| |

| |

ANOVA Computation | |

| |

| |

Sum of Squares Between, SSB | |

| |

| |

Sum of Squares Within, SSW | |

| |

| |

ANOVA Computational Illustration | |

| |

| |

ANOVA Theory | |

| |

| |

Mean Square Between Groups, MSB | |

| |

| |

Mean Square Within Groups, MSW | |

| |

| |

The F-Test | |

| |

| |

ANOVA with Equal n's | |

| |

| |

A Statistical Model for the Data | |

| |

| |

Estimates of the Terms in the Model | |

| |

| |

Sum of Squares | |

| |

| |

Restatement of the Null Hypothesis in Terms of Population Means | |

| |

| |

Degrees of Freedom | |

| |

| |

Mean Squares: The Expected Value of MSW | |

| |

| |

The Expected Value of MSB | |

| |

| |

Some Distribution Theory | |

| |

| |

The F-Test of the Null Hypothesis: Rationale and Procedure | |

| |

| |

Type-I versus Type-II Errors: a and b | |

| |

| |

A Summary of Procedures for One-Factor ANOVA | |

| |

| |

Consequences of Failure to Meet the ANOVA Assumptions: The "Robustness" of ANOVA | |

| |

| |

The Welch and Brown-Forsythe Modifications of ANOVA: What Does One Do When <F128>-'s and n's Differ? | |

| |

| |

The Power of the F-Test | |

| |

| |

An Illustration | |

| |

| |

Power When s is Unknown | |

| |

| |

A Table for Estimating Power When J=2 | |

| |

| |

The Non-Parametric Alternative: The Krukal-Wallis Test | |

| |

| |

Inferences About Variances | |

| |

| |

Chi-Square Distributions | |

| |

| |

Chi-Square Distributions with u1: c 2u | |

| |

| |

The Chi-Square Distribution with u Degrees of Freedom, c2u | |

| |

| |

Inferences about the Population Variance: H0: s2 = K | |

| |

| |

F-Distributions | |

| |

| |

Inferences about Two Independent Variances: H0: s21 = s22 | |

| |

| |

Testing Homogeneity of Variance: Hartley's Fmax Test | |

| |

| |

Testing Homogeneity Variance from J Independent Samples: The Bartlett Test | |

| |

| |

Other Tests of Homogeneity of Variance: The Levene and Brown-Forsythe Tests | |

| |

| |

Inferences about H0: s21 = s22 with Paired Observations | |

| |

| |

Relationships among the Normal, t, c2 and F-Distributions | |

| |

| |

Multiple Comparisons and Trend Analysis | |

| |

| |

Testing All Pairs of Means: The Studentized Range Statistic, q | |

| |

| |

The Tukey Method of Multiple Comparisons | |

| |

| |

The Effect Size of Mean Differences | |

| |

| |

The Basis for Type-I Error Rate: Contrast vs. Family | |

| |

| |

The Newman-Keuls Method | |

| |

| |

The Tukey and Newman-Keuls Methods Compared | |

| |

| |

The Definition of a Contrast | |

| |

| |

Simple versus Complex Contrasts | |

| |

| |

The Standard Error of a Contrast | |

| |

| |

The t-Ratio for a Contrast | |

| |

| |

Planned versus Post Hoc Comparisons | |

| |

| |

Dunn (Bonferroni) Method of Multiple Comparisons | |

| |

| |

Dunnett Method of Multiple Comparisons | |

| |

| |

Scheffe Method of Multiple Comparisons | |

| |

| |

Planned Orthogonal Contrasts | |

| |

| |

Confidence Intervals for Contrasts | |

| |

| |

Relative Power of Multiple Comparison Techniques | |

| |

| |

Trend Analysis | |

| |

| |

Significance of Trend Components | |

| |

| |

Relation to Trends to Correlation Coefficients | |

| |

| |

Assumptions of MC Methods | |

| |

| |

Multiple Comparisons for Other Statistics | |

| |

| |

Chapter Summary and Criteria for Selecting a Multiple Comparison Method | |

| |

| |

Two and Three Factor ANOVA: An Introduction to Factorial Designs | |

| |

| |

The Meaning of Interaction | |

| |

| |

Interactions and Generalizability: Factors Do Not Interact | |

| |

| |

Interactions and Generalizability: Factors Interact | |

| |

| |

Interpreting when Interaction is Present | |

| |

| |

Statistical Significance and Interaction | |

| |

| |

Data Layout and Notation | |

| |

| |

A Model for the Data | |

| |

| |

Least-Squares of the Model | |

| |

| |

Statement of Null Hypotheses | |

| |

| |

Sums of Squares in the Two-Factor ANOVA | |

| |

| |

Degrees of Freedom | |

| |

| |

Mean Squares | |

| |

| |

Illustration of the Computation for the Two-Factor ANOVA | |

| |

| |

Expected Values of Mean Squares | |

| |

| |

The Distribution of the Mean Squares | |

| |

| |

Determining Power in Factorial Designs | |

| |

| |

Multiple Comparisons in Factorial ANOVA Designs | |

| |

| |

Confidence Intervals for Means in Two-Factor ANOVA | |

| |

| |

Three-Factor ANOVA | |

| |

| |

Three-Factor ANOVA: An Illustration | |

| |

| |

Three-Factor ANOVA Computation | |

| |

| |

The Interpretation of Three-Factor Interaction | |

| |

| |

Confidence Intervals in Three-Factor ANOVA | |

| |

| |

How Factorial Designs Increase Power | |

| |

| |

Factorial ANOVA with Unequal n's | |

| |

| |

Multi-Factor ANOVA Designs: Random, Mixed, and Fixed Effects | |

| |

| |

The Random-Effects ANOVA Model | |

| |

| |

Assumptions of the Random ANOVA Model | |

| |

| |

An Example | |

| |

| |

Mean Square, MSW | |

| |

| |

Mean Square, MSB | |

| |

| |

The Variance Component, sa2 | |

| |

| |

Confidence Interval for sa2/se2 | |

| |

| |

Summary of Random ANOVA Model | |

| |

| |

The Mixed-Effects ANOVA Model | |

| |

| |

Mixed-Model ANOVA Assumptions | |

| |

| |

Mixed-Model ANOVA Computation | |

| |

| |

Multiple Comparisons in the Two-Factor Mixed Model | |

| |

| |

Crossed and Nested Factors | |

| |

| |

Computation of Sums of Squares for Nested Factors | |

| |

| |

Determining the Sources of Variation in the ANOVA Table | |

| |

| |

Degrees of Freedom for Nested Factors | |

| |

| |

Determining Expected Mean Squares | |

| |

| |

Error Mean Square in Complex ANOVA Designs | |

| |

| |

The Incremental Generalization Strategy: Inferential "Concentric Circles." | |

| |

| |

Model Simplification and Pooling | |

| |

| |

The Experimental Unit and the Observational Unit | |

| |

| |

Repeated- Measures ANOVA | |

| |

| |

A Simple Repeated-Measures ANOVA | |

| |

| |

Repeated-Measures Assumptions | |

| |

| |

Trend Analysis on Repeated-Measures Factors | |

| |

| |

Estimating Reliability via Repeated-Measures ANOVA | |

| |

| |

Repeated-Measures Designs with a Between-Subjects Factor | |

| |

| |

Repeated-Measures ANOVA with Two Between-Subjects Factors | |

| |

| |

Trend Analysis on Between-Subjects Factors | |

| |

| |

Repeated-Measures ANOVA with Two Within-Subjects Factors and Two Between-Subjects Factors | |

| |

| |

Repeated-Measures ANOVA vs. MANOVA | |

| |

| |

An Introduction to the Analysis of Covariance | |

| |

| |

The Functions of ANCOVA | |

| |

| |

ANOVA Results | |

| |

| |

ANCOVA Model | |

| |

| |

ANCOVA Computations, SStotal | |

| |

| |

The Adjusted Within Sum of Squares, SS'W | |

| |

| |

The Adjusted Sum of Squares Between Groups, SS'B | |

| |

| |

Degrees of Freedom in ANCOVA and the ANCOVA Table | |

| |

| |

Adjusted Means, Y'j | |

| |

| |

Confidence Intervals and Multiple Comparisons for Adjusted Means | |

| |

| |

ANCOVA Illustrated Graphically | |

| |

| |

ANCOVA Assumptions | |

| |

| |

ANCOVA Precautions | |

| |

| |

Covarying and Stratifying | |

| |

| |

Appendix: Tables | |

| |

| |

Unit-Normal (z) Distribution | |

| |

| |

Random Digits | |

| |

| |

t-Distribution | |

| |

| |

c2-Distribution | |

| |

| |

Fisher Z-Transformation | |

| |

| |

F-Distribution | |

| |

| |

Power Curves for the F-Test | |

| |

| |

Hartley's Fmax Distribution | |

| |

| |

Studentized Range Statistic: q-Distribution | |

| |

| |

Critical Values of r | |

| |

| |

Critical Values of rranks, Spearman's Rank Correlation | |

| |

| |

Critical Values for the Dunn (Bonferroni) t-Statistic | |

| |

| |

Critical Values for the Dunnett t-Statistic | |

| |

| |

Coefficients (Orthogonal Polynomials) for Trend Analysis | |

| |

| |

Binomial Probabilities when P = .5 | |

| |

| |

Glossary of Symbols | |

| |

| |

Bibliography | |

| |

| |

Author Index | |

| |

| |

Subject Index | |