Skip to content

Statistical Methods in Education and Psychology

Best in textbook rentals since 2012!

ISBN-10: 0205142125

ISBN-13: 9780205142125

Edition: 3rd 1996 (Revised)

Authors: Gene V. Glass, Kenneth D. Hopkins

List price: $185.00
Blue ribbon 30 day, 100% satisfaction guarantee!
what's this?
Rush Rewards U
Members Receive:
Carrot Coin icon
XP icon
You have reached 400 XP and carrot coins. That is the daily max!

Description:

The approach of Statistical Methods in Education and Psychology, Third Edition, is conceptual rather than mathematical. The authors stress the understanding, applications, and interpretation of concepts rather than derivation and proof or hand-computation. Selection of topics in the book was guided by three considerations: (1) What are the most useful statistical methods?; (2) Which statistical methods are the most widely used in journals in the behavioral and social sciences?; and (3) Which statistical methods are fundamental to further study?
Customers also bought

Book details

List price: $185.00
Edition: 3rd
Copyright year: 1996
Publisher: Allyn & Bacon, Incorporated
Publication date: 8/10/1995
Binding: Hardcover
Pages: 608
Size: 7.50" wide x 9.75" long x 1.25" tall
Weight: 2.596

Preface
Introduction
The "Image" of Statistics
Descriptive Statistics
Inferential Statistics
Statistics and Mathematics
Case Method
Our Targets
Measurement, Variables, and Scales
Variables and their Measurement
Measurement: The Observation of Variables
Measurement Scales; Nominal Measurement
Ordinal Measurement
Interval Measurement
Ratio Measurement
Interrelationships among Measurement Scales
Continuous and Discrete Variables
Frequency Distributions and Visual Displays of Data
Tabulating Data
Grouped Frequency Distributions
Grouping and Loss of Information
Graphing a Frequency Distribution: The Histogram
Frequency and Percentage Polygons
Type of Distribution
Cumulative Distributions and the Ogive Curve
Percentiles
Box-and-Whisker Plots
Stem-and-Leaf Displays
Time-Series Graphs
Misleading Graphs-How to Lie with Statistics
Measures of Central Tendency
The Mode
The Median
Summation Notation
The Mean
More Summation Notation
Adding or Subtracting a Constant
Multiplying or Dividing by a Constant
Sum of Deviations
Sum of Squared Deviations
The Mean of the Sum of Two or More Scores
The Mean of a Difference
Mean, Median, and Mode of Two or More Groups
Interpretation of Mode, Median, and Mean
Central Tendency and Skewness
Measures of Central Tendency as Inferential Statistics
Which Measure is Best?
Measures of Variability
The Range
H-Spread and the Interquartile Range
Deviation Scores
Sum of Squares
More about the Summation Operator, <F128>-
The Variance of a Population
The Variance Estimated From a Sample
The Standard Deviation
The Effect of Adding or Subtracting a Constant on Measures of Variability
The Effect of Multiplying or Dividing a Constant on Measures of Variability
Variance of a Combined Distribution
Inferential Properties of the Range, s2, and s
The Normal Distribution and Standard Scores
The Importance of the Normal Distribution
God Loves the Normal Curve
The Standard Normal Distribution as a Standard Reference Distribution: z-Scores
Ordinates of the Normal Distribution
Areas Under the Normal Curve
Other Standard Scores
T-Scores
Areas Under the Normal Curve in Samples
Skewness
Kurtosis
Transformations
Normalized Scores
Correlation: Measures of Relationship Between Two Variables
The Concept of Correlation
Scatterplots
The Measurement of Correlation
The Use of Correlation Coefficients
Interpreting r as a Percent
Linear and Curvilinear Relationships
Calculating the Pearson Product-Moment Correlation Coefficient, r
Scatterplots
Correlation Expressed in Terms of z-Scores
Linear Transformations and Correlation
The Bivariate Normal Distribution
Effects of Variability on Correlation
Correcting for Restricted Variability
Effect of Measurement Error on r and the Correction for Attenuation
The Pearson r and Marginal Distributions
The Effect of the Unit of Analysis on Correlation: Ecological Correlations
The Variance of a Sum
The Variance of a Difference
Additional Measures of Relationship: The Spearman Rank Correlation
The Phi Coefficient: Both X and Y are Dichotomies
The Point Biserial Coefficient
The Biserial Correlation
Biserial versus Point-Biserial Correlation Coefficients
The Tetrachoric Coefficient
Causation and Correlation
Regression and Prediction
Purposes of Regression Analysis
The Regression Effect
The Regression Equation Expressed in Standard z-Scores
Use of Regression Equations
Cartesian Coordinates
Estimating Y from X: The Raw-score Regression Equation
Error of Estimate
Proportion of Predictable Variance
Least-squares Criterion
Homoscedasticity and the Standard Error of Estimate
Regression and Pretest-Posttest Gains
Part Correlation
Partial Correlation
Second-Order Partial Correlation
Multiple Regression and Multiple Correlation
The Standardized Regression Equation
The Raw-Score Regression Equation
Multiple Correlation
Multiple Regression Equation with Three or More Independent Variables
Stepwise Multiple Regression
Illustration of Stepwise Multiple Regression
Dichotomous and Categorical Variables as Predictors
The Standard Error of Estimate in Multiple Regression
The Multiple Correlation as an Inferential Statistic: Correction for Bias
Assumptions
Curvilinear Regression and Correlation
Measuring Non-linear Relationships between Two Variables
Transforming Non-linear Relationships into Linear Relationships
Dichotomous Dependent Variables: Logistic Regression
Categorical Dependent Variables more than Two Categories: Discriminant Analysis
Probability
Probability as a Mathematical System
First Addition Rule of Probabilities
Second Addition Rule of Probabilities
Multiplication Rule of Probabilities
Conditional Probability
Bayes's Theorem
Permutations
Combinations
Binomial Probabilities
The Binomial and Sign Test
Intuition and Probability
Probability as an Area
Combining Probabilities
Expectations and Moments
Statistical Inference: Sampling and Interval Estimation
Overview
Populations and Samples: Parameters and Statistics
Infinite versus Finite Populations
Randomness and Random Sampling
Accidental or Convenience Samples
Random Samples
Independence
Systematic Sampling
Point and Interval Estimates
Sampling Distributions
The Standard Error of the Mean
Relationship of sx to n
Confidence Intervals
Confidence Intervals when s is Known: An Example
Central Limit Theorem: A Demonstration
The Use of Sampling Distributions
Proof that s2 = s2/n
Properties of Estimators
Unbiasedness
Consistency
Relative Efficiency
Introduction to Hypothesis Testing
Statistical Hypotheses and Explanations
Statistical versus Scientific Hypotheses
Testing Hypotheses about ��
Testing H0: �� = K, a One-Sample z-Test
Two Types of Errors in Hypothesis Testing
Hypothesis Testing and Confidence Intervals
Type-II Error, b, and Power
Power
Effect of a on Power
Power and the Value Hypothesized in the Alternative Hypothesis
Methods of Increasing Power
Non-Directional and Directional Alternatives: Two-Tailed versus One- Tailed Tests
Statistical Significance versus Practical Significance
Confidence Limits for the Population Median
Inference Regarding �� when s is not Known: t versus z
The t-Distribution
Confidence Intervals Using the t-Distribution
Accuracy of Confidence Intervals when Sampling Non-Normal Distributions
Inferences about the Difference Between Two Means
Testing Statistical Hypotheses Involving Two Means
The Null Hypotheses
The t-Test for Comparing Two Independent Means
Computing sx1-x2
An Illustration
Confidence Intervals about Mean Differences
Effect Size
t-Test Assumptions and Robustness
Homogeneity of Variance
What if Sample Sizes Are Unequal and Variances Are Heterogeneous: The Welch t' Test
Independence of Observations
Testing H0: ��1 = ��2 with Paired Observations
Direct Difference for the t-Test for Paired Observations
Cautions Regarding the Matched-Pairs Designs in Research
Power when Comparing Means
Non-Parametric Alternatives: The Mann-Whitney Test and the Wilcoxon Signed-Rank Test
Statistics for Categorical Dependent Variables: Inferences about Proportions
Overview
The Proportion as a Mean
The Variance of a Proportion
The Sampling distribution of a Proportion: The Standard Error of p
The Influence of n on sp
Influence of the Sampling Fraction on sp
The Influence of P on sp
Confidence Intervals for P
Quick Confidence Intervals for P
Testing H0: P = K
Testing Empirical versus Theoretical Distributions: Chi-Square Goodness of Fit Test
Testing Differences among Proportions: The Chi-Square Test of Association
Other Formulas for the Chi-Square Test of Association
The C2 Median Test
Chi-Square and the Phi Coefficient
Independence of Observations
Inferences about H0: P1 = P2 when Observations are Paired: McNemar's Test for Correlated Proportions
Inferences about Correlation Coefficient
Testing Statistical Hypotheses Regarding r
Testing H0: r = 0 Using the t-Test
Directional Alternatives: "Two-Tailed" vs. "One- Tailed" Tests
Sampling Distribution of r
The Fisher Z-Transformation
Setting Confidence Intervals for r
Determining Confidence Intervals Graphically
Testing the Difference between Independent Correlation Coefficients: H0: r1 = e2 = ...ej
Averaging r's
Testing Differences between Two Dependent Correlation Coefficients: H0: e31 = r32
Inferences about Other Correlation Coefficients
The Point-Biserial Correlation Coefficient rpr
Spearman's Rank Correlation: H0: ranks = 0
Partial Correlation: H0: r12.3 = 0
Significance of a Multiple Correlation Coefficient
Statistical Significance in Stepwise Multiple Regression
Significance of the Biserial Correlation Coefficient rbis
Significance of the Tetrachoric Correlation Coefficient rtet
Significance of the Correlation Ratio Eta
Testing for Non-linearity of Regression
One-Factor Analysis of Variance
Why Not Several t-Tests?
ANOVA Nomenclature
ANOVA Computation
Sum of Squares Between, SSB
Sum of Squares Within, SSW
ANOVA Computational Illustration
ANOVA Theory
Mean Square Between Groups, MSB
Mean Square Within Groups, MSW
The F-Test
ANOVA with Equal n's
A Statistical Model for the Data
Estimates of the Terms in the Model
Sum of Squares
Restatement of the Null Hypothesis in Terms of Population Means
Degrees of Freedom
Mean Squares: The Expected Value of MSW
The Expected Value of MSB
Some Distribution Theory
The F-Test of the Null Hypothesis: Rationale and Procedure
Type-I versus Type-II Errors: a and b
A Summary of Procedures for One-Factor ANOVA
Consequences of Failure to Meet the ANOVA Assumptions: The "Robustness" of ANOVA
The Welch and Brown-Forsythe Modifications of ANOVA: What Does One Do When <F128>-'s and n's Differ?
The Power of the F-Test
An Illustration
Power When s is Unknown
A Table for Estimating Power When J=2
The Non-Parametric Alternative: The Krukal-Wallis Test
Inferences About Variances
Chi-Square Distributions
Chi-Square Distributions with u1: c 2u
The Chi-Square Distribution with u Degrees of Freedom, c2u
Inferences about the Population Variance: H0: s2 = K
F-Distributions
Inferences about Two Independent Variances: H0: s21 = s22
Testing Homogeneity of Variance: Hartley's Fmax Test
Testing Homogeneity Variance from J Independent Samples: The Bartlett Test
Other Tests of Homogeneity of Variance: The Levene and Brown-Forsythe Tests
Inferences about H0: s21 = s22 with Paired Observations
Relationships among the Normal, t, c2 and F-Distributions
Multiple Comparisons and Trend Analysis
Testing All Pairs of Means: The Studentized Range Statistic, q
The Tukey Method of Multiple Comparisons
The Effect Size of Mean Differences
The Basis for Type-I Error Rate: Contrast vs. Family
The Newman-Keuls Method
The Tukey and Newman-Keuls Methods Compared
The Definition of a Contrast
Simple versus Complex Contrasts
The Standard Error of a Contrast
The t-Ratio for a Contrast
Planned versus Post Hoc Comparisons
Dunn (Bonferroni) Method of Multiple Comparisons
Dunnett Method of Multiple Comparisons
Scheffe Method of Multiple Comparisons
Planned Orthogonal Contrasts
Confidence Intervals for Contrasts
Relative Power of Multiple Comparison Techniques
Trend Analysis
Significance of Trend Components
Relation to Trends to Correlation Coefficients
Assumptions of MC Methods
Multiple Comparisons for Other Statistics
Chapter Summary and Criteria for Selecting a Multiple Comparison Method
Two and Three Factor ANOVA: An Introduction to Factorial Designs
The Meaning of Interaction
Interactions and Generalizability: Factors Do Not Interact
Interactions and Generalizability: Factors Interact
Interpreting when Interaction is Present
Statistical Significance and Interaction
Data Layout and Notation
A Model for the Data
Least-Squares of the Model
Statement of Null Hypotheses
Sums of Squares in the Two-Factor ANOVA
Degrees of Freedom
Mean Squares
Illustration of the Computation for the Two-Factor ANOVA
Expected Values of Mean Squares
The Distribution of the Mean Squares
Determining Power in Factorial Designs
Multiple Comparisons in Factorial ANOVA Designs
Confidence Intervals for Means in Two-Factor ANOVA
Three-Factor ANOVA
Three-Factor ANOVA: An Illustration
Three-Factor ANOVA Computation
The Interpretation of Three-Factor Interaction
Confidence Intervals in Three-Factor ANOVA
How Factorial Designs Increase Power
Factorial ANOVA with Unequal n's
Multi-Factor ANOVA Designs: Random, Mixed, and Fixed Effects
The Random-Effects ANOVA Model
Assumptions of the Random ANOVA Model
An Example
Mean Square, MSW
Mean Square, MSB
The Variance Component, sa2
Confidence Interval for sa2/se2
Summary of Random ANOVA Model
The Mixed-Effects ANOVA Model
Mixed-Model ANOVA Assumptions
Mixed-Model ANOVA Computation
Multiple Comparisons in the Two-Factor Mixed Model
Crossed and Nested Factors
Computation of Sums of Squares for Nested Factors
Determining the Sources of Variation in the ANOVA Table
Degrees of Freedom for Nested Factors
Determining Expected Mean Squares
Error Mean Square in Complex ANOVA Designs
The Incremental Generalization Strategy: Inferential "Concentric Circles."
Model Simplification and Pooling
The Experimental Unit and the Observational Unit
Repeated- Measures ANOVA
A Simple Repeated-Measures ANOVA
Repeated-Measures Assumptions
Trend Analysis on Repeated-Measures Factors
Estimating Reliability via Repeated-Measures ANOVA
Repeated-Measures Designs with a Between-Subjects Factor
Repeated-Measures ANOVA with Two Between-Subjects Factors
Trend Analysis on Between-Subjects Factors
Repeated-Measures ANOVA with Two Within-Subjects Factors and Two Between-Subjects Factors
Repeated-Measures ANOVA vs. MANOVA
An Introduction to the Analysis of Covariance
The Functions of ANCOVA
ANOVA Results
ANCOVA Model
ANCOVA Computations, SStotal
The Adjusted Within Sum of Squares, SS'W
The Adjusted Sum of Squares Between Groups, SS'B
Degrees of Freedom in ANCOVA and the ANCOVA Table
Adjusted Means, Y'j
Confidence Intervals and Multiple Comparisons for Adjusted Means
ANCOVA Illustrated Graphically
ANCOVA Assumptions
ANCOVA Precautions
Covarying and Stratifying
Appendix: Tables
Unit-Normal (z) Distribution
Random Digits
t-Distribution
c2-Distribution
Fisher Z-Transformation
F-Distribution
Power Curves for the F-Test
Hartley's Fmax Distribution
Studentized Range Statistic: q-Distribution
Critical Values of r
Critical Values of rranks, Spearman's Rank Correlation
Critical Values for the Dunn (Bonferroni) t-Statistic
Critical Values for the Dunnett t-Statistic
Coefficients (Orthogonal Polynomials) for Trend Analysis
Binomial Probabilities when P = .5
Glossary of Symbols
Bibliography
Author Index
Subject Index