Preface |
|
xxi | |
Acknowledgments |
|
xxiii | |
About the Author |
|
xxiv | |
Chapter 1 A Gentle Introduction |
|
1 | (42) |
|
How Much Math Do I Need to Do Statistics? |
|
|
2 | (1) |
|
The General Purpose of Statistics: Understanding the World |
|
|
2 | (1) |
|
Another Purpose of Statistics: Making an Argument or a Decision |
|
|
2 | (1) |
|
|
3 | (3) |
|
One Role: The Curious Detective |
|
|
3 | (1) |
|
Another Role: The Honest Attorney |
|
|
4 | (1) |
|
A Final Role: A Good Storyteller |
|
|
5 | (1) |
|
Liberal and Conservative Statisticians |
|
|
6 | (1) |
|
Descriptive and Inferential Statistics |
|
|
7 | (1) |
|
Experiments Are Designed to Test Theories and Hypotheses |
|
|
8 | (1) |
|
|
9 | (1) |
|
|
9 | (2) |
|
Eight Essential Questions of Any Survey or Study |
|
|
11 | (6) |
|
1 Who Was Surveyed or Studied? |
|
|
11 | (1) |
|
2 Why Did the People Participate in the Study? |
|
|
11 | (1) |
|
3 Was There a Control Group, and Did the Control Group Receive a Placebo? |
|
|
12 | (1) |
|
4 How Many People Participated in the Study? |
|
|
13 | (1) |
|
5 How Were the Questions Worded to the Participants in the Study? |
|
|
13 | (2) |
|
6 Was Causation Assumed From a Correlational Study? |
|
|
15 | (1) |
|
7 Who Paid for the Study? |
|
|
15 | (1) |
|
8 Was the Study Published in a Peer-Reviewed Journal? |
|
|
16 | (1) |
|
On Making Samples Representative of the Population |
|
|
17 | (1) |
|
Experimental Design and Statistical Analysis as Controls |
|
|
18 | (1) |
|
The Language of Statistics |
|
|
19 | (1) |
|
On Conducting Scientific Experiments |
|
|
20 | (1) |
|
The Dependent Variable and Measurement |
|
|
20 | (1) |
|
|
21 | (1) |
|
|
21 | (1) |
|
Measurement Scales: The Difference Between Continuous and Discrete Variables |
|
|
22 | (1) |
|
Types of Measurement Scales |
|
|
22 | (2) |
|
|
22 | (1) |
|
|
23 | (1) |
|
|
23 | (1) |
|
|
24 | (1) |
|
Rounding Numbers and Rounding Error |
|
|
24 | (2) |
|
|
26 | (1) |
|
|
26 | (2) |
|
|
28 | (1) |
|
History Trivia: Achenwall to Nightingale |
|
|
29 | (1) |
|
|
30 | (1) |
|
|
31 | (1) |
|
|
32 | (4) |
|
|
36 | (7) |
|
|
37 | (1) |
|
|
37 | (1) |
|
Changing the Display Format for New Numeric Variables |
|
|
38 | (1) |
|
|
39 | (1) |
|
|
39 | (1) |
|
Numeric Variables Versus String Variables |
|
|
40 | (1) |
|
|
41 | (1) |
|
Changing the Folder Where Your Data Are Saved |
|
|
41 | (1) |
|
Opening and Saving Your Data Files |
|
|
42 | (1) |
Chapter 2 Descriptive Statistics: Understanding Distributions of Numbers |
|
43 | (40) |
|
The Purpose of Graphs and Tables: Making Arguments and Decisions |
|
|
44 | (4) |
|
How a Good Graph Stopped a Cholera Epidemic |
|
|
45 | (2) |
|
How Bad Graphs and Tables Contributed to the Space Shuttle Challenger Explosion |
|
|
47 | (1) |
|
How a Poor PowerPoint® Presentation Contributed to the Space Shuttle Columbia Disaster |
|
|
48 | (1) |
|
A Summary of the Purpose of Graphs and Tables |
|
|
48 | (3) |
|
1 Document the Sources of Statistical Data and Their Characteristics |
|
|
48 | (1) |
|
2 Make Appropriate Comparisons |
|
|
49 | (1) |
|
3 Demonstrate the Mechanisms of Cause and Effect and Express the Mechanisms Quantitatively |
|
|
49 | (1) |
|
4 Recognize the Inherent Multivariate Nature of Analytic Problems |
|
|
50 | (1) |
|
5 Inspect and Evaluate Alternative Hypotheses |
|
|
50 | (1) |
|
|
51 | (1) |
|
|
52 | (3) |
|
Shapes of Frequency Distributions |
|
|
55 | (1) |
|
Grouping Data Into Intervals |
|
|
56 | (2) |
|
Advice on Grouping Data Into Intervals |
|
|
58 | (1) |
|
1 Choose Interval Widths That Reduce Your Data to 5 to 10 Intervals |
|
|
58 | (1) |
|
2 Choose the Size of Your Interval Widths Based on Understandable Units, for Example, Multiples of 5 or 10 |
|
|
59 | (1) |
|
3 Make Sure That Your Chosen Intervals Do Not Overlap |
|
|
59 | (1) |
|
The Cumulative Frequency Distribution |
|
|
59 | (1) |
|
Cumulative Percentages, Percentiles, and Quartiles |
|
|
60 | (2) |
|
|
62 | (1) |
|
Non-normal Frequency Distributions |
|
|
63 | (1) |
|
On the Importance of the Shapes of Distributions |
|
|
64 | (1) |
|
Additional Thoughts About Good Graphs Versus Bad Graphs |
|
|
64 | (2) |
|
|
64 | (1) |
|
|
64 | (1) |
|
Changing Scales Midstream (or Mid-Axis) |
|
|
65 | (1) |
|
|
65 | (1) |
|
|
65 | (1) |
|
PowerPoint® Graphs and Presentations |
|
|
66 | (1) |
|
History Trivia: De Moivre to Tukey |
|
|
66 | (2) |
|
|
68 | (1) |
|
|
68 | (1) |
|
|
69 | (4) |
|
|
73 | (10) |
|
Creating a Frequency Distribution |
|
|
73 | (3) |
|
|
76 | (1) |
|
|
77 | (1) |
|
Understanding Skewness and Kurtosis |
|
|
78 | (1) |
|
Describing the Total Autistic Symptoms Data |
|
|
79 | (2) |
|
Describing the Schizoid Personality Disorder Data |
|
|
81 | (2) |
Chapter 3 Statistical Parameters: Measures of Central Tendency and Variation |
|
83 | (25) |
|
Measures of Central Tendency |
|
|
83 | (5) |
|
|
84 | (1) |
|
|
85 | (2) |
|
|
85 | (1) |
|
|
86 | (1) |
|
|
87 | (1) |
|
Choosing Among Measures of Central Tendency |
|
|
88 | (1) |
|
|
89 | (1) |
|
Uncertain or Equivocal Results |
|
|
90 | (1) |
|
|
90 | (3) |
|
|
91 | (1) |
|
|
91 | (2) |
|
Correcting for Bias in the Sample Standard Deviation |
|
|
93 | (1) |
|
How the Square Root of x2 Is Almost Equivalent to Taking the Absolute Value of x |
|
|
93 | (1) |
|
The Computational Formula for Standard Deviation |
|
|
94 | (1) |
|
|
94 | (1) |
|
The Sampling Distribution of Means, the Central Limit Theorem, and the Standard Error of the Mean |
|
|
95 | (1) |
|
The Use of the Standard Deviation for Prediction |
|
|
96 | (1) |
|
Practical Uses of the Empirical Rule: As a Definition of an Outlier |
|
|
97 | (1) |
|
Practical Uses of the Empirical Rule: Prediction and IQ Tests |
|
|
97 | (1) |
|
|
98 | (1) |
|
History Trivia: Fisher to Eels |
|
|
98 | (1) |
|
|
99 | (1) |
|
|
99 | (1) |
|
|
100 | (3) |
|
|
103 | (5) |
|
Generating Central Tendency and Variation Statistics |
|
|
103 | (5) |
Chapter 4 Standard Scores, the z Distribution, and Hypothesis Testing |
|
108 | (35) |
|
|
109 | (1) |
|
The Classic Standard Score: The z Score and the z Distribution |
|
|
110 | (1) |
|
|
111 | (1) |
|
More Practice on Converting Raw Data Into z Scores |
|
|
111 | (2) |
|
Converting z Scores to Other Types of Standard Scores |
|
|
113 | (2) |
|
|
115 | (1) |
|
Interpreting Negative z Scores |
|
|
116 | (1) |
|
Testing the Predictions of the Empirical Rule With the z Distribution |
|
|
116 | (1) |
|
Why Is the z Distribution so Important? |
|
|
117 | (1) |
|
How We Use the z Distribution to Test Experimental Hypotheses |
|
|
117 | (1) |
|
More Practice With the z Distribution and TScores |
|
|
118 | (13) |
|
Example 1: Finding the Area in a z Distribution That Falls Above a Known Score Where the Known Score Is Above the Mean |
|
|
118 | (1) |
|
Example 2: Finding the Area in a z Distribution That Falls Below a Known Score Where the Known Score Is Above the Mean |
|
|
119 | (3) |
|
Example 3: Finding the Area in a z Distribution That Falls Below a Known Score Where the Known Score Is Below the Mean |
|
|
122 | (1) |
|
Example 4: Finding The Area in a z Distribution That Falls Above a Known Score Where the Known Score Is Below the Mean |
|
|
123 | (2) |
|
Example 5: Finding The Area in a z Distribution That Falls Between Two Known Scores Where Both Known Scores Are Above the Mean |
|
|
125 | (2) |
|
Example 6: Finding The Area in a z Distribution That Falls Between Two Known Scores Where One Known Score Is Above the Mean and One Is Below the Mean |
|
|
127 | (1) |
|
Example 7: Finding the Area in a z Distribution That Falls Between Two Known Scores |
|
|
128 | (3) |
|
Summarizing Scores Through Percentiles |
|
|
131 | (1) |
|
History Trivia: Karl Pearson to Egon Pearson |
|
|
132 | (2) |
|
|
134 | (1) |
|
|
134 | (1) |
|
|
135 | (2) |
|
|
137 | (6) |
|
Transforming Raw Scores Into z Scores |
|
|
138 | (2) |
|
Transforming z Scores Into T Scores |
|
|
140 | (3) |
Chapter 5 Inferential Statistics: The Controlled Experiment, Hypothesis Testing, and the z Distribution |
|
143 | (38) |
|
Hypothesis Testing in the Controlled Experiment |
|
|
145 | (1) |
|
Hypothesis Testing: The Big Decision |
|
|
146 | (1) |
|
How the Big Decision Is Made: Back to the z Distribution |
|
|
146 | (2) |
|
The Parameter of Major Interest in Hypothesis Testing: The Mean |
|
|
148 | (1) |
|
Nondirectional and Directional Alternative Hypotheses |
|
|
149 | (1) |
|
A Debate: Retain the Null Hypothesis or Fait to Reject the Null Hypothesis |
|
|
149 | (1) |
|
The Null Hypothesis as a Nonconservative Beginning |
|
|
150 | (1) |
|
The Four Possible Outcomes in Hypothesis Testing |
|
|
151 | (1) |
|
1 Correct Decision: Retain Ho When Ho Is Actually True |
|
|
151 | (1) |
|
2 Type I Error: Reject Ho When Ho Is Actually True |
|
|
151 | (1) |
|
3 Correct Decision: Reject Ho When Ho Is Actually False |
|
|
151 | (1) |
|
4 Type II Error: Retain Ho When Ho Is Actually False |
|
|
152 | (1) |
|
|
152 | (1) |
|
Significant and Nonsignificant Findings |
|
|
152 | (1) |
|
Trends, and Does God Really Love the 05 Level of Significance More Than the .06 Level? |
|
|
153 | (1) |
|
Directional or Nondirectional Alternative Hypotheses: Advantages and Disadvantages |
|
|
154 | (5) |
|
Did Nuclear Fusion Occur? |
|
|
154 | (1) |
|
|
155 | (1) |
|
How Reliable Is the Source of the Claim? |
|
|
155 | (1) |
|
Does This Source Often Make Similar Claims? |
|
|
156 | (1) |
|
Have the Claims Been Verified by Another Source? |
|
|
156 | (1) |
|
How Does the Claim Fit With Known Natural Scientific Laws? |
|
|
157 | (1) |
|
Can the Claim Be Disproven, or Has Only Supportive Evidence Been Sought? |
|
|
158 | (1) |
|
Do the Claimants' Personal Beliefs and Biases Drive Their Conclusions or Vice Versa? |
|
|
159 | (1) |
|
Conclusions About Science and Pseudoscience |
|
|
159 | (1) |
|
The Most Critical Elements in the Detection of Baloney in Suspicious Studies and Fraudulent Claims |
|
|
160 | (1) |
|
Can Statistics Solve Every Problem? |
|
|
161 | (1) |
|
|
161 | (7) |
|
|
161 | (1) |
|
The Definition of the Probability of an Event |
|
|
162 | (1) |
|
The Multiplication Theorem of Probability |
|
|
162 | (1) |
|
Combinations Theorem of Probability |
|
|
163 | (1) |
|
Permutations Theorem of Probability |
|
|
164 | (3) |
|
|
167 | (1) |
|
|
167 | (1) |
|
|
168 | (1) |
|
|
168 | (1) |
|
History Trivia: Egon Pearson to Karl Pearson |
|
|
168 | (2) |
|
|
170 | (1) |
|
|
170 | (1) |
|
|
170 | (4) |
|
|
174 | (7) |
|
Removing a Case From a Data Set |
|
|
174 | (1) |
|
|
174 | (1) |
|
|
175 | (1) |
|
|
176 | (1) |
|
|
176 | (1) |
|
Selecting a Particular Condition for Analysis Within a Data Set |
|
|
177 | (2) |
|
Copying Selected Cases or Conditions to a New Data Set |
|
|
179 | (2) |
Chapter 6 An Introduction to Correlation and Regression |
|
181 | (49) |
|
Correlation: Use and Abuse |
|
|
183 | (2) |
|
A Warning: Correlation Does Not Imply Causation |
|
|
185 | (3) |
|
1 Marijuana Use and Heroin Use Are Positively Correlated |
|
|
186 | (1) |
|
2 Milk Use Is Positively Correlated to Cancer Rates |
|
|
186 | (1) |
|
3 Weekly Church Attendance Is Negatively Correlated With Drug Abuse |
|
|
186 | (1) |
|
4 Lead Levels Are Positively Correlated With Antisocial Behavior |
|
|
187 | (1) |
|
5 The Risk of Getting Alzheimer's Disease Is Negatively Correlated With Smoking Cigarettes |
|
|
187 | (1) |
|
6 Sexual Activity Is Negatively Correlated With Increases in Education |
|
|
187 | (1) |
|
7 An Active Sex Life Is Positively Correlated With Longevity |
|
|
187 | (1) |
|
8 Coffee Drinking Is Negatively Correlated With Suicidal Risk |
|
|
188 | (1) |
|
9 Excessive Drinking and Smoking Causes Women to Be Abused |
|
|
188 | (1) |
|
Another Warning: Chance Is Lumpy |
|
|
188 | (1) |
|
Correlation and Prediction |
|
|
189 | (1) |
|
The Four Common Types of Correlation |
|
|
189 | (1) |
|
The Pearson Product-Moment Correlation Coefficient |
|
|
189 | (3) |
|
Testing for the Significance of a Correlation Coefficient |
|
|
192 | (1) |
|
Obtaining the Critical Values of the t Distribution |
|
|
193 | (2) |
|
Step 1: Choose a One-Tailed or Two-Tailed Test of Significance |
|
|
194 | (1) |
|
Step 2: Choose the Level of Significance |
|
|
194 | (1) |
|
Step 3: Determine the Degrees of Freedom (df) |
|
|
194 | (1) |
|
Step 4: Determine Whether the t from the Formula (Called the Derived t) Exceeds the Tabled Critical Values From the t Distribution |
|
|
194 | (1) |
|
If the Null Hypothesis Is Rejected |
|
|
195 | (1) |
|
Representing the Pearson Correlation Graphically: The Scatterplot |
|
|
195 | (1) |
|
Fitting the Points With a Straight Line: The Assumption of a Linear Relationship |
|
|
196 | (1) |
|
Interpretation of the Slope of the Best-Fitting Line |
|
|
197 | (1) |
|
The Assumption of Homoscedasticity |
|
|
198 | (1) |
|
The Coefficient of Determination: How Much One Variable Accounts for Variation in Another Variable-The Interpretation of r2 |
|
|
199 | (1) |
|
Quirks in the Interpretation of Significant and Nonsignificant Correlation Coefficients |
|
|
200 | (1) |
|
|
201 | (1) |
|
Reading the Regression Line |
|
|
202 | (5) |
|
The World Is a Complex Place: Any Single Behavior Is Most Often Caused by Multiple Variables |
|
|
203 | (1) |
|
|
204 | (1) |
|
|
205 | (1) |
|
|
205 | (2) |
|
Final Thoughts About Multiple Regression Analyses: A Warning About the Interpretation of the Significant Beta Coefficients |
|
|
207 | (1) |
|
|
208 | (1) |
|
Significance Test for Spearman's r |
|
|
209 | (1) |
|
|
210 | (1) |
|
Point-Biserial Correlation |
|
|
211 | (3) |
|
Testing for the Significance of the Point-Biserial Correlation Coefficient |
|
|
214 | (1) |
|
|
215 | (1) |
|
Testing for the Significance of Phi |
|
|
216 | (1) |
|
History Trivia: Galton to Fisher |
|
|
217 | (1) |
|
|
218 | (1) |
|
|
219 | (1) |
|
|
219 | (5) |
|
|
224 | (6) |
|
Analyzing the Pearson Product-Moment Correlation |
|
|
224 | (1) |
|
|
225 | (3) |
|
Using the Paste Function in the Syntax Editor |
|
|
228 | (2) |
Chapter 7 The t Test for Independent Groups |
|
230 | (27) |
|
The Statistical Analysis of the Controlled Experiment |
|
|
230 | (1) |
|
One t Test but Two Designs |
|
|
231 | (2) |
|
Assumptions of the Independent tTest |
|
|
232 | (1) |
|
|
232 | (1) |
|
Normality of the Dependent Variable |
|
|
233 | (1) |
|
|
233 | (1) |
|
The Formula for the Independent t Test |
|
|
233 | (1) |
|
You Must Remember This! An Overview of Hypothesis Testing With the tTest |
|
|
234 | (1) |
|
What Does the t Test Do? Components of the t Test Formula |
|
|
234 | (1) |
|
What If the Two Variances Are Radically Different From One Another? |
|
|
235 | (1) |
|
|
235 | (5) |
|
Steps in the tTest Formula |
|
|
236 | (2) |
|
Testing the Null Hypothesis |
|
|
238 | (1) |
|
Steps in Determining Significance |
|
|
238 | (1) |
|
When Ho Has Been Rejected |
|
|
239 | (1) |
|
|
240 | (1) |
|
The Power of a Statistical Test |
|
|
240 | (1) |
|
|
241 | (1) |
|
The Correlation Coefficient of Effect Size |
|
|
241 | (1) |
|
Another Measure of Effect Size: Cohen's d |
|
|
242 | (1) |
|
|
243 | (3) |
|
Estimating the Standard Error |
|
|
246 | (2) |
|
History Trivia: Gosset and Guinness Brewery |
|
|
248 | (1) |
|
|
248 | (1) |
|
|
249 | (1) |
|
|
249 | (4) |
|
|
253 | (4) |
|
Conducting a t Test for Independent Groups |
|
|
253 | (1) |
|
Interpreting a t Test for Independent Groups |
|
|
254 | (1) |
|
Conducting a tTest for Independent Groups for a Different Variable |
|
|
255 | (1) |
|
Interpreting a t Test for Independent Groups for a Different Variable |
|
|
256 | (1) |
Chapter 8 The t Test for Dependent Groups |
|
257 | (24) |
|
Variations on the Controlled Experiment |
|
|
257 | (2) |
|
|
258 | (1) |
|
|
258 | (1) |
|
|
258 | (1) |
|
|
258 | (1) |
|
|
259 | (3) |
|
|
259 | (1) |
|
Assumptions of the Dependent t Test |
|
|
259 | (1) |
|
Why the Dependent t Test May Be More Powerful Than the Independent t Test |
|
|
259 | (1) |
|
How to Increase the Power of a t Test |
|
|
260 | (1) |
|
Drawbacks of the Dependent t Test Designs |
|
|
260 | (1) |
|
One-Tailed or Two-Tailed Tests of Significance |
|
|
261 | (1) |
|
Hypothesis Testing and the Dependent t Test: Design 1 |
|
|
261 | (1) |
|
Design 1 (Same Participants or Repeated Measures): A Computational Example |
|
|
262 | (4) |
|
Determination of Effect Size |
|
|
265 | (1) |
|
Design 2 (Matched Pairs): A Computational Example |
|
|
266 | (3) |
|
Determination of Effect Size |
|
|
269 | (1) |
|
Design 3 (Same Participants and Balanced Presentation): A Computational Example |
|
|
269 | (4) |
|
Determination of Effect Size |
|
|
272 | (1) |
|
History Trivia: Fisher to Pearson |
|
|
273 | (1) |
|
|
273 | (1) |
|
|
274 | (1) |
|
|
274 | (4) |
|
|
278 | (3) |
|
Conducting a t Test for Dependent Groups |
|
|
278 | (1) |
|
Interpreting a t Test for Dependent Groups |
|
|
279 | (2) |
Chapter 9 Analysis of Variance (ANOVA): One-Factor Completely Randomized Design |
|
281 | (23) |
|
A Limitation of Multiple t Tests and a Solution |
|
|
282 | (1) |
|
The Equally Unacceptable Bonferroni Solution |
|
|
282 | (1) |
|
The Acceptable Solution: An Analysis of Variance |
|
|
282 | (1) |
|
The Null and Alternative Hypotheses in ANOVA |
|
|
283 | (1) |
|
The Beauty and Elegance of the FTest Statistic |
|
|
283 | (1) |
|
|
284 | (1) |
|
How Can There Be Two Different Estimates of Within-Groups Variance? |
|
|
285 | (1) |
|
|
286 | (1) |
|
|
287 | (1) |
|
|
287 | (1) |
|
What a Significant ANOVA Indicates |
|
|
288 | (1) |
|
|
288 | (3) |
|
Degrees of Freedom for the Numerator |
|
|
291 | (1) |
|
Degrees of Freedom for the Denominator |
|
|
291 | (1) |
|
Determining Effect Size in ANOVA: Omega Squared |
|
|
292 | (1) |
|
Another Measure of Effect Size: Eta |
|
|
293 | (1) |
|
History Trivia: Gosset to Fisher |
|
|
294 | (2) |
|
|
296 | (1) |
|
|
296 | (1) |
|
|
297 | (4) |
|
|
301 | (3) |
|
|
301 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
301 | (1) |
|
Conducting a One-Factor Completely Randomized ANOVA |
|
|
301 | (2) |
|
Interpreting a One-Factor Completely Randomized ANOVA |
|
|
303 | (1) |
Chapter 10 After a Significant ANOVA: Multiple Comparison Tests |
|
304 | (19) |
|
Conceptual Overview of Tukey's Test |
|
|
305 | (1) |
|
Computation of Tukey's HSD Test |
|
|
305 | (3) |
|
What to Do If the Number of Error Degrees of Freedom Is Not Listed in the Table of Tukey's q Values |
|
|
308 | (1) |
|
Determining What It All Means |
|
|
308 | (1) |
|
|
309 | (1) |
|
On the Importance of Nonsignificant Mean Differences |
|
|
310 | (1) |
|
|
310 | (1) |
|
|
311 | (1) |
|
|
311 | (1) |
|
|
311 | (1) |
|
|
312 | (1) |
|
|
312 | (3) |
|
|
315 | (8) |
|
|
315 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
315 | (1) |
|
Conducting a Multiple Comparison Test |
|
|
316 | (2) |
|
Interpreting a Multiple Comparison Test |
|
|
318 | (1) |
|
Conducting a Multiple Comparison Test for Another Variable |
|
|
319 | (2) |
|
Interpreting a Multiple Comparison Test for Another Variable |
|
|
321 | (2) |
Chapter 11 Analysis of Variance (ANOVA): One-Factor Repeated-Measures Design |
|
323 | (16) |
|
The Repeated-Measures ANOVA |
|
|
323 | (1) |
|
Assumptions of the One-Factor Repeated-Measures ANOVA |
|
|
324 | (1) |
|
|
324 | (4) |
|
Determining Effect Size in ANOVA |
|
|
328 | (1) |
|
|
329 | (1) |
|
|
329 | (1) |
|
|
330 | (2) |
|
|
332 | (7) |
|
|
332 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
333 | (1) |
|
Conducting a One-Factor Repeated-Measures Design ANOVA |
|
|
333 | (3) |
|
Interpreting a One-Factor Repeated-Measures Design ANOVA |
|
|
336 | (3) |
Chapter 12 Factorial ANOVA: Two-Factor Completely Randomized Design |
|
339 | (16) |
|
|
339 | (1) |
|
The Most Important Feature of a Factorial Design: The Interaction |
|
|
340 | (1) |
|
Fixed and Random Effects and In Situ Designs |
|
|
340 | (1) |
|
The Null Hypotheses in a Two-Factor ANOVA |
|
|
341 | (1) |
|
Assumptions and Unequal Numbers of Participants |
|
|
341 | (1) |
|
|
341 | (6) |
|
Computation of the First Main Effect |
|
|
343 | (1) |
|
Computation of the Second Main Effect |
|
|
343 | (1) |
|
Computation of the Interaction Between the Two Main Effects |
|
|
344 | (2) |
|
Interpretation of the Results |
|
|
346 | (1) |
|
|
347 | (1) |
|
|
347 | (1) |
|
|
348 | (3) |
|
|
351 | (4) |
|
|
351 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
352 | (1) |
|
Conducting an ANOVA Two-Factor Completely Randomized Design |
|
|
352 | (1) |
|
Interpreting an ANOVA Two-Factor Completely Randomized Design |
|
|
353 | (2) |
Chapter 13 Post Hoc Analysis of Factorial ANOVA |
|
355 | (26) |
|
Main Effect Interpretation: Gender |
|
|
355 | (1) |
|
Why a Multiple Comparison Test Is Unnecessary for a Two-Level Main Effect, and When Is a Multiple Comparison Test Necessary? |
|
|
356 | (1) |
|
|
357 | (1) |
|
Multiple Comparison Test for the Main Effect for Age |
|
|
358 | (2) |
|
Warning: Limit Your Main Effect Conclusions When the Interaction Is Significant |
|
|
360 | (1) |
|
Multiple Comparison Tests |
|
|
360 | (1) |
|
Interpretation of the Interaction Effect |
|
|
361 | (4) |
|
|
364 | (1) |
|
|
365 | (1) |
|
ADHD Men Versus ADHD Women |
|
|
365 | (1) |
|
|
365 | (1) |
|
Writing Up the Results Journal Style |
|
|
365 | (1) |
|
|
366 | (1) |
|
Exploring the Possible Outcomes in a Two-Factor ANOVA |
|
|
366 | (2) |
|
Determining Effect Size in a Two-Factor ANOVA |
|
|
368 | (1) |
|
History Trivia: Fisher and Smoking |
|
|
369 | (1) |
|
|
370 | (1) |
|
|
371 | (1) |
|
|
371 | (3) |
|
|
374 | (7) |
|
|
374 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
374 | (1) |
|
Conducting a Post Hoc Analysis of Factorial ANOVA |
|
|
375 | (2) |
|
Interpreting a Post Hoc Analysis of Factorial ANOVA for the Main Effect |
|
|
377 | (1) |
|
Conducting a Post Hoc Analysis of a Significant Interaction in Factorial ANOVA With a Group Variable |
|
|
378 | (1) |
|
Interpreting a Post Hoc Analysis of a Significant Interaction in Factorial ANOVA With a Group Variable |
|
|
379 | (2) |
Chapter 14 Factorial ANOVA: Additional Designs |
|
381 | (31) |
|
|
381 | (1) |
|
Overview of the Split-Plot ANOVA |
|
|
382 | (1) |
|
|
382 | (7) |
|
Main Effect: Social Facilitation |
|
|
388 | (1) |
|
|
388 | (1) |
|
Interaction: Social Facilitation x Trials |
|
|
389 | (1) |
|
Two-Factor ANOVA: Repeated Measures on Both Factors Design |
|
|
389 | (1) |
|
Overview of the Repeated-Measures ANOVA |
|
|
389 | (1) |
|
|
390 | (8) |
|
|
398 | (1) |
|
|
398 | (1) |
|
|
399 | (3) |
|
|
402 | (10) |
|
|
402 | (1) |
|
Downloading the Data Set to Your Desktop |
|
|
403 | (1) |
|
Conducting a Split-Plot ANOVA |
|
|
403 | (3) |
|
A Second Two-Factor ANOVA Design in SPSS |
|
|
406 | (1) |
|
|
407 | (1) |
|
Conducting a Repeated-Measures ANOVA |
|
|
408 | (4) |
Chapter 15 Nonparametric Statistics: The Chi-Square Test and Other Nonparametric Tests |
|
412 | (30) |
|
Overview of the Purpose of Chi-Square |
|
|
413 | (1) |
|
Overview of Chi-Square Designs |
|
|
414 | (1) |
|
Chi-Square Test: Two-Cell Design (Equal Probabilities Type) |
|
|
414 | (3) |
|
Computation of the Two-Cell Design |
|
|
415 | (2) |
|
The Chi-Square Distribution |
|
|
417 | (1) |
|
Assumptions of the Chi-Square Test |
|
|
417 | (1) |
|
Chi-Square Test: Two-Cell Design (Different Probabilities Type) |
|
|
418 | (2) |
|
Computation of the Two-Cell Design |
|
|
418 | (2) |
|
Interpreting a Significant Chi-Square Test for a Newspaper |
|
|
420 | (1) |
|
Chi-Square Test: Three-Cell Experiment (Equal Probabilities Type) |
|
|
420 | (2) |
|
Computation of the Three-Cell Design |
|
|
420 | (2) |
|
Chi-Square Test: Two-by-Two Design |
|
|
422 | (3) |
|
Computation of the Two-by-Two Design |
|
|
423 | (2) |
|
What to Do After a Chi-Square Test Is Significant |
|
|
425 | (1) |
|
When Cell Frequencies Are Less Than 5 Revisited |
|
|
426 | (6) |
|
Other Nonparametric Tests |
|
|
427 | (1) |
|
|
427 | (3) |
|
Wilcoxon Test for Two Dependent Groups |
|
|
430 | (2) |
|
History Trivia: Pearson and Biometrika |
|
|
432 | (1) |
|
|
432 | (1) |
|
|
433 | (1) |
|
|
433 | (3) |
|
|
436 | (6) |
|
Building a Data Set for a Chi-Square Test |
|
|
436 | (2) |
|
Conducting a Chi-Square Test |
|
|
438 | (2) |
|
Interpreting a Chi-Square Test |
|
|
440 | (2) |
Chapter 16 Other Statistical Topics, Parameters, and Tests |
|
442 | (18) |
|
|
443 | (1) |
|
Health Science Statistics |
|
|
443 | (7) |
|
|
443 | (4) |
|
|
447 | (1) |
|
Parameters of Mortality and Morbidity |
|
|
448 | (2) |
|
Additional Statistical Analyses and Multivariate Statistics |
|
|
450 | (5) |
|
|
450 | (1) |
|
Multivariate Analysis of Variance |
|
|
450 | (1) |
|
Multivariate Analysis of Covariance |
|
|
451 | (1) |
|
|
451 | (1) |
|
|
452 | (1) |
|
Structural Equation Modeling |
|
|
453 | (1) |
|
|
453 | (1) |
|
|
454 | (1) |
|
Linear Discriminant Function Analysis |
|
|
454 | (1) |
|
A Summary of Multivariate Statistics |
|
|
455 | (1) |
|
|
456 | (1) |
|
|
456 | (1) |
|
|
457 | (1) |
|
|
457 | (3) |
Appendix A: z Distribution |
|
460 | (14) |
Appendix B: t Distribution |
|
474 | (2) |
Appendix C: Spearman's Correlation |
|
476 | (1) |
Appendix D: Chi-Square Distribution |
|
477 | (2) |
Appendix E: F Distribution |
|
479 | (6) |
Appendix F: Tukey's Table |
|
485 | (2) |
Appendix G: Mann-Whitney U Critical Values |
|
487 | (3) |
Appendix H: Wilcoxon Signed-Rank Test Critical Values |
|
490 | (1) |
Appendix I: Answers to Odd-Numbered Test Yourself Questions |
|
491 | (3) |
Glossary |
|
494 | (8) |
References |
|
502 | |
Index |
|
50 | |