100,99 €
Features an integrated approach of statistical scenarios and simulations to aid readers in developing key intuitions needed to understand the wide ranging concepts and methods of statistics and inference
Illuminating Statistical Analysis Using Scenarios and Simulations presents the basic concepts of statistics and statistical inference using the dual mechanisms of scenarios and simulations. This approach helps readers develop key intuitions and deep understandings of statistical analysis. Scenario-specific sampling simulations depict the results that would be obtained by a very large number of individuals investigating the same scenario, each with their own evidence, while graphical depictions of the simulation results present clear and direct pathways to intuitive methods for statistical inference. These intuitive methods can then be easily linked to traditional formulaic methods, and the author does not simply explain the linkages, but rather provides demonstrations throughout for a broad range of statistical phenomena. In addition, induction and deduction are repeatedly interwoven, which fosters a natural "need to know basis" for ordering the topic coverage.
Examining computer simulation results is central to the discussion and provides an illustrative way to (re)discover the properties of sample statistics, the role of chance, and to (re)invent corresponding principles of statistical inference. In addition, the simulation results foreshadow the various mathematical formulas that underlie statistical analysis.
In addition, this book:
• Features both an intuitive and analytical perspective and includes a broad introduction to the use of Monte Carlo simulation and formulaic methods for statistical analysis
• Presents straight-forward coverage of the essentials of basic statistics and ensures proper understanding of key concepts such as sampling distributions, the effects of sample size and variance on uncertainty, analysis of proportion, mean and rank differences, covariance, correlation, and regression
• Introduces advanced topics such as Bayesian statistics, data mining, model cross-validation, robust regression, and resampling
• Contains numerous example problems in each chapter with detailed solutions as well as an appendix that serves as a manual for constructing simulations quickly and easily using Microsoft® Office Excel®
Illuminating Statistical Analysis Using Scenarios and Simulations is an ideal textbook for courses, seminars, and workshops in statistics and statistical inference and is appropriate for self-study as well. The book also serves as a thought-provoking treatise for researchers, scientists, managers, technicians, and others with a keen interest in statistical analysis.
Jeffrey E. Kottemann, Ph.D., is Professor in the Perdue School at Salisbury University. Dr. Kottemann has published articles in a wide variety of academic research journals in the fields of business administration, computer science, decision sciences, economics, engineering, information systems, psychology, and public administration. He received his Ph.D. in Systems and Quantitative Methods from the University of Arizona.
Sie lesen das E-Book in den Legimi-Apps auf:
Seitenzahl: 288
Veröffentlichungsjahr: 2017
Cover
Title Page
Copyright
Preface
Acknowledgements
Part I: Sample Proportions and the Normal Distribution
Chapter 1: Evidence and Verdicts
Chapter 2: Judging Coins I
Chapter 3: Brief on Bell Shapes
Chapter 4: Judging Coins II
Chapter 5: Amount of Evidence I
Chapter 6: Variance of Evidence I
Chapter 7: Judging Opinion Splits I
Chapter 8: Amount of Evidence II
Chapter 9: Variance of Evidence II
Chapter 10: Judging Opinion Splits II
Chapter 11: It Has Been the Normal Distribution All Along
A Note on Stricter Thresholds for Type I Error
Chapter 12: Judging Opinion Split Differences
Chapter 13: Rescaling to Standard Errors
Chapter 14: The Standardized Normal Distribution Histogram
Chapter 15: The z-Distribution
Chapter 16: Brief on Two-Tail Versus One-Tail
Chapter 17: Brief on Type I Versus Type II Errors
The Bigger Picture
Part II: Sample Means and the Normal Distribution
Chapter 18: Scaled Data and Sample Means
Chapter 19: Distribution of Random Sample Means
Chapter 20: Amount of Evidence
Chapter 21: Variance of Evidence
Variance and Standard Deviation
Chapter 22: Homing in on the Population Mean I
Chapter 23: Homing in on the Population Mean II
Chapter 24: Homing in on the Population Mean III
Chapter 25: Judging Mean Differences
Chapter 26: Sample Size, Variance, and Uncertainty
Chapter 27: The t-Distribution
Part III: Multiple Proportions and Means: The X2- and F-Distributions
Chapter 28: Multiple Proportions and the X2-Distribution
Chapter 29: Facing Degrees of Freedom
Chapter 30: Multiple Proportions: Goodness of Fit
A Note on Using Chi-squared to Test the Distribution of a Scaled Variable
Chapter 31: Two-Way Proportions: Homogeneity
Chapter 32: Two-Way Proportions: Independence
Chapter 33: Variance Ratios and the F-Distribution
Chapter 34: Multiple Means and Variance Ratios: ANOVA
Chapter 35: Two-Way Means and Variance Ratios: ANOVA
Part IV: Linear Associations: Covariance, Correlation, and Regression
Chapter 36: Covariance
Chapter 37: Correlation
Chapter 38: What Correlations Happen Just by Chance?
Special Considerations: Confidence Intervals for Sample Correlations
Chapter 39: Judging Correlation Differences
Special Considerations: Sample Correlation Differences
Chapter 40: Correlation with Mixed Data Types
Chapter 41: A Simple Regression Prediction Model
Chapter 42: Using Binomials Too
Getting More Sophisticated #1
Getting More Sophisticated #2
Chapter 43: A Multiple Regression Prediction Model
Getting More Sophisticated
Chapter 44: Loose End I (Collinearity)
Chapter 45: Loose End II (Squaring R)
Chapter 46: Loose End III (Adjusting R-Squared)
Chapter 47: Reality Strikes
Part V: Dealing with Unruly Scaled Data
Chapter 48: Obstacles and Maneuvers
Chapter 49: Ordered Ranking Maneuver
Chapter 50: What Rank Sums Happen Just by Chance?
Chapter 51: Judging Rank Sum Differences
Chapter 52: Other Methods Using Ranks
Chapter 53: Transforming the Scale of Scaled Data
Chapter 54: Brief on Robust Regression
Chapter 55: Brief on Simulation and Resampling
Part VI: Review and Additional Concepts
Chapter 56: For Part I
Chapter 57: For Part II
Chapter 58: For Part III
Chapter 59: For Part IV
Chapter 60: For Part V
Appendices
A: Data Types and Some Basic Statistics
Some Basic Statistics (Primarily for Scaled and Binomial Variables)
B: Simulating Statistical Scenarios
Random Variation
General Guidelines
Scenario-Specific Instructions
C: Standard Error as Standard Deviation
D: Data Excerpt
E: Repeated Measures
F: Bayesian Statistics
A Note on Priors
Getting More Sophisticated
G: Data Mining
Index
End User License Agreement
Table E.1
Table E.2
Table F.1
Table F.2
Table F.3
Table F.4
Table F.5
Table G.1
Table 1.1
Table 2.1
Table 4.1
Table 4.2
Table 4.3
Table 4.4
Table 7.1
Table 8.1
Table 8.2
Table 9.1
Table 10.1
Table 12.1
Table 12.2
Table 13.1
Table 13.2
Table 15.1
Table 15.2
Table 15.3
Table 17.1
Table 17.2
Table 17.3
Table 17.4
Table 26.1
Table 27.1
Table 28.1
Table 30.1
Table 31.1
Table 31.2
Table 31.3
Table 32.1
Table 32.2
Table 32.3
Table 32.4
Table 32.5
Table 33.1
Table 33.2
Table 35.1
Table 35.2
Table 35.3
Table 35.4
Table 35.5
Table 35.6
Table 35.7
Table 42.1
Table 42.2
Table 43.1
Table 44.1
Table 44.2
Table 44.3
Table 44.4
Table 44.5
Table 47.1
Table 47.2
Table 49.1
Table 51.1
Table 51.2
Table 56.1
Table 56.2
Table 57.1
Table 57.2
Table 58.1
Table 58.2
Table 58.3
Table 58.4
Table 58.5
Table 58.6
Table 58.7
Table 59.1
Table 59.2
Table 59.3
Table 59.4
Table 59.5
Table 59.6
Table 59.7
Table 60.1
Table 60.2
Table 60.3
Figure A.1
Figure B.1
Figure B.2
Figure C.1
Figure E.1
Figure F.1
Figure G.1
Figure G.2
Figure G.3
Figure G.4
Figure G.5
Figure G.6
Figure G.7
Figure G.8
Figure 2.1
Figure 2.2
Figure 3.1
Figure 3.2
Figure 4.1
Figure 4.2
Figure 4.3
Figure 4.4
Figure 4.5
Figure 5.1
Figure 5.2
Figure 5.3
Figure 5.4
Figure 5.5
Figure 6.1
Figure 6.2
Figure 6.3
Figure 7.1
Figure 8.1
Figure 9.1
Figure 9.2
Figure 9.3
Figure 10.1
Figure 12.1
Figure 14.1
Figure 14.2
Figure 14.3
Figure 15.1
Figure 15.2
Figure 15.3
Figure 15.4
Figure 16.1
Figure 16.2
Figure 16.3
Figure 17.1
Figure 17.2
Figure 17.3
Figure 19.1
Figure 20.1
Figure 20.2
Figure 21.1
Figure 21.2
Figure 22.1
Figure 23.1
Figure 24.1
Figure 24.2
Figure 25.1
Figure 25.2
Figure 26.1
Figure 26.2
Figure 26.3
Figure 26.4
Figure 26.5
Figure 26.6
Figure 27.1
Figure 27.2
Figure 27.3
Figure 27.4
Figure 27.5
Figure 28.1
Figure 28.2
Figure 28.3
Figure 28.4
Figure 28.5
Figure 28.6
Figure 30.1
Figure 33.1
Figure 33.2
Figure 33.3
Figure 33.4
Figure 34.1
Figure 34.2
Figure 34.3
Figure 34.4
Figure 34.5
Figure 36.1
Figure 36.2
Figure 36.3
Figure 36.4
Figure 38.1
Figure 38.2
Figure 38.3
Figure 38.4
Figure 38.5
Figure 38.6
Figure 39.1
Figure 39.2
Figure 39.3
Figure 40.1
Figure 41.1
Figure 41.2
Figure 42.1
Figure 43.1
Figure 43.2
Figure 45.1
Figure 45.2
Figure 47.1
Figure 47.2
Figure 47.3
Figure 47.4
Figure 48.1
Figure 48.2
Figure 48.3
Figure 48.4
Figure 50.1
Figure 53.1
Figure 56.1
Figure 57.1
Figure 57.2
Figure 59.1
Figure 59.2
Figure 59.3
Figure 59.4
Figure 59.5
Figure 59.6
Figure 60.1
Cover
Table of Contents
Begin Reading
Part 1
Chapter 1
i
ii
iii
iv
ix
x
xi
xii
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
Jeffrey E Kottemann Ph.D.
Copyright © 2017 by John Wiley & Sons, Inc. All rights reserved
Published by John Wiley & Sons, Inc., Hoboken, New JerseyPublished simultaneously in Canada
No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning, or otherwise, except as permitted under Section 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 750-4470, or on the web at www.copyright.com. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permission.
Limit of Liability/Disclaimer of Warranty: While the publisher and author have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives or written sales materials. The advice and strategies contained herein may not be suitable for your situation. You should consult with a professional where appropriate. Neither the publisher nor author shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages.
For general information on our other products and services or for technical support, please contact our Customer Care Department within the United States at (800) 762-2974, outside the United States at (317) 572-3993 or fax (317) 572-4002.
Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic formats. For more information about Wiley products, visit our web site at www.wiley.com.
Library of Congress Cataloging-in-Publication Data:
Names: Kottemann, Jeffrey E.
Title: Illuminating statistical analysis using scenarios and simulations/Jeffrey E Kottemann, Ph.D.
Description: Hoboken, New Jersey: John Wiley & Sons, Inc. [2017], | Includes index.
Identifiers: LCCN 2016042825| ISBN 9781119296331 (cloth) | ISBN 9781119296362 (epub)
Subjects: LCSH: Mathematical statistics. | Distribution (Probability theory)
Classification: LCC QA276 .K676 2017 | DDC 519.5—dc23 LC record available at https://lccn.loc.gov/2016042825
The goal of this book is to help people develop an assortment of key intuitions about statistics and inference and use those intuitions to make sense of statistical analysis methods in a conceptual as well as a practical way. Moreover, I hope to engender good ways of thinking about uncertainty. The book is comprised of a series of short, concise chapters that build upon each other and are best read in order. The chapters cover a wide range of concepts and methods of classical (frequentist) statistics and inference. (There are also appendices on Bayesian statistics and on data mining techniques.)
Examining computer simulation results is central to our investigation. Simulating pollsters, for example, who survey random people for responses to an agree or disagree opinion question not only mimics reality but also has the added advantage of being able to employ 1000 independent pollsters simultaneously. The results produced by such simulations provide an eye-opening way to (re)discover the properties of sample statistics, the role of chance, and to (re)invent corresponding principles of statistical inference. The simulation results also foreshadow the various mathematical formulas that underlie statistical analysis.
Mathematics used in the book involves basic algebra. Of particular relevance is interpreting the relationships found in formulas. Take, for example, . As increases, increases because is the numerator of the fraction. And as increases, decreases because is the denominator. Going one step further, we could have . Here, as increases, decreases, increases, so increases. These functional forms mirror most of the statistical formulas we will encounter.
My thanks go to Dan Dolk, Gene Hahn, Fati Salimian, and Kathie Wright for their feedback and encouragement. At John Wiley & Sons, thanks go to Susanne Steitz-Filler, Kathleen Pagliaro, Vishnu Narayanan, and Shikha Pahuja.
Before we focus in on using statistics as evidence to be used in making judgments, let's take a look at a widely used “verdict outcomes framework.” This general framework is useful for framing judgments in a wide range of situations, including those encountered in statistical analysis.
Anytime we use evidence to arrive at a judgment, there are four generic outcomes possible, as shown in Table 1.1. Two outcomes correspond to correct judgments and two correspond to incorrect judgments, although we rarely know whether our judgments are correct or incorrect. Consider a jury trial in U.S. criminal court. Ideally, the jury is always correct, judging innocent defendants not guilty and judging guilty defendants guilty. Evidence is never perfect, though, and so juries will make erroneous judgments, judging innocent defendants guilty or guilty defendants not guilty.
Table 1.1 Verdict outcomes framework.
In U.S. criminal court, the presumption is that a defendant is innocent until “proven” guilty. Further, convention in U.S. criminal court has it that we are more afraid of punishing an innocent person (type I error) than we are of letting a guilty person go unpunished (type II error). Because of this fear, the threshold for a guilty verdict is set high: “Beyond a reasonable doubt.” So, convicting an innocent person should be a relatively unlikely outcome. In U.S. criminal court, we are willing to have a greater chance of letting a guilty person go unpunished than we are of punishing an innocent person. In short, we need to be very sure before we reject the presumption of innocence and render a verdict of guilty in U.S. criminal court.
We can change the relative chances of the two types of error by changing the threshold. Say we change from “beyond a reasonable doubt” to “a preponderance of evidence.” The former is the threshold used in U.S. criminal court, and the latter is the threshold used in U.S. civil court. Let's say that the former corresponds to being 95% sure before judging a defendant guilty and that the latter corresponds to being 51% sure before judging a defendant guilty. You can imagine cases where the same evidence results in different verdicts in criminal and civil court, which indeed does happen. For example, say that the evidence leads to the jury being 60% sure of the defendant's guilt. The jury verdict in criminal court would be not guilty (60% < 95%) but the jury verdict in civil court would be guilty (60% > 51%). Compared to criminal court, civil court is more likely to declare an innocent person guilty (type I error), but is also less likely to declare a guilty person not guilty (type II error).
Changing the verdict threshold either decreases type I error while increasing type II error (criminal court) or increases type I error while decreasing type II error (civil court)
Statistical analysis is conducted as if in criminal court. Below are a number of jury guidelines that have parallels in statistical analysis, as we'll see repeatedly.
Erroneously rejecting the presumption of innocence (type I error) is feared most.
The possible verdicts are “guilty” and “not guilty.” There is no verdict of “innocent.”
Reasons other than perceived innocence can lead to a not guilty verdict, such as insufficient evidence.
To reject the presumption of innocence and render a guilty verdict, there must be a sufficient amount of (unbiased) evidence.
To reject the presumption of innocence and render a guilty verdict, the pieces of evidence must be sufficiently consistent (not at variance with each other).
Statistical analysis formally evaluates evidence in order to determine whether to reject or not reject a stated presumption, and it is primarily concerned with limiting the chances of type I error. Further, the amount of evidence and the variance of evidence are key characteristics of evidence that are formally incorporated into the evaluation process. In what follows, we'll see how this is accomplished.
Let's start with the simplest statistical situation: that of judging whether a coin is fair or not fair. Later we'll see that this situation is statistically equivalent to agree or disagree opinion polling. A coin is fair if it has a 50% chance of coming up heads, and a 50% chance of coming up tails when you flip it. Adjusting the verdict table to the coin-flipping context gives us Table 2.1.
Table 2.1 Coin flipping outcomes.
You want to see if a coin is fair or not. To gather evidence, you plan on flipping it 10 times to see how many heads come up. Beforehand, you want to set the verdict rule.
Where do you draw the two lines for declaring the coin to be not fair?
0
1
2
3
4
5
6
7
8
9
10
Number of heads
Draw one line toward the left for your threshold of “too few heads,” and another line toward the right for your threshold of “too many heads.”
Where you draw the lines represents your choice of thresholds.
Just use your intuition; don't do any arithmetic.
Intuitively, it seems extremely unlikely for a fair coin to come up heads only 0 or 1 times out of 10, and most people would arrive at the verdict that the coin is not fair. Likewise, it seems extremely unlikely for a fair coin to come up heads 9 or 10 times out of 10, and most people would arrive at the verdict that the coin is not fair. On the other hand, it seems fairly likely for a fair coin to come up heads 4, 5, or 6 times out of 10, and so most people would say that the coin seems fair. But what about 2, 3, 7, or 8 heads? Let's experiment.
Shown in Figure 2.1 is a histogram of what actually happened (in simulation) when 1000 people each flipped a fair coin 10 times. This shows us how fair coins tend to behave. The horizontal axis is the number of heads that came up out of 10. The vertical axis shows the number of people out of the 1000 who came up with the various numbers of heads.
Figure 2.1
Appendix B gives step-by-step instructions for constructing this simulation using common spreadsheet software; guidelines are also given for constructing additional simulations found in the book.
Sure enough, fair coins very rarely came up heads 0, 1, 9, or 10 times. And, sure enough, they very often came up heads 4, 5, or 6 times. What about 2, 3, 7, or 8 heads?
Notice that 2 heads came up a little less than 50 times out of 1000, or near 5% of the time. Same with 8 heads. And, 3 heads came up well over 100 times out of 1000, or over 10% of the time. Same with 7 heads.
Where do you want to draw the threshold lines now?
0
1
2
3
4
5
6
7
8
9
10
Number of heads
How about the lines shown on the histogram in Figure 2.2?
Figure 2.2
Eyeballing the histogram, we can add up the frequencies for 0, 1, 2, 8, 9, and 10 heads out of the 1000 total. Fair coins came up with 0, 1, 2, 8, 9, or 10 heads a total of about and of the time.
So, a “verdict rule” that has about an 11% chance of type I error can be stated as:
If the number of heads flipped is outside the interval
Number of heads ≥3 and number of heads ≤7
Then reject the presumption that the coin is fair.
Fair coins will be inside the interval about 89% of the time, and we will correctly judge them to be fair coins.
Fair coins will be outside the interval about 11% of the time, and we will incorrectly judge them to be unfair coins.
The “verdict rule” defines an interval outside of which we will reject the presumption.
Type I error
occurs when a fair coin's head count is outside our interval.
Type II error
occurs when an unfair coin's head count is inside our interval.
Why is the histogram bell-shaped?
Before expanding the previous Statistical Scenario let's briefly explore why the histogram, reproduced in Figure 3.1, is shaped the way it is: bell-shaped. It tapers off symmetrically on each side from a single peak in the middle.
Figure 3.1
Since each coin flip has two possible outcomes and we are considering ten separate outcomes together, there are a total of unique possible patterns (permutations) of heads and tails with 10 flips of a coin. Of these, there is only one with 0 heads and only one with 10 heads. These are the least likely outcomes.
TTTTTTTTTT
HHHHHHHHHH
There are ten with 1 head, and ten with 9 heads:
HTTTTTTTTT
THHHHHHHHH
THTTTTTTTT
HTHHHHHHHH
TTHTTTTTTT
HHTHHHHHHH
TTTHTTTTTT
HHHTHHHHHH
TTTTHTTTTT
HHHHTHHHHH
TTTTTHTTTT
HHHHHTHHHH
TTTTTTHTTT
HHHHHHTHHH
TTTTTTTHTT
HHHHHHHTHH
TTTTTTTTHT
HHHHHHHHTH
TTTTTTTTTH
HHHHHHHHHT
Since there are 10 times more ways to get 1 or 9 heads than 0 or 10 heads, we expect to flip 1 or 9 heads 10 times more often than 0 or 10 heads.
Further, there are 45 ways to get 2 or 8 heads, 120 ways to get 3 or 7 heads, and 210 ways to get 4 or 6 heads. Finally, there are 252 ways to get 5 heads, which is the most likely outcome and therefore the most frequently expected outcome. Notice how the shape of the histogram of simulation outcomes we saw in Figure 3.1 closely mirrors the number of ways (#Ways) chart that is shown in Figure 3.2.
Figure 3.2
You don't need to worry about calculating #ways. Soon we won't need such calculations. Just for the record, the formula for the #ways is where is the number of flips, h is the number of heads you are interested in, and ! is the factorial operation (example: ). In official terms, #ways is the number of combinations of things taken at a time.
Let's revisit Statistical Scenario–Coins #1, now with additional information on each of the possible outcomes. Table 4.1 summarizes this additional information. As noted, there are a total of different unique patterns of heads & tails possible when we flip a coin 10 times. For any given number of heads, as we have just seen, there are one or more ways to get that number of heads.
Using Table 4.1, where do you draw the two lines for declaring the coin to be unfair?
Table 4.1 Coin flipping details.
#Heads
#Ways
Expected relative frequency
Probability
as Percent
Rounded
0
1
1/1024
0.00098
0.098%
0.1%
1
10
10/1024
0.00977
0.977%
1.0%
2
45
45/1024
0.04395
4.395%
4.4%
3
120
120/1024
0.11719
11.719%
11.7%
4
210
210/1024
0.20508
20.508%
20.5%
5
252
252/1024
0.24609
24.609%
24.6%
6
210
210/1024
0.20508
20.508%
20.5%
7
120
120/1024
0.11719
11.719%
11.7%
8
45
45/1024
0.04395
4.395%
4.4%
9
10
10/1024
0.00977
0.977%
1.0%
10
1
1/1024
0.00098
0.098%
0.1%
Totals:
1024
1024/1024
1.0
100%
100%
The #ways divided by 1024 gives us the expectedrelative frequency for that number of heads expressed as a fraction. For example, we expect to get 5 heads 252/1024ths of the time. The fraction can also be expressed as a decimal value. This decimal value can be viewed as the probability that a certain number of heads will come up in 10 flips. For example, the probability of getting 5 heads is approximately 0.246. We can also express this as a percentage, 24.6%.
A probability of 1 (100%) means something will always happen and a probability of 0 (0%) means something will never happen. A probability of 0.5 (50%) means something will happen half the time. The sum of the probabilities of the entire set of possible outcomes is the sum of all the probabilities and always equals 1 (100%). The probability of a subset of possible outcomes can be calculated by summing the probabilities of each of the outcomes. For example, using the rounded percentages from the table, the probability of 2 or fewer heads is .
Notice how the bars of our simulation histogram, reproduced in Figure 4.1, reflect the corresponding probabilities in Table 4.1.
Figure 4.1
Say someone gives you a coin to test. When you flip the coin 10 times, you are sampling the coin's behavior 10 times. The number of heads you toss is your evidence. Based on this evidence you must decide whether to reject your presumption of fairness and judge the coin as not fair.
What happens if you make your “verdict rule” to be:
Verdict “coin is not fair” if outside the interval #heads ≥ 1 and ≤ 9 as shown in Table 4.2 and the accompanying Figure 4.2?
From the Statistical ScenarioTable 4.1, we can see that a fair coin will come up 0 heads about 0.1% of the time, and 10 heads about 0.1% of the time. The sum is about 0.2% of the time, or about 2 out of 1000. So, it will be extremely rare for us to make a type I error and erroneously call a fair coin unfair because fair coins will almost never come up with 0 or 10 heads. However, what about 1 head or 9 heads? Our rule says not to call those coins unfair. But a fair coin will only come up 1 head or 9 heads about of the time. Therefore, we may end up misjudging many unfair coins that come up heads one or nine times because we'll declare them to be fair coins. That is type II error.
Table 4.2 First verdict rule scenario.
Figure 4.2
Determining the chance of type II error is too involved for discussion now (that is Chapter 17), but recall from Chapter 1 that increasing the chance of type I error decreases the chance of type II error, and vice versa.
To lower the chances of type II error, we can narrow our “verdict rule” interval to #heads ≥ 2 and ≤ 8 as shown in Table 4.3 and Figure 4.3. Now the probability of making a type I error is about . This rule will decrease the chances of type II error, while increasing the chances of type I error from 0.2 to 2.2%.
If we narrow our “verdict rule” interval even more to #heads ≥ 3 and ≤ 7, we get Table 4.4 and Figure 4.4.
Table 4.3 Second verdict rule scenario.
Figure 4.3
Now the probability of making a type I error is about because a fair coin will come up 0, 1, 2, 8, 9, or 10 heads about 11% of the time. We can express this uncertainty by saying either that there will be an 11% chance of a type I error, or that we are 89% confident that there will not be a type I error. Notice that this is what we came up with earlier by simply eyeballing the histogram of actual simulation outcomes in Chapter 2.
Table 4.4 Third verdict rule scenario.
Figure 4.4
As noted earlier, type I error is feared most. And an 11% chance of type I error is usually seen as excessive. So, we can adopt this rule:
Verdict:
If outside the interval #Heads ≥ 2 and ≤ 8, Judge Coin to be Not Fair.
This gives us about a 2% chance of type I error.
From now on, we'll typically use the following threshold levels for type I error: 10% (0.10), 5% (0.05), and 1% (0.01). We'll see the effects of using various thresholds as we go along. Also as we go along we'll need to replace some common words with statistical terminology. Below are statistical terms to replace the common words we have been using.
Alpha-Level is the threshold probability we stipulate for the occurrence of type I error. Commonly used alpha-levels are 0.05 and 0.01. Alpha-levels used in sciences such as physics are typically much lower.
Confidence level is the complement of alpha-level and is the threshold percentage we stipulate for the nonoccurrence of type I error. 95% confidence corresponds to an alpha-level of 0.05. 99% confidence corresponds to an alpha-level of 0.01.
Null hypothesis is the presumption we have been referring to above. Example: the coin is not unusual, it is a fair coin. Most of the simulations in this book simulate outcomes to expect when the null hypothesis is true: like a fair coin.
Binomial is a variable with only two possible values, such as heads or tails. The term comes from the Latin for “two names.” Other examples of binomials are agree or disagree, male or female. When we represent the binomial values as 0 or 1, then the average will give us the proportion of 1s. For example, the average of 0, 1, 0, and 1 is 0.5 and the average of 1, 1, 0, and 1 is 0.75. (Appendix A overviews all the types of data we'll be working with as the book progresses.)
Sample is the set of observations we have, which is the set of heads and tails flipped for the above coin flipping cases. A sample constitutes evidence.
Sample size is the number of separate observations, which equals 10 for the above cases. The italicized letter is often used: the sample .
Sample statistic is a calculated value that serves to summarize the sample. They are summarizations of evidence. Examples: number of heads in the sample or Proportion of heads in the sample (number of heads divided by sample size). Counts and proportions are the basic sample statistics for binomial variables.
Sampling distributions are illustrated by the histograms of simulation results. They reflect the distributions of the values for sample statistics we get with repeated sampling.
It is important to emphasize that simulation histograms represent sampling distributions that tell us what to expect when the null hypothesis is true. We'll look at many, many sampling distribution histograms in this book. For the remainder of Part I, we'll switch from using counts as our sample statistic to using proportions as our sample statistic. The sampling distribution histogram in Figure 4.5 shows the use of proportions rather than counts on the horizontal axis.
Figure 4.5
