-
Graphing Data 4
-
Lecture1.1
-
Lecture1.2
-
Lecture1.3
-
Lecture1.4
-
-
Mean and Standard Deviation 5
-
Lecture2.1
-
Lecture2.2
-
Lecture2.3
-
Lecture2.4
-
Lecture2.5
-
-
Distributions 6
-
Lecture3.1
-
Lecture3.2
-
Lecture3.3
-
Lecture3.4
-
Lecture3.5
-
Lecture3.6
-
-
Correlation and Linear Regression 7
-
Lecture4.1
-
Lecture4.2
-
Lecture4.3
-
Lecture4.4
-
Lecture4.5
-
Lecture4.6
-
Lecture4.7
-
-
Probability 3
-
Lecture5.1
-
Lecture5.2
-
Lecture5.3
-
-
Counting Principles 3
-
Lecture6.1
-
Lecture6.2
-
Lecture6.3
-
-
Binomial Distribution 3
-
Lecture7.1
-
Lecture7.2
-
Lecture7.3
-
-
Confidence Interval 7
-
Lecture8.1
-
Lecture8.2
-
Lecture8.3
-
Lecture8.4
-
Lecture8.5
-
Lecture8.6
-
Lecture8.7
-
-
Proportion Confidence Interval 3
-
Lecture9.1
-
Lecture9.2
-
Lecture9.3
-
-
Hypothesis Testing 5
-
Lecture10.1
-
Lecture10.2
-
Lecture10.3
-
Lecture10.4
-
Lecture10.5
-
-
Comparing Two Means 5
-
Lecture11.1
-
Lecture11.2
-
Lecture11.3
-
Lecture11.4
-
Lecture11.5
-
-
Chi-squared Test 3
-
Lecture12.1
-
Lecture12.2
-
Lecture12.3
-
Z Scores
Solution
print(len(zScores[abs(zScores) < 1])/10000)
print(len(zScores[abs(zScores) < 2])/10000)
print(len(zScores[abs(zScores) < 3])/10000)
It holds up pretty well!
What does the distribution of the z-scores look like?
plt.hist(zScores,bins=100)
plt.show()
Let’s formally define the cdf function.
Equation
$$\text{x = The cutoff point x we want to see if our X is less than or equal to}$$
$$\text{X = Our random variable X}$$
By proxy, if we wanted to see what the probability of our variable being greater than or equal to x was, we could take 1-CDF(x) to get the opposite.
Equation
Now what if we wanted to find the bounds for 1, 2 and 3 standard deviations like we did before, but this time with the CDF function?
Let’s break it down the equation for one standard deviation.
Explanation
Step 1
$$ P(- \sigma < X < \sigma) $$
Step 2
$$ 1-(P(X > \sigma) + P(X < – \sigma))$$
Step 3
$$ 1-(1-CDF(\sigma) + CDF(-\sigma)) $$
Step 4
$$ CDF(\sigma) – CDF(-\sigma) $$
The solution makes sense, we want the probability that we are under a positive standard deviation away, but we do not want to include the probability of being all the way under a negative standard deviation away.
Challenge