Tuesday, February 06, 2007

Chapter 10 Beginning of Inference

This chapter introducecs important methods under the highly unrealistic conditions where we know the population standard deviation but not the population mean.

Point estimates for the average value of X found through samples are generally good estimates, but they are wrong. You can generate a better estimate by creating a confidence interval.

The confidence interval =
x-bar +/- Z* times sigma of x / (Sqrt n).

We get Z* from the t-table for a specific confidence level, for instance when we want a 95% we use 1.96.

In creating a complete solution we first write down all of the given information. Define your variable. Then we determine whether the central limit theorem has "kicked in" or if the underlying data were already normally distributed. Be sure to address whether the data were from a SRS. Graph them if you have them to make sure there are no gaps or outliers. Is the sample size less than 1/10 of the population size???

Identify what you are trying to produce-- a 95% Z interval for mu and give the formula. Show how the numbers are plugged in and calculate the interval.

Write the interpretation of your interval.

We are 95% confident that the true population mean value of [insert the contextual information here] falls between [lower bound] and [upper bound].

Refer to your notes for all of the baaaaaaaaaaaaad interpretations of a confidence interval and NEVER use them. :)

If a value of mu had been proposed before we collected our sample, we could see if the value falls within our interval. If the proposed value does fall in the interval, then it is a reasonable value, although not necessarily correct. If it does not fall in the interval, then it is not a reasonable value according to our sample values.

Hypothesis tests

For hypothesis tests, you develop a null and alternate hypothesis BEFORE you collect data. Both hypotheses use the parameter (NEVER THE STATISTIC) and they are considered logical opposites. The null hypotheses ALWAYS has an "equals" aspect to it: the alternate hypothesis is always <, >, or not equal to.

For instance: H0: mu = 15
Ha: mu > 15.

Although these are not actually opposites, finding evidence that mu is less than 15 provides no support of the alternative hypothesis. You can think of the null hypothesis in this case as mu<=15, which still has an "equals" in it. This is the way I learned hypothesis-writing back in the day and it is still acceptable, but not as common.

Alpha, Beta, Type I error, Type II error, and Power

Alpha is the likelihood of a Type I error--accidentally rejecting the null hypothesis when it was actually correct. (Like convicting the wrong guy.)

Beta is the likelihood of a Type II error--failing to reject the null when it was wrong. (Kind of an error of omission, or not enough evidence to convict.)

Power is the likelihood that the test would have been sensitive enough to pick up the difference between the hypothesized mu and the actual mu (given some other new value for mu). This is the complement of Beta. Yes, 1 - Beta = Power. 1 - power = beta. Power + beta = 1.

Notice that alpha and beta are NEVER added together. They don't live under the same conditions--one assumes that the null was true and the other that the null was false. DO not fall into the trap of EVER adding alpha and beta together (unless you are TOLD to do it and then only if they offer you a lot of money or a passing grade on a test).

Calculating beta is easier than people think on the calculator.
(1) Figure out what the critical values are for rejection of Ho in terms of x-bar.

(2) Find the area under the curve centered at the NEW mu that falls between these critical values. You can use normalcdf(left_critical_value, right_critical_value, new mean, standard dev or error of x-bar).

11 comments:

nanahcubbackwards said...

I'm still kind of confused about the power probabilities or their called something like that. I understand what they're asking for but I don't know how to get it.

jfulton said...

On #71 from the homework I did exactly what it said to do to solve for the type two error and my answer matched the back of the book but the question asked for the power of the test... Therefore isn't it 1- P(type two error)?

jfulton said...

Yeah one more question regarding which z value to use... In class today when we were using the 5% significance we chose the z value that only had a probability of 5% on one tail alone (not both). Is that only when the Alternative Hypothesis is < or >? And if thats true when the alternative hypothesis is not equal to whatever value would you still go back and find the appropriate z value the normal way?

Mrs.L said...

B7P--verify all assumptions you need to proceed with an inference (SRS, normal, std dev)

HB-- Google the power of a test and see if you can find an applet. The power of the test is the probability that the test will be able to differentiate between your original null hyp and some new and improved mean. It is the probability that you would have rejected Ho given some other mu is correct instead of the null one.

To find it, take the probability that X-bar falls in the rejection region, but using the new mu and the standard deviation. You may use Normalcdf for this.


JF: You calculated a critical x-bar value of 299.013, then found the probability that the null would be rejected (P(x-bar < 299.013)) as described in the HB note above. That is the power.

re: 2nd question Yes. The total area in the tails of interest is alpha, the significance level. If you're looking for errors in both directions, then you split up the alpha into two tails. If you're only interested in one direction, then all the alpha goes in the one tail.

Mrs.L said...

Power applet:
http://wise.cgu.edu/power/power_applet.html

Pretty cool. Click and drag the distribution right and left.

nkhat said...

Is power of a test the probability that we would reject the null hypothesis when the alternative true?

nkhat said...

Also, why does increasing the sample size of a distribution increase its height and decrease its width?

Mrs.L said...

The distribution gets taller because the area under the curve always equals 1, but when n increases the std dev of x-bar decreases (sigma-sub-x /n gets smaller!!!).

Mrs.L said...

Both if these questions are incredibly important.

Extreme means significant, that it is so unlikely that it could not have happened just because of randomness, that you can't believe that the null hypothesis is true, that you REJECT it.

NormalCdf(lowerbound, upperbound, mean, std dev) gives the area under the curve between the two bounds given the mean and std dev listed. When you use z-values the mean and std dev are 0 and 1.

Follow the patterns of the probability statements that I ALWAYS do when I solve these problems. For instance

P(x-bar > your sample avg) =
P((x-bar - mu)/sigma of x-bar > (your sample avg - mu)/std dev of x-bar)

P(z > the z-statistic you calculated)= the probability in the tail to the right of your sample average. You find this probability by using
normalcdf(z-statistic, infinity, 0, 1).

If your tail goes to the left, then you would use less thans.

Go back to the examples throughout this chapter and the many problems we worked in class.

sean 7th said...

ok so i was totally lost on this. thx 4 clearing it up a lil... still need 2 come in early one day for the finer points but it helps alot... when will chap 11 get posted? 2 days in and im already lost...

Mrs.L said...

See, that's the great thing about teaching AP Statistics students!

I've posted the first parts of Chapter 11. I hope that they help.