This page will examine and step through multiple solutions for hypothesis tests based on two samples using the TI-83/84 calculators.
We have a number of different instances to examine:
We have two treatment groups, A and B, from which we have taken two samples. The statistics for those samples are given by:
A | B | |
Sample size | 49 | 44 |
Sample mean | 69.4 | 68.7 |
Sample standard deviation | 2.5 | 1.7 |
![]() | The standard error of the difference of the means is given by `sqrt((s_1^2)/(n_1) + (s_2^2)/(n_2))`. This would have been a pain to compute had we not developed the CALCVSDF program in the last last chapter. We start a run of the program here and give it the sample size and standard deviation for the first sample. |
![]() |
Figure 2 shows the completion of the data input for the program. |
![]() |
Figure 3 provides the output from the program. This gives us the standard error as .4395825731 and the "complex" degrees of freedom as B85.00205449. |
![]() |
Because we do not know the standard deviations of the populations, we use the standard deviation of the samples as our estimate. Therefore, we are using the Student's-t distribution. To get a "critical value" we can go to the DIST menu and select invT(. [Note that users of the TI-83, TI-83-Plus and early software for the 84-Plus can use the INVT program to get the same values.] |
![]() |
This is a one-sided test so our command is
invT(.9,85.00205449).
The result is our "critical value, 1.291590577.
The test statistic tht we want is `t = ( (barx_1 - barx_2) - (mu_1 - mu_2) )/ sqrt((s_1^2)/(n_1) + (s_2^2)/(n_2))` but under the null hypothesis `mu_1-mu_2` is zero. thus, we really calculate `t = ( barx_1 - barx_2 )/ sqrt((s_1^2)/(n_1) + (s_2^2)/(n_2))` and we already have the value of that denominator. Thus the command (69.4-68.7)/.4395825731 will give us our test statistic. The calculator produces the value 1.592419816. This is more extreme than was our "critical value". Therefore, using the "critical value" approach, we will reject the null hypothesis in favor of the alternative hypothesis. |
![]() |
To use the "P-value" approach we need to see how "extreme" our test statistic turned out to be. TO do this we use the tcdf( command, shown in Figure 6 in the DIST menu. |
![]() |
The result, computed with our "complex" degrees of freedom, is .0575012782 which is less than the 0.10 given in the problem. Therefore, this result is that we "reject" the null hypothesis. |
![]() |
We just completed the problem by working it out using the formula and variaous commands and programs. An alternative approach is to have the calcualtor do all the work. We find, in the STAT menu, under the TESTS sub-menu, the item 2-SampTTest.... Choosing that option opens a data input page. |
![]() |
On this screen we have made sure that we are in the Stats mode and that we have placed the appropriate values from the problem into the fields of the screen. |
![]() |
Moving down the screen we complete the settigns by choosing the
`>mu2` option. Then we highlight the Calculate option and
press ![]() |
![]() |
The result, shown in Figure 11, gives us both the appropraite test statistic, namely t=1.592419816 and the P-value from that, namely p=.0575012782. These are the same values that we computed in Figures 1 through 7. |
Use these data values to test the hypothesis `H_0: mu_1 = mu_2` against `H_1: mu_1 > mu_2` at the 0.05 significance level.
![]() |
The first thing to do is to generate the data values.
As in earlier examples from the last chapter, we will generate the second list first and then copy to L2. Figure 12 has the key values to generate the second table. |
![]() |
Figure 13 shows the start of the data, the command to copy L1 to L2, and the command to find the one variable statistics for the data. |
![]() |
Figure 14 gives us the values that we need to be able to follow the initial approach from the previous example. In particualr, Figure 14 gives us the sample size, the sample mean, and the sample standard deviation. |
![]() |
Then we generate the first list of values. |
![]() |
Again we see the start of that list and the command to get the one variables statistics on that data. |
![]() |
Now we have the sample size, sample mean,
and sample standard deviation from the second list.
At this point we are just where we were at the start of the first example: we know all the needed values to do the computation. |
![]() |
We use the information from Figure 17 to enter the sample size and standard deviation into the CALCVSDF program. |
![]() |
We complete the data entry for that program with the information from Figure 14 for the second list. |
![]() |
The program gives us the two values that we will need, namely, the combined standard deviation 1.076684266 and the number of degrees of freedom 48.87017928. |
![]() |
Before we go on to do the actual work here, we might recall that
the program stored the standard deviation in S and that
it found the degrees of freedom as the ratio N/M.
[One could look back at the program listing to confirm this.]
Therefore, when we ask for the value of S we get the
1.076684266 value, and when we ask for N/M
we get the 48.87017928 value.
We use this in forming the invT(.95,N/M) command to get the "critical value" for this problem. |
![]() |
Again, following the earlier approach, we compute,
as we did in Figure 5,
the desired test statistic as the difference of
the sample means divided by the standard deviation.
However, in this case we shorten our typing
by using S for that standard deviation.
This gives us the test statistic as 1.297331868.
Using the "critical value" method, we note that this value is not
as extreme as was our critical value.
Therefore, we do not reject the null hypothesis, we accept it.
Alternatively, using the P-value method, we compute the probability of getting a value as extreme or more extreme than the one we got via the tcdf( command. The result is .100304044, which is larger than the allowed .05 given in the problem statement. Again, we accept the null hypothesis. |
![]() |
As with the first example, we really did not have to do all this work. The calculator will do almost everything for us. Figure 23 returns to the 2-SampTTest input screen. GHere we have changed to the Data option and we have given the names of the two lists holding the sample data. in addition, we have chosen the `>mu2` option. |
![]() |
We move down to the Calculate option and press
![]() |
![]() |
The calcualtor processes the two samples and generates the screen shown in Figure 25. This gives us the same values that we just computed the hard way. |
We are interested in the proportion of a characteristic A in two large sub-groups, Group 1 and Group 2, of a larger population. We have taken a sample of each group and we have found the following:
Group 1 | Group 2 | |
Sample Size | 172 | 186 |
Num with A | 60 | 40 |
![]() |
Unlike the process for finding confidence intervals,
the hypothesis test uses the pooled proportion
`hatp = (x_1 + x_2) / (n_1 + n_2)`.
Figure 26 shows the computation of that
value and storing it into P.
The desired standard deviation is then ` sqrt(( hatp*(1-hatp) )/(n_1) + ( hatp*(1-hatp) )/(n_2) ) = sqrt( hatp*(1-hatp)*(1/n_1 +1/n_2))` which is also comptued in Figure 26 and stored in S Finally, Figure 26 sets up the computation of the appropriate test statistic. |
![]() |
Figure 27 gives us the value of that test
statistic as 2.818738152.
We go on to find the probability of a value this extreme or more extreme, remembering that we are using a normal distribution. Then, because this is a two-sided test, we remember to multiply that result by 2 to get the total probability, namely, .0048214068. This is less than the problem stated significance level of 0.005. Therefore, we reject the null hypothesis in favor of the alternative hypothesis. |
![]() |
Alternatively, we could have used the calculator option 2-PropZTest... found tin the STAT menu under the TESTS sub-menu. |
![]() |
Figure 29 shows the resulting input form with data from our problem supplied. |
![]() |
Figrue 30 gives all the required results, identical to those we found by our computations. |
![]() |
Naturally, we start by generating the data. |
![]() |
Once we have the data, we compute L3
as the difference of the paired values.
Then, we want to see the one variable statistics for L3. |
![]() |
Figure 33 gives us all the values we need. |
![]() |
Our test statistic is the difference of the means divided by
the standard deviation of the mean of the differences.
But the difference of the original means is just the mean of the
differences, namely, -1.466666667 and the
standard deviation of the mean of the differences is
`3.72869795/sqrt(24)`.
We compute the desired quotient and store the result in
S.
Then we find the probability of getting a value this extreme or more extreme. using the student's-t distribution with 23 degrees of freedom, this turns out to be .0332114791. This is well below the given significance level of 0.100. Therefore, we reject the null hypothesis in favor of the alternative hypothesis. |
![]() |
Alternatively, we could have gone back and selected the
T-Test option in the STAT menu, TESTS sub-menu,
told it that we want to process the data in
L3, set the test for ` |
We want to run a test on the null hypohtesis that the standard deviations of two populations are the same versus an alternative hypothesis. We are given the following values:
Sample 1 | sample 2 | |
sample size | 17 | 28 |
sample standard deviation | 6.1 | 4.4 |
Thus, we want to test `H_0: sigma_1 = sigma_2` against `H_1: sigma_1 > sigma_2`, and do so at the 0.1 significance level.
![]() |
The test statistic that we want is the quotient
`(max(sigma_1^2,sigma_2^2))/(min(sigma_1^2,sigma_2^2))`.
In our problem the larger variance is going to be 6.12
and the smaller will be 4.42.
Thus, we compute the desired quotient of the variances to
get a test statistic of 1.922004132.
Note that because we divide the larger by the smaller the result will
always be geater than 1.
The next step is to see the probability of getting a value this strange or stranger. The distribution that we use is the F distribution. We want the cumulative probability of values from 1.922004132 to positive infinity, but we will settle for values from 1.922004132 to 9999. The F distribution requires us to specify first the number of degrees of freedom in the numerator of the test statistic, and, second, the number of degrees of freedom in the denominator of the test statistic. Each is one less than the respective sample size. In our problem the numberator came from Sample 1 and it has sample size of 17. Therefore, the value 16 becomes the third parameter for the Fcdf( command. The denominator of the test statistic came from Sample 2 and it has sample size of 28. Therefore, the value 27 becmes the fourth parameter for the Fcdf( command. The result, the achieved level of significance, is .0650661498. This is less than the problem specified 0.10. Therefore, we reject the null hypothesis. |
This is a case where we have two data sets, perhaps one from each of two treatments, and we want to test the null hypothesis that the two populations have the same standard deviation. In this example, we will test that against the alternative that the standard deviation of the first is less than the standard deviation of the second. We will set the level of significance at 0.100.
![]() |
The most straight forward way to do this is to generate the
two lists, obtain the statistics of each list,
pull out the necessary values, do the computation just as we did
in Figure 35.
As before, we start by generating the second list. |
![]() |
Once generated, we copy it to L2. Then we ask for the 1-Var Stats. |
![]() |
From Figure 38 we can get the sample size, 25, and the sample standard deviation, 2.850713361. |
![]() |
Now generate the first list. |
![]() |
Ask for the 1-Var Stats. |
![]() |
The first list has sample size of 27 and a standard deviation of 2.456672987. |
![]() |
The test statistic is the quotient of the
two variances, with the large one in the numerator.
Figure 42 shows that this becomes 1.346509391.
Note that we had to use the variance of the second list (even though we generated it first), as the numerator. Therefore, the third parameter of the Fcmd( command will be one less than the sample size of the second list, or 24. That leaves one less than the sample size of the first list, or 26, as the fourth parameter. having formed and performed the command, the result is .2292632314, a value that is not less than the problem specified significance level of 0.10. Therefore, we accept the null hypothesis. |
![]() |
The computations that we just went through were pretty straight forward. However, teh calculator does have a command to do such a test. In the STAT menu, in the TESTS sub-menu, there is the 2-sampFTest.... |
![]() |
Figure 44 shows the input screen for that test. The values for this problem have been given. We can go ahead and perform the test. |
![]() |
The result is shown in Figure 45.
However, there is a problem here. Indeed, the
resulting probability, .2292632314, is just what we
computed in Figure 42. As such, we will still accept the null hypothesis.
But the F statistic is all wrong. Figure 45 gives it as
.7426558364. Not only is this different from the
1.346509391 we computed above, but it is less than 1.
This seems impossible given that our quotient has the larger
variance in the numerator.
What is wrong? |
![]() |
In order to help find the problem, we will first find the variable that the calculator uses to hold the F statistic. We find it in the VARS, Statistics, TEST menu. |
![]() |
In Figure 47 we have used that variable to get 1/F.
This turns out to be exactly the value that we thought we would get
for the F statistic.
It seems that the calcualtor has computed the quotient, but done so with the numerator and denominator reversed. It appears that the calculator uses the List1: list in the numerator and the List2: list in the denominator. Just looking at the two original lists we do not realize that the second list has the larger variance. Looking at the results that we obtained in Figures 38 and 41 we see that this is indeed the case. |
![]() |
Perhaps the solution is to reverse the order of the two lists in the input screen. In Figure 48 we have done that. Now we can try again. |
![]() |
These results are different, but they still pose a problem.
Now the F statistic is correct, but the probability is wrong.
The new probability is just 1-.2292632314.
This should lead us to understand that
we reversed the the two lists but we did not reverse the
direction of the alternative. The original alternative
was `sigma_1 |
![]() |
Figrue 50 shows that change. |
![]() |
Figure 51 shows the results. We finally have the right answers.
The lesson here is that the calculator does its thing the way it wants to. We can use the approach shown in Figures 36 through 42 and get pretty clear results. Alternatively, we can use the 2-SampFTest process but we better be ready to change our input to get the desired output. |
©Roger M. Palay
Saline, MI 48176
November, 2012