Difference between revisions of "AP Statistics Curriculum 2007 Infer 2Means Indep"

From SOCR
Jump to: navigation, search
(Analysis Protocol for Independent Designs)
(Detailed specifications for independent-sample design studies)
 
(18 intermediate revisions by 3 users not shown)
Line 20: Line 20:
  
 
* '''Confidence Intervals''': <math>(1-\alpha)100%</math> confidence interval for <math>\mu_1-\mu_2</math> will be  
 
* '''Confidence Intervals''': <math>(1-\alpha)100%</math> confidence interval for <math>\mu_1-\mu_2</math> will be  
: <math>CI(\alpha): \overline{x}-\overline{y} \pm z_{\alpha\over 2} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm z_{\alpha\over 2}  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}</math>. Note that the <math>SE(\overline{x} -\overline{x})=\sqrt{SE(\overline{x})+SE(\overline{y})}</math>, as the samples are independent. Also, <math>z_{\alpha\over 2}</math> is the [[AP_Statistics_Curriculum_2007_Normal_Critical | critical value]] for a [[AP_Statistics_Curriculum_2007_Normal_Std |Standard Normal]] distribution at <math>{\alpha\over 2}</math>.
+
: <math>CI(\alpha): \overline{x}-\overline{y} \pm z_{\alpha\over 2} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm z_{\alpha\over 2}  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}</math>. Note that the <math>SE(\overline{x} -\overline{y})=\sqrt{SE^2(\overline{x})+SE^2(\overline{y})}</math>, as the samples are independent. Also, <math>z_{\alpha\over 2}</math> is the [[AP_Statistics_Curriculum_2007_Normal_Critical | critical value]] for a [[AP_Statistics_Curriculum_2007_Normal_Std |Standard Normal]] distribution at <math>{\alpha\over 2}</math>.
  
 
====Small Samples====
 
====Small Samples====
 
*'''Significance Testing''': Again, we have a standard null-hypothesis <math>H_o: \mu_X -\mu_Y = \mu_o</math> (e.g., <math>\mu_o=0</math>). Then the test statistics is:
 
*'''Significance Testing''': Again, we have a standard null-hypothesis <math>H_o: \mu_X -\mu_Y = \mu_o</math> (e.g., <math>\mu_o=0</math>). Then the test statistics is:
: <math>T_o = {\overline{x}-\overline{y}-\mu_o \over SE(\overline{x}-\overline{y})} \sim T(df)</math>.
+
: <math>T_o = {\overline{x}-\overline{y}-\mu_o \over SE(\overline{x}-\overline{y})} \sim T_{(df)}</math>.
: The ''degrees of freedom'' is: <math>df={( SE^2(\overline{x})+SE^2(\overline{x}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.</math> Always round down the degrees of freedom to the next smaller integer.
+
: The ''degrees of freedom'' (df) is: <math>df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.</math> Always round down the degrees of freedom to the next smaller integer.
 
: <math>t_o= {\overline{x}-\overline{y}- \mu_o \over  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}</math>
 
: <math>t_o= {\overline{x}-\overline{y}- \mu_o \over  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}</math>
  
 
* '''Confidence Intervals''': <math>(1-\alpha)100%</math> confidence interval for <math>\mu_1-\mu_2</math> will be  
 
* '''Confidence Intervals''': <math>(1-\alpha)100%</math> confidence interval for <math>\mu_1-\mu_2</math> will be  
: <math>CI(\alpha): \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}}  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}</math>. Note that the <math>SE(\overline{x} -\overline{x})=\sqrt{SE(\overline{x})+SE(\overline{y})}</math>, as the samples are independent.  
+
: <math>CI(\alpha): \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}}  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}</math>. Note that the <math>SE(\overline{x} -\overline{y})=\sqrt{SE^2(\overline{x})+SE^2(\overline{y})}</math>, as the samples are independent.  
: The ''degrees of freedom'' is: <math>df={( SE^2(\overline{x})+SE^2(\overline{x}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.</math> Always round down the degrees of freedom to the next smaller integer. Also, <math>t_{df, {\alpha\over 2}}</math> is the [[AP_Statistics_Curriculum_2007_Normal_Critical | critical value]] for a [[AP_Statistics_Curriculum_2007_StudentsT |Student's T]] distribution at <math>{\alpha\over 2}</math>.
+
: The ''degrees of freedom'' is: <math>df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.</math> Always round down the degrees of freedom to the next smaller integer. Also, <math>t_{df, {\alpha\over 2}}</math> is the [[AP_Statistics_Curriculum_2007_Normal_Critical | critical value]] for a [[AP_Statistics_Curriculum_2007_StudentsT |Student's T]] distribution at <math>{\alpha\over 2}</math>.
  
 
===Example===
 
===Example===
Nine observations of surface soil [http://en.wikipedia.org/wiki/PH pH] were made at two different locations.  Does the data suggest that the true mean soil [http://en.wikipedia.org/wiki/PH pH] values differ for the two locations?  Formulate testable hypothesis and make inference about the effect of the treatment at <math>\alpha=0.05</math>. Check any necessary assumptions for the validity of your test.
+
Nine observations of surface soil [http://en.wikipedia.org/wiki/PH pH] were made at two different locations.  Does the data suggest that the true mean soil [http://en.wikipedia.org/wiki/PH pH] values differs for the two locations?  Formulate testable hypothesis and make inference about the effect of the treatment at <math>\alpha=0.05</math>. Check any necessary assumptions for the validity of your test.
  
 
====Data in row format====
 
====Data in row format====
Line 89: Line 89:
  
 
* Test statistics: We can use the sample summary statistics to compute the degrees of freedom and the [[AP_Statistics_Curriculum_2007_Infer_2Means_Dep#Test_Statistics |T-statistic]]
 
* Test statistics: We can use the sample summary statistics to compute the degrees of freedom and the [[AP_Statistics_Curriculum_2007_Infer_2Means_Dep#Test_Statistics |T-statistic]]
: The ''degrees of freedom'' is: <math>df={( SE^2(\overline{x})+SE^2(\overline{x}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } ={( 0.0335^2+0.074^2)^2 \over {0.0335^4 \over 8} + {0.074^4 \over 8} } = 11.03.</math> So, we round down '''df=11'''.  
+
: The ''degrees of freedom'' is: <math>df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } ={( 0.0335^2+0.074^2)^2 \over {0.0335^4 \over 8} + {0.074^4 \over 8} } = 11.03.</math> So, we round down '''df=11'''.  
  
 
: <math>t_o= {\overline{x}-\overline{y}- \mu_o \over  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}</math>
 
: <math>t_o= {\overline{x}-\overline{y}- \mu_o \over  \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}</math>
  
: <math>t_o = {7.9322-7.460-0 \over  0.081}=58.27</math>.
+
: <math>t_o = {7.9322-7.460-0 \over  0.081}=5.829</math>.
  
: <math>p-value=P(T_{(df=11)}>T_o=5.827)=0.00003</math> for this (two-sided) test. Therefore, we '''can reject''' the null hypothesis at <math>\alpha=0.05</math>! The left white area at the tails of the T(df=11) distribution depict graphically the probability of interest, which represents the strenght of the evidence (in the data) against the Null hypothesis. In this case, this area is 0.00003, which is much smaller than the initially set [[AP_Statistics_Curriculum_2007_Hypothesis_Basics | Type I]] error <math>\alpha = 0.05</math> and we reject the null hypothesis.
+
: <math>p-value=P(T_{(df=11)}>T_o=5.829)=0.00003</math> for this (two-sided) test. Therefore, we '''can reject''' the null hypothesis at <math>\alpha=0.05</math>! The left white area at the tails of the T(df=11) distribution depicts graphically the probability of interest, which represents the strength of the evidence (in the data) against the Null hypothesis. In this case, this area is 0.00003, which is much smaller than the initially set [[AP_Statistics_Curriculum_2007_Hypothesis_Basics | Type I]] error <math>\alpha = 0.05</math> and we reject the null hypothesis.
 
<center>[[Image:SOCR_EBook_Dinov_Infer_2Means_Indep_020908_Fig4.jpg|600px]]</center>
 
<center>[[Image:SOCR_EBook_Dinov_Infer_2Means_Indep_020908_Fig4.jpg|600px]]</center>
  
Line 113: Line 113:
  
 
====Independent T-test Validity====
 
====Independent T-test Validity====
Both the confidence intervals and the hypothesis testing methods in the independent-sample design require Normality of both samples. If the sample sizes are large (say >50), Normality is not as critical, as the [[AP_Statistics_Curriculum_2007_Limits_CLT | CLT]] implies the sampling disitributions of the means are approximately Normal. If these parametric assumptions are invalid we must use a [[AP_Statistics_Curriculum_2007_NonParam_2MeansIndep | not-parametric (distribution free test)]], even if the latter is less powerful.
+
Both the confidence intervals and the hypothesis testing methods in the independent-sample design require Normality of both samples. If the sample sizes are large (say >50), Normality is not as critical, as the [[AP_Statistics_Curriculum_2007_Limits_CLT | CLT]] implies the sampling distributions of the means are approximately Normal. If these parametric assumptions are invalid we must use a [[AP_Statistics_Curriculum_2007_NonParam_2MedianIndep | non-parametric (distribution free test)]], even if the latter is less powerful.
  
The plots below indicate that Normal assumptions are not unreasonable for these data, and hence we may be justified in using the two independent sample T-test in this case.
+
The plots below indicate that Normal assumptions are not unreasonable for these data, and hence we may be justified in using the two independent sample T-tests in this case.
  
 
* [[AP_Statistics_Curriculum_2007_Normal_Prob#Assessing_Normality | QQ-Normal plot]] of the first sample:
 
* [[AP_Statistics_Curriculum_2007_Normal_Prob#Assessing_Normality | QQ-Normal plot]] of the first sample:
Line 125: Line 125:
 
<hr>
 
<hr>
  
===References===
+
===Detailed specifications for independent-sample design studies===
 +
There are several different situations that arise in studies involving independent samples inference. These cases are separated by whether the sample sizes are equal and the sample variances (approximately) '''equivalent''' (i.e., <math>\frac{\sigma_1}{2}\le \sigma_2 \le 2\sigma_1</math>). The table below illustrates the differences in the statistical inference in each of these situations. This table uses the following notation:
 +
: Indexes 1 and 2 denote group one and group two, respectively
 +
: ''T'' is the test statistics
 +
: ''S'' is the standard error of the difference between two means
 +
: df = degrees of freedom
 +
: $n_1$ and $n_2$ are the number of observations in each group.
 +
<center>
 +
{| class="wikitable" style="text-align:center; width:55%" border="1"
 +
|-
 +
! colspan="2" rowspan="2"|Design || colspan="2"|Sample Size
 +
|-
 +
! Equal  || Unequal (different)
 +
|-
 +
| rowspan="2"|Population Variance || Equivalent ||
 +
<math>T = \frac{\bar {X}_1 - \bar{X}_2}{S_{X_1X_2} \cdot \sqrt{\frac{2}{n}}}</math>
 +
where
 +
<math>\ S_{X_1X_2} = \sqrt{\frac{1}{2}(S_{X_1}^2+S_{X_2}^2)}</math> is the [http://en.wikipedia.org/wiki/Pooled_standard_deviation pooled standard deviation] and df =2<math>n_1</math>−2
 +
||
 +
<math>T = \frac{\bar {X}_1 - \bar{X}_2}{S_{X_1X_2} \cdot \sqrt{\frac{1}{n_1}+\frac{1}{n_2}}}</math>, where <math> S_{X_1X_2} = \sqrt{\frac{(n_1-1)S_{X_1}^2+(n_2-1)S_{X_2}^2}{n_1+n_2-2}}</math> is an estimator of the common standard deviation of the two samples and df=''n''<sub>1</sub>&nbsp;+&nbsp;''n''<sub>2</sub>&nbsp;&minus;&nbsp;2
 +
|-
 +
| Non-equivalent (different) || colspan="2"|
 +
<math>T = {\overline{X}_1 - \overline{X}_2 \over s_{\overline{X}_1 - \overline{X}_2}}</math>, where <math>s_{\overline{X}_1 - \overline{X}_2} = \sqrt{{s_1^2 \over n_1} + {s_2^2  \over n_2}}
 +
</math> (not a pooled variance), <math> \mathrm{df} = \frac{(s_1^2/n_1 + s_2^2/n_2)^2}{(s_1^2/n_1)^2/(n_1-1) + (s_2^2/n_2)^2/(n_2-1)}</math>, which is called the [http://en.wikipedia.org/wiki/Welch%E2%80%93Satterthwaite_equation Welch–Satterthwaite equation]
 +
|}
 +
</center>
 +
 
 +
===Effect-Sizes===
 +
Effect size refers to a statistic calculated from a sample of data and are most commonly estimated with their corresponding errors. Sample-driven estimates of effect sizes are different from test statistics used in hypothesis testing as the former evaluate the strength of an apparent relationship and the latter assign a significance level reflecting whether the relationship could be due to chance. The effect size does not determine the significance level, or vice-versa. Given a large enough sample size, a statistical comparison will always show a significant difference unless the population effect size is trivial (i.e., zero). For example, a sample Pearson correlation coefficient of 0.1 would be strongly statistically significant when the sample size is 1000, but would be insignificant if the sample size is just 10. Reporting only the significant p-value from this analysis could be misleading if a correlation of 0.1 is too small to be of interest in a particular application.
 +
 
 +
===Cohen's d===
 +
Cohen's ''d'' is the normalized difference between two means, i.e., the difference of means divided by a standard deviation for the data, i.e.,
 +
: \(d = \frac{\bar{x}_1 - \bar{x}_2}{s}\).
 +
 
 +
Cohen's ''d'' is frequently used in estimating sample sizes.
 +
* A lower Cohen's ''d'' indicates the necessity of larger sample sizes, i.e., need to have larger cohorts to be able to assess/detect if there are between-group differences in the observed data.
 +
* The standard deviation \(s\) represents the population standard deviation (assuming the [[AP_Statistics_Curriculum_2007_Infer_BiVar#Comparing_Two_Variances_.28.5C.28.5Csigma_1.5E2_.3D_.5Csigma_2.5E2.5C.29.3F.29|groups have approximately the same variances]]).
 +
: \(s = \sqrt{\frac{(n_1-1)s^2_1 + (n_2-1)s^2_2}{n_1+n_2-2}}\), where
 +
: \(s_1^2 = \frac{1}{n_1-1} \sum_{i=1}^{n_1} (x_{1,i} - \bar{x}_1)^2\) and \(s_2^2 = \frac{1}{n_2-1} \sum_{i=1}^{n_2} (x_{2,i} - \bar{x}_2)^2\).
 +
This is referred to as the maximum likelihood estimator "Cohen's ''d''", and it is related to [http://en.wikipedia.org/wiki/Effect_size#Hedges.27s_g Hedges's g] by a scaling factor.
 +
 
 +
* Example: observing the heights of male and female visitors in England, the data ([http://www.ic.nhs.uk/pubs/hlthsvyeng2004upd Aaron,Kromrey,& Ferron, 1998, November] from a 2004 UK representative sample of 2436 men and 3311 women): '''Men''': mean height = 1.75m; standard deviation = 89.9mm, and '''Women''': mean height = 1.61m; standard deviation = 69.0mm. The effect size (using Cohen's ''d'') would equal 1.72 (95% confidence interval: 1.66 – 1.78). This effect-size is large indicating that there is a consistent height difference, on average, between men and women visiting England.
 +
 
 +
===[[EBook_Problems_Infer_2Means_Indep|Problems]]===
  
 
<hr>
 
<hr>
* SOCR Home page: http://www.socr.ucla.edu
+
* SOCR Home page: http://www.socr.umich.edu
  
{{translate|pageName=http://wiki.stat.ucla.edu/socr/index.php?title=AP_Statistics_Curriculum_2007_Infer_2Means_Indep}}
+
{{translate|pageName=http://wiki.socr.umich.edu/index.php?title=AP_Statistics_Curriculum_2007_Infer_2Means_Indep}}

Latest revision as of 12:20, 16 November 2014

General Advance-Placement (AP) Statistics Curriculum - Inferences about Two Means: Independent Samples

In the previous section we discussed the inference on two paired random samples. Now, we show how to do inference on two independent samples.

Indepenent Samples Designs

Independent samples designs refer to design of experiments or observations where all measurements are individually independent from each other within their groups and the groups are independent. The groups may be drawn from different populations with different distribution characteristics.

Background

  • Recall that for a random sample {\(X_1, X_2, X_3, \cdots , X_n\)} of the process, the population mean may be estimated by the sample average, \(\overline{X_n}={1\over n}\sum_{i=1}^n{X_i}\).
  • The standard error of \(\overline{x}\) is given by \({{1\over \sqrt{n}} \sqrt{\sum_{i=1}^n{(x_i-\overline{x})^2\over n-1}}}\).

Analysis Protocol for Independent Designs

To study independent samples, we would like to examine the differences between two group means. Suppose {\(X_1, X_2, X_3, \cdots , X_{n_1}\)} and {\(Y_1, Y_2, Y_3, \cdots , Y_{n_2}\)} represent the two independent samples. Then we want to study the differences of the two group means relative to the internal sample variations. If the two samples were drawn from populations that had different centers, then we would expect that the two sample averages will be distinct.

Large Samples

  • Significance Testing: We have a standard null-hypothesis \(H_o: \mu_X -\mu_Y = \mu_o\) (e.g., \(\mu_o=0\)). Then the test statistics is:

\[Z_o = {\overline{x}-\overline{y}-\mu_o \over SE(\overline{x}-\overline{y})} \sim N(0,1)\]. \[z_o= {\overline{x}-\overline{y} \over \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}\]

  • Confidence Intervals\[(1-\alpha)100%\] confidence interval for \(\mu_1-\mu_2\) will be

\[CI(\alpha): \overline{x}-\overline{y} \pm z_{\alpha\over 2} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm z_{\alpha\over 2} \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}\]. Note that the \(SE(\overline{x} -\overline{y})=\sqrt{SE^2(\overline{x})+SE^2(\overline{y})}\), as the samples are independent. Also, \(z_{\alpha\over 2}\) is the critical value for a Standard Normal distribution at \({\alpha\over 2}\).

Small Samples

  • Significance Testing: Again, we have a standard null-hypothesis \(H_o: \mu_X -\mu_Y = \mu_o\) (e.g., \(\mu_o=0\)). Then the test statistics is:

\[T_o = {\overline{x}-\overline{y}-\mu_o \over SE(\overline{x}-\overline{y})} \sim T_{(df)}\].

The degrees of freedom (df) is\[df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.\] Always round down the degrees of freedom to the next smaller integer.

\[t_o= {\overline{x}-\overline{y}- \mu_o \over \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}\]

  • Confidence Intervals\[(1-\alpha)100%\] confidence interval for \(\mu_1-\mu_2\) will be

\[CI(\alpha): \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}\]. Note that the \(SE(\overline{x} -\overline{y})=\sqrt{SE^2(\overline{x})+SE^2(\overline{y})}\), as the samples are independent.

The degrees of freedom is\[df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } \approx n_1+n_2-2.\] Always round down the degrees of freedom to the next smaller integer. Also, \(t_{df, {\alpha\over 2}}\) is the critical value for a Student's T distribution at \({\alpha\over 2}\).

Example

Nine observations of surface soil pH were made at two different locations. Does the data suggest that the true mean soil pH values differs for the two locations? Formulate testable hypothesis and make inference about the effect of the treatment at \(\alpha=0.05\). Check any necessary assumptions for the validity of your test.

Data in row format

Location 1 8.1,7.89,8,7.85,8.01,7.82,7.99,7.8,7.93
Location 2 7.85,7.3,7.73,7.27,7.58,7.27,7.5,7.23,7.41

Data in column format

Index Location 1 Location 2
1 8.10 7.85
2 7.89 7.30
3 8.00 7.73
4 7.85 7.27
5 8.01 7.58
6 7.82 7.27
7 7.99 7.50
8 7.80 7.23
9 7.93 7.41
Mean 7.9322 7.4600
SD 0.1005 0.2220

Exploratory Data Analysis

We begin first by exploring the data visually using various SOCR EDA Tools.

SOCR EBook Dinov Infer 2Means Indep 020908 Fig1.jpg
SOCR EBook Dinov Infer 2Means Indep 020908 Fig2.jpg

Inference

  • Null Hypothesis\[H_o: \mu_{1}-\mu_{2}=0\]
  • (Two-sided) alternative Research Hypotheses\[H_1: \mu_{1} -\mu_{2} \not= 0\].
  • Test statistics: We can use the sample summary statistics to compute the degrees of freedom and the T-statistic
The degrees of freedom is\[df={( SE^2(\overline{x})+SE^2(\overline{y}))^2 \over {SE^4(\overline{x}) \over n_1-1} + {SE^4(\overline{y}) \over n_2-1} } ={( 0.0335^2+0.074^2)^2 \over {0.0335^4 \over 8} + {0.074^4 \over 8} } = 11.03.\] So, we round down df=11.

\[t_o= {\overline{x}-\overline{y}- \mu_o \over \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}}\]

\[t_o = {7.9322-7.460-0 \over 0.081}=5.829\].

\[p-value=P(T_{(df=11)}>T_o=5.829)=0.00003\] for this (two-sided) test. Therefore, we can reject the null hypothesis at \(\alpha=0.05\)! The left white area at the tails of the T(df=11) distribution depicts graphically the probability of interest, which represents the strength of the evidence (in the data) against the Null hypothesis. In this case, this area is 0.00003, which is much smaller than the initially set Type I error \(\alpha = 0.05\) and we reject the null hypothesis.

SOCR EBook Dinov Infer 2Means Indep 020908 Fig4.jpg
SOCR EBook Dinov Infer 2Means Indep 020908 Fig3.jpg
  • \(95%=(1-0.05)100%\) (\(\alpha=0.05\)) Confidence interval:

\[CI(\mu_{1}-\mu_{2})\]\[\overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} SE(\overline{x}-\overline{y})= \overline{x}-\overline{y} \pm t_{df, {\alpha\over 2}} \sqrt{{1\over {n_1}} {\sum_{i=1}^{n_1}{(x_i-\overline{x})^2\over n_1-1}} + {1\over {n_2}} {\sum_{i=1}^{n_2}{(y_i-\overline{y})^2\over n_2-1}}}\]

\[CI: {7.932-7.460 \pm 2.201\times 0.081 }= [0.294 ; 0.650].\]

Conclusion

These data show that there is a statistically significant mean difference in the pH of Location 1 and Location 2 (p < 0.001).


Independent T-test Validity

Both the confidence intervals and the hypothesis testing methods in the independent-sample design require Normality of both samples. If the sample sizes are large (say >50), Normality is not as critical, as the CLT implies the sampling distributions of the means are approximately Normal. If these parametric assumptions are invalid we must use a non-parametric (distribution free test), even if the latter is less powerful.

The plots below indicate that Normal assumptions are not unreasonable for these data, and hence we may be justified in using the two independent sample T-tests in this case.

SOCR EBook Dinov Infer 2Means Indep 020908 Fig6.jpg
SOCR EBook Dinov Infer 2Means Indep 020908 Fig5.jpg

Detailed specifications for independent-sample design studies

There are several different situations that arise in studies involving independent samples inference. These cases are separated by whether the sample sizes are equal and the sample variances (approximately) equivalent (i.e., \(\frac{\sigma_1}{2}\le \sigma_2 \le 2\sigma_1\)). The table below illustrates the differences in the statistical inference in each of these situations. This table uses the following notation:

Indexes 1 and 2 denote group one and group two, respectively
T is the test statistics
S is the standard error of the difference between two means
df = degrees of freedom
$n_1$ and $n_2$ are the number of observations in each group.
Design Sample Size
Equal Unequal (different)
Population Variance Equivalent

\(T = \frac{\bar {X}_1 - \bar{X}_2}{S_{X_1X_2} \cdot \sqrt{\frac{2}{n}}}\) where \(\ S_{X_1X_2} = \sqrt{\frac{1}{2}(S_{X_1}^2+S_{X_2}^2)}\) is the pooled standard deviation and df =2\(n_1\)−2

\(T = \frac{\bar {X}_1 - \bar{X}_2}{S_{X_1X_2} \cdot \sqrt{\frac{1}{n_1}+\frac{1}{n_2}}}\), where \( S_{X_1X_2} = \sqrt{\frac{(n_1-1)S_{X_1}^2+(n_2-1)S_{X_2}^2}{n_1+n_2-2}}\) is an estimator of the common standard deviation of the two samples and df=n1 + n2 − 2

Non-equivalent (different)

\(T = {\overline{X}_1 - \overline{X}_2 \over s_{\overline{X}_1 - \overline{X}_2}}\), where \(s_{\overline{X}_1 - \overline{X}_2} = \sqrt{{s_1^2 \over n_1} + {s_2^2 \over n_2}} \) (not a pooled variance), \( \mathrm{df} = \frac{(s_1^2/n_1 + s_2^2/n_2)^2}{(s_1^2/n_1)^2/(n_1-1) + (s_2^2/n_2)^2/(n_2-1)}\), which is called the Welch–Satterthwaite equation

Effect-Sizes

Effect size refers to a statistic calculated from a sample of data and are most commonly estimated with their corresponding errors. Sample-driven estimates of effect sizes are different from test statistics used in hypothesis testing as the former evaluate the strength of an apparent relationship and the latter assign a significance level reflecting whether the relationship could be due to chance. The effect size does not determine the significance level, or vice-versa. Given a large enough sample size, a statistical comparison will always show a significant difference unless the population effect size is trivial (i.e., zero). For example, a sample Pearson correlation coefficient of 0.1 would be strongly statistically significant when the sample size is 1000, but would be insignificant if the sample size is just 10. Reporting only the significant p-value from this analysis could be misleading if a correlation of 0.1 is too small to be of interest in a particular application.

Cohen's d

Cohen's d is the normalized difference between two means, i.e., the difference of means divided by a standard deviation for the data, i.e.,

\(d = \frac{\bar{x}_1 - \bar{x}_2}{s}\).

Cohen's d is frequently used in estimating sample sizes.

  • A lower Cohen's d indicates the necessity of larger sample sizes, i.e., need to have larger cohorts to be able to assess/detect if there are between-group differences in the observed data.
  • The standard deviation \(s\) represents the population standard deviation (assuming the groups have approximately the same variances).
\(s = \sqrt{\frac{(n_1-1)s^2_1 + (n_2-1)s^2_2}{n_1+n_2-2}}\), where
\(s_1^2 = \frac{1}{n_1-1} \sum_{i=1}^{n_1} (x_{1,i} - \bar{x}_1)^2\) and \(s_2^2 = \frac{1}{n_2-1} \sum_{i=1}^{n_2} (x_{2,i} - \bar{x}_2)^2\).

This is referred to as the maximum likelihood estimator "Cohen's d", and it is related to Hedges's g by a scaling factor.

  • Example: observing the heights of male and female visitors in England, the data (Aaron,Kromrey,& Ferron, 1998, November from a 2004 UK representative sample of 2436 men and 3311 women): Men: mean height = 1.75m; standard deviation = 89.9mm, and Women: mean height = 1.61m; standard deviation = 69.0mm. The effect size (using Cohen's d) would equal 1.72 (95% confidence interval: 1.66 – 1.78). This effect-size is large indicating that there is a consistent height difference, on average, between men and women visiting England.

Problems




Translate this page:

(default)
Uk flag.gif

Deutsch
De flag.gif

Español
Es flag.gif

Français
Fr flag.gif

Italiano
It flag.gif

Português
Pt flag.gif

日本語
Jp flag.gif

България
Bg flag.gif

الامارات العربية المتحدة
Ae flag.gif

Suomi
Fi flag.gif

इस भाषा में
In flag.gif

Norge
No flag.png

한국어
Kr flag.gif

中文
Cn flag.gif

繁体中文
Cn flag.gif

Русский
Ru flag.gif

Nederlands
Nl flag.gif

Ελληνικά
Gr flag.gif

Hrvatska
Hr flag.gif

Česká republika
Cz flag.gif

Danmark
Dk flag.gif

Polska
Pl flag.png

România
Ro flag.png

Sverige
Se flag.gif