AP Statistics Curriculum 2007 IntroDesign

From SOCR
Revision as of 21:02, 16 June 2007 by IvoDinov (talk | contribs)
Jump to: navigation, search

General Advance-Placement (AP) Statistics Curriculum - Design and Experiments

Design and Experiments

Design of experiments refers of the blueprint for planning a study or experiment, performing the data collection protocol and controlling the study parameters for accuracy and consistency. Design of experiments only makes sense in studies where variation, chance and uncertainly are present and unavoidable. Data, or information, is typically collected in regard to a specific process or phenomenon being studied to investigate the effects of some controlled variables (independent variables or predictors) on other observed measurements (responses or dependent variables). Both types of variables are associated with specific observational units (living beings, components, objects, materials, etc.)

Approach

The following are the most common components used in Experimental Design.

  • Comparison: To make inference about effects, associations or predictions, one typically has to compare different groups subjected to distinct conditions. This allows contrasting observed responses and underlying group differences which ultimately may lead to inference on relations and influence between controlled and observed variables.
  • Randomization: The second fundamental design principle is randomization. It requires that we make allocation of (controlled variables) treatments to units using some random mechanism. This will simply guarantees that effects that may be present is the units, but not incorporated in the model, are equidistributed amongst all groups and are therefore unlikely to significantly effect our group comparisons at the end of the statistical inference or analysis (as these effects, if present, will be similar within each group).
  • Replication: All measurements we make, observations we acquire or data we collect is subject to variation, as there are no completely deterministic processes. As we try to make inference about the process that generated the observed data (not the sample data itself, even though our statistical analysis are data-driven and therefore based on the observed measurements), the more data we collect (unbiasly) the stronger our inference is likely to be. Therefore, repeated measurements intuitively would allow is to tame the variability associated with the phenomenon we study.
  • Blocking: Blocking is related to randomization. The difference is that we use blocking when we know a priori of certain effects of the observational units on the response measurements (e.g., when studying the effects of hormonal treatments on humans, gender plays a significant role). We arrange units into groups (blocks) that are similar to one another when we design an experiment in which certain unit characteristics are known to affect the response measurements. Blocking reduces known and irrelevant sources of variation between units and allows greater precision in the estimation of the source of variation in the study.
  • Orthogonality: Orthogonality allows division of complex relations, variation into separate (independent/orthogonal) contrasts, or factors, that can be studies efficiently and autonomously. Often, these contrasts may be represented by vectors where sets of orthogonal contrasts are uncorrelated and may be independently distributed. Independence implies that each orthogonal contrast provides complementary information to other contrasts (i.e., other treatments). The goal is to completely decompose the variance or the relations of the observed measurements into independent components (e.g., like Taylor expension allows polynomial desomposition of smooth functions, where the polynomial base functions are easy to differentiate, integrate, etc.) This will, of course, allow easier interpretation of the statistical analysis and the findings of the study.

Model Validation

Checking/affirming underlying assumptions.

  • TBD

Computational Resources: Internet-based SOCR Tools

Examples & Hands-on activities

  • A study of aortic valve-sparing repair*
    • This study sought to establish whether there was a difference in outcome after aortic valve repair with autologous pericardial leaflet extension in acquired versus congenital valvular disease. One 128 patients underwent reparative aortic valve surgery at UCLA from 1997 through 2005 for acquired or congenital aortic valve disease. The acquired group (43/128) (34%) had a mean age of 56.4 \(\pm\) 20.3 years (range, 7.8—84.6 years) and the congenital group (85/128) (66%) had a mean age of 16.9 � \(\pm\)19.2 years (range, 0.3—82 years). The endpoints of the study were mortality and reoperation rates.
    • In this case the units are heart disease patients. These were split into two groups (acquired or congenital) and locked by gender (male/female). The treatment allied on the two groups was aortic valve repair with autologous pericardial leaflet extension.

References




Translate this page:

(default)
Uk flag.gif

Deutsch
De flag.gif

Español
Es flag.gif

Français
Fr flag.gif

Italiano
It flag.gif

Português
Pt flag.gif

日本語
Jp flag.gif

България
Bg flag.gif

الامارات العربية المتحدة
Ae flag.gif

Suomi
Fi flag.gif

इस भाषा में
In flag.gif

Norge
No flag.png

한국어
Kr flag.gif

中文
Cn flag.gif

繁体中文
Cn flag.gif

Русский
Ru flag.gif

Nederlands
Nl flag.gif

Ελληνικά
Gr flag.gif

Hrvatska
Hr flag.gif

Česká republika
Cz flag.gif

Danmark
Dk flag.gif

Polska
Pl flag.png

România
Ro flag.png

Sverige
Se flag.gif