SMHS CenterSpreadShape
Contents
- 1 Scientific Methods for Health Sciences - Measures of Centrality, Variability and Shape
Scientific Methods for Health Sciences - Measures of Centrality, Variability and Shape
Overview
Three main features of sample data are commonly reported as critical in understanding and interpreting the population, or process, the data represents. These include Center, Spread and Shape. The main measures of centrality are Mean, Median and Mode(s). Common measures of variability include the range, the variance, the standard deviation, and mean absolute deviation. The shape of a (sample or population) distribution is an important characterization of the process and its intrinsic properties.
Motivation
Suppose you want to synthesize the information contained in a large dataset (e.g., SOCR Data) and communicate this summary quickly and effectively to someone else. One approach is to estimate the point of centrality, the amount of dispersion and the shape of the histogram of the data and use these as concise intrinsic characteristics of the process generating the observed data.
Theory
Measures of Centrality
The main measures of centrality are mean, median and mode.
Suppose we are interested in the survival times for some terminally ill patients undergoing an experimental intervention. If we conduct an ethical IRB-approved study including nine male patients and observe their survival times, as shown below. How can we determine the expected mean (center) survival time?
74 | 78 | 106 | 80 | 68 | 64 | 60 | 76 | 98 |
Mean
The sample-mean is the arithmetic average of a finite sample of numbers. For instance, the mean of the sample \(x_1, x_2, x_3, \cdots, x_{n-1}, x_n\). Short hand notation\[\{x_{i}\}_{i=1}^n,\] is given by\[\bar{x}={1\over n}\sum_{i=1}^{n}{x_{i}}.\]
In the survival time example, the sample-mean is calculated as follows\[\overline{y} = {1 \over 9} (74+78+106+80+68+64+60+76+98)=78.22 days.\]
Median
The sample-median can be thought of as the point that divides a distribution in half (50/50). The following steps are used to find the sample-median:
- Arrange the data in ascending order
- If the sample size is odd, the median is the middle value of the ordered collection
- If the sample size is even, the median is the average of the middle two values in the ordered collection.
For the survival time example data above we have:
- Ordered data:
60 | 64 | 68 | 74 | 76 | 78 | 80 | 98 | 106 |
- \(Median = 76\).
Mode(s)
The modes represent the most frequently occurring values (The numbers that appear the most). The term mode is applied both to probability distributions and to collections of experimental data.
For instance, for the Hot dogs data file, there appears to be three modes for the calorie variable. This is presented by the histogram of the Calorie content of all hotdogs, shown in the image below. Note the clear separation of the calories into three distinct sub-populations - the highest points in these three sub-populations are the three modes for these data.
Resistance
A statistic is said to be resistant if the value of the statistic is relatively unchanged by changes in a small portion of the data. Referencing the formulas for the median, mean and mode which statistic seems to be more resistant?
If you remove the student with the long jump distance of 106 and recalculate the median and mean, which one is altered less (therefore is more resistant)? Notice that the mean is very sensitive to outliers and atypical observations, and hence less resistant than the median.
The following two sample measures of population centrality estimate resemble the calculations of the mean, however they are much more resistant to change in the presence of outliers.
K-times trimmed mean
\(\bar{y}_{t,k}={1\over n-2k}\sum_{i=k+1}^{n-k}{y_{(i)}}\), where \(k\geq 0\) is the trim-factor (large k, yield less variant estimates of center), and \(y_{(i)}\) are the order statistics (small to large). That is, we remove the smallest and the largest k observations from the sample, before we compute the arithmetic average.
Winsorized k-times mean
The Winsorized k-times mean is defined similarly by \(\bar{y}_{w,k}={1\over n}( k\times y_{(k)}+\sum_{i=k+1}^{n-k}{y_{(i)}}+k\times y_{(n-k+1)})\), where \(k\geq 0\) is the trim-factor and \(y_{(i)}\) are the order statistics (small to large). In this case, before we compute the arithmetic average, we replace the k smallest and the k largest observations with the kth and (n-k)th largest observations, respectively.
Other Centrality Measures
The arithmetic mean answers the question, if all observations were equal, what would that value (center) have to be in order to achieve the same total? \[n\times \bar{x}=\sum_{i=1}^n{x_i}\]
In some situations, there is a need to think of the average in different terms, not in terms of arithmetic average.
Harmonic Mean
If we study speeds (velocities) the arithmetic mean is inappropriate. However the harmonic mean (computed differently) gives the most intuitive answer to what the "middle" is for a process. The harmonic mean answers the question if all the observations were equal, what would that value have to be in order to achieve the same sample sum of reciprocals?
- Harmonic mean\[\hat{\hat{x}}= \frac{n}{\frac{1}{x_1} + \frac{1}{x_2} + \frac{1}{x_3} + \ldots + \frac{1}{x_n}}\]
Geometric Mean
In contrast, the geometric mean answers the question if all the observations were equal, what would that value have to be in order to achieve the same sample product?
- Geometric mean\[\tilde{x}^n={\prod_{i=1}^n x_i}\]
- Alternatively\[\tilde{x}= \exp \left( \frac{1}{n} \sum_{i=1}^n\log(x_i) \right)\]
Measures of Variation and Dispersion
There are many measures of (population or sample) variation, e.g., the range, the variance, the standard deviation, mean absolute deviation, etc. These are used to assess the dispersion or spread of the population.
Range
The range is the easiest measure of dispersion to calculate, yet, perhaps not the best measure. The Range = max - min. For example, for the survival time data, the range is calculated by:
Note that the range is only sensitive to the extreme values of a sample and ignores all other information. So, two completely different distributions may have the same range.
Quartiles and IQR
The first quartile (\(Q_1\)) and the third quartile (\(Q_3\)) are defined values that split the dataset into bottom-25% vs. top-75% and bottom-75% vs. top-25%, respectively. Thus the inter-quartile range (IQR), which is the difference \(Q_3 - Q_1\), represents the central 50% of the data and can be considered as a measure of data dispersion or variation. The wider the IQR, the more variant the data.
For example, \(Q_1=68\), \(Q_3=80\) and \(IQR=Q_3-Q_1=12\), for the survival time data shown above. Thus, we expect the middle half of all survival times (for that patient cohort and clinical intervention) to be between 68 and 80 days.
Coefficient of Variation
For a given process, the coefficient of variation (CV) is defined as the ratio of the standard deviation (\(\sigma \)) to the mean (\(\mu \)): \[CV = \frac{\sigma}{\mu}.\]
Obviously, the CV is well-defined for processes with well-defined first two moments (mean and variance), but also requires a non-trivial mean (\(\mu \not= 0\)). If the CV is expressed as percentage, than this ratio is multiplied by 100. The sample coefficient of variation is computed mostly for data measured on a ratio scale. For instance, if a set of distances are measured, the standard deviation does not depend on whether the distances were measured in kilometers (metric) or miles. This is because changes in the particle/object's distances by 1 kilometer also changes its distance by 1 mile. However the mean distance of the data would differ in each measurement scale (as 1 mile is approximately 1.7 kilometers) and thus the coefficient of variation would differ. In general, the CV may not have any meaning for data on an interval scale.
The sample-coefficient of variation is computed by plugin the sample-driven estimates of the standard deviation (sample-standard deviation, \(s\), and the sample-average, \(\bar{x}\)). In image processing, the reciprocal of the coefficient of variation is μ/σ is called signal-to-noise-ratio (SNR).
Five-number summary
The five-number summary for a dataset is the 5-tuple \(\{min, Q_1, Q_2, Q_3, max\}\), containing the sample minimum, first-quartile, second-quartile (median), third-quartile, and maximum.
Variance and Standard Deviation
The logic behind the variance and standard deviation measures is to measure the difference between each observation and the mean (i.e., dispersion). Suppose we have n > 1 observations, \(\left \{ y_1, y_2, y_3, ..., y_n \right \}\). The deviation of the \(i^{th}\) measurement, \(y_i\), from the mean (\(\overline{y}\)) is defined by \((y_i - \overline{y})\).
Does the average of these deviations seem like a reasonable way to find an average deviation for the sample or the population? No, because the sum of all deviations is trivial:
To solve this problem we employ different versions of the mean absolute deviation:
In particular, the variance is defined as:
And the standard deviation is defined as:
For the survival time data, the standard deviation is:
Applications
Try to pair each of the 4 samples whose numerical summaries are reported below with one of the 4 frequency plots below. Explain your answers.
Sample | Mean | Median | StdDev |
A | 4.688 | 5.000 | 1.493 |
B | 4.000 | 4.000 | 1.633 |
C | 3.933 | 4.000 | 1.387 |
D | 4.000 | 4.000 | 2.075 |
Distribution Shape
- A distribution is unimodal if it has one mode. Unimodal distributions include:
- Bell shaped distributions (symmetric, Normal)
- Skewed right or skewed left
- We can use the mean and median to help interpret the shape of a distribution. For an unimodal distribution we have these properties:
- If mean = median, then the distribution is symmetric
- If mean > median, then the distribution is right skewed
- If mean < median, then the distribution is left skewed
- Multimodal distributions have two or more modes. Examples of multimodal distributions are:
Other Measures of Shape
This section also provides moment-based characterization of distribution shape.
Software
- SOCR Charts
- R Example
data(faithful) str(faithful) attach(faithful) # attach this data so that we can easily look at the variables inside and "detach" it when we're done with these data! length(waiting) sum(is.na(waiting)) length(waiting) - sum(is.na(waiting)) sampsize = function(x) length(x) - sum(is.na(x)) sampsize(waiting) ls()
[1] "AD_Associations_Data" "avgmod.95p" "avpred" [4] "Beetle" "cc" "Cement" [7] "confset.95p" "DF" "dsd" [10] "epsilon" "faithful" "fm1" [13] "fm2" "fmList" "globmod" [16] "hv" "MCI_Associations_Data" "ms1" [19] "ms2" "msubset" "myData" [22] "NC_Associations_Data" "newdata" "newdata1" [25] "Orthodont" "p" "rc" [28] "sampsize" "sd" "test" [31] "varying.link" "x"
mean(waiting) median(waiting) range(waiting) IQR(waiting) var(waiting) sd(waiting)
Problems
- What seems like a logical choice for the shape of the hot dog calorie data? Try looking at the histogram of the calories for the Hot-dogs dataset.
- Collect data, draw the sample histogram or dot-plot and classify the shape of the distribution accordingly. Also, if unimodal, classify symmetry (symmetric, skewed right or skewed left).
- Data collected on height of randomly sampled college students.
- Data collected on height of randomly sampled female college students.
- The salaries of all persons employed by a large university.
- The amount of time spent by students on a difficult exam.
- The grade distribution on a difficult exam.
References
- Probability and Statistics EBook Section on Centrality
- Probability and Statistics EBook Section on Variability
- Probability and Statistics EBook Section on Centrality
- SOCR Home page: http://www.socr.umich.edu
Translate this page: