1 / 58

Introduction to Statistics

Introduction to Statistics. Colm O’Dushlaine. Neuropsychiatric Genetics, TCD codushlaine@gmail.com. Overview. Descriptive Statistics & Graphical Presentation of Data Statistical Inference Hypothesis Tests & Confidence Intervals T-tests (Paired/Two-sample)

bronwen
Download Presentation

Introduction to Statistics

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Introduction to Statistics Colm O’Dushlaine Neuropsychiatric Genetics, TCD codushlaine@gmail.com

  2. Overview • Descriptive Statistics & Graphical Presentation of Data • Statistical Inference • Hypothesis Tests & Confidence Intervals • T-tests (Paired/Two-sample) • Regression (SLR & Multiple Regression) • ANOVA/ANCOVA • Intended as an interview. Will provide slides after lectures • What’s in the lectures?...

  3. Lecture 1 Lecture 2 Lecture 3 Lecture 4 Descriptive Statistics and Graphical Presentation of Data • Terminology • Frequency Distributions/Histograms • Measures of data location • Measures of data spread • Box-plots • Scatter-plots • Clustering (Multivariate Data)

  4. Lecture 1 Lecture 2 Lecture 3 Lecture 4 Statistical Inference • Distributions & Densities • Normal Distribution • Sampling Distribution & Central Limit Theorem • Hypothesis Tests • P-values • Confidence Intervals • Two-Sample Inferences • Paired Data

  5. Lecture 1 Lecture 2 Lecture 3 Lecture 4 Sample Inferences • Two-Sample Inferences • Paired t-test • Two-sample t-test • Inferences for more than two samples • One-way ANOVA • Two-way ANOVA • Interactions in Two-way ANOVA • DataDesk demo

  6. Lecture 1 Lecture 2 Lecture 3 Lecture 4 • Regression • Correlation • Multiple Regression • ANCOVA • Normality Checks • Non-parametrics • Sample Size Calculations • Useful tools and websites

  7. FIRST, A REALLY USEFUL SITE Explanations of outputs Videos with commentary Help with deciding what test to use with what data

  8. TerminologyPopulations & Samples • Population: the complete set of individuals, objects or scores of interest. • Often too large to sample in its entirety • It may be real or hypothetical (e.g. the results from an experiment repeated ad infinitum) • Sample: A subset of the population. • A sample may be classified asrandom(each member has equal chance of being selected from a population) or convenience (what’s available). • Random selection attempts to ensure the sample is representative of the population.

  9. Variables • Variables are the quantities measured in a sample.They may be classified as: • Quantitativei.e. numerical • Continuous (e.g. pH of a sample, patient cholesterol levels) • Discrete (e.g. number of bacteria colonies in a culture) • Categorical • Nominal (e.g. gender, blood group) • Ordinal(ranked e.g. mild, moderate or severe illness). Often ordinal variables are re-coded to be quantitative.

  10. Variables • Variables can be further classified as: • Dependent/Response. Variable of primary interest (e.g. blood pressure in an antihypertensive drug trial). Not controlled by the experimenter. • Independent/Predictor • called a Factor whencontrolled by experimenter. It is often nominal (e.g. treatment) • Covariate when not controlled. • If the value of a variable cannot be predicted in advance then the variable is referred to as a random variable

  11. Parameters & Statistics • Parameters: Quantities that describe a population characteristic. They are usually unknown and we wish to make statistical inferences about parameters. Different to perimeters. • Descriptive Statistics: Quantities and techniques used to describe a sample characteristic or illustrate the sample data e.g. mean, standard deviation, box-plot

  12. 2. Frequency Distributions • An (Empirical) Frequency Distribution or Histogram for a continuous variable presents the counts of observations grouped within pre-specified classes or groups • A Relative Frequency Distribution presents the corresponding proportions of observations within the classes • A Barchart presents the frequencies for a categorical variable

  13. Example – Serum CK • Blood samples taken from 36 male volunteers as part of a study to determine the natural variation in CK concentration. • The serum CK concentrations were measured in (U/I) are as follows:

  14. Serum CK Data for 36 male volunteers

  15. Relative Frequency Table

  16. Frequency Distribution

  17. Relative Frequency Distribution Mode Shaded area is percentage of males with CK values between 60 and 100 U/l, i.e. 42%. Right tail (skewed) Left tail

  18. 3. Measures of Central Tendency (Location) Measures of location indicate where on the number line the data are to be found. Common measures of location are: (i)the ArithmeticMean, (ii)theMedian, and (iii)the Mode

  19. The Mean • Let x1,x2,x3,…,xnbe the realised valuesof a random variable X, from a sample of size n. The sample arithmetic mean is defined as:

  20. Example Example 2: The systolic blood pressure of seven middle aged men were as follows: 151, 124, 132, 170, 146, 124 and 113. The mean is

  21. The Median and Mode • If the sample data are arranged in increasing order, the median is • the middle value if n is an odd number, or • midway between the two middle values if n is an even number • The mode is the most commonly occurring value.

  22. Example 1 – n is odd The reordered systolic blood pressure data seen earlier are: 113, 124, 124, 132, 146, 151, and 170. The Median is the middle value of the ordered data, i.e. 132. Two individuals have systolic blood pressure = 124 mm Hg, so the Mode is 124.

  23. Example 2 – n is even Six men with high cholesterol participated in a study to investigate the effects of diet on cholesterol level. At the beginning of the study, their cholesterol levels (mg/dL) were as follows: 366, 327, 274, 292, 274 and 230. Rearrange the data in numerical order as follows: 230, 274, 274, 292, 327 and 366. The Median is half way between the middle two readings, i.e. (274+292)  2 = 283. Two men have the same cholesterol level- the Mode is 274.

  24. Mean versus Median • Large sample values tend to inflate the mean. This will happen if the histogram of the data is right-skewed. • The median is not influenced by large sample values and is a better measure of centrality if the distribution is skewed. • Note if mean=median=mode then the data are said to be symmetrical • e.g. In the CK measurement study, the sample mean = 98.28. The median = 94.5, i.e. mean is larger than median indicating that mean is inflated by two large data values 201 and 203.

  25. 4. Measures of Dispersion • Measures of dispersion characterise how spread out the distribution is, i.e., how variable the data are. • Commonly used measures of dispersion include: • Range • Variance & Standard deviation • Coefficient of Variation (or relative standard deviation) • Inter-quartile range

  26. Range • the sample Range is the difference between the largest and smallest observations in the sample • easy to calculate; • Blood pressure example: min=113 and max=170, so the range=57 mmHg • useful for “best” or “worst” case scenarios  • sensitive to extreme values 

  27. Sample Variance • The sample variance, s2, is the arithmetic mean of the squared deviations from the sample mean: >

  28. Standard Deviation • The sample standard deviation, s, is the square-root of the variance • s has the advantage of being in the same units as the original variable x

  29. Example

  30. Example (contd.) Therefore,

  31. Coefficient of Variation • The coefficient of variation (CV) or relative standard deviation (RSD) is the sample standard deviation expressed as a percentage of the mean, i.e. • The CV is not affected by multiplicative changes in scale • Consequently, a useful way of comparing the dispersion of variables measured on different scales

  32. Example The CV of the blood pressure data is: i.e., the standard deviation is 14.3% as large as the mean.

  33. Inter-quartile range • The Median divides a distribution into two halves. • The first and third quartiles (denoted Q1 and Q3) are defined as follows: • 25% of the data lie below Q1 (and 75% is above Q1), • 25% of the data lie above Q3 (and 75% is below Q3) • The inter-quartile range (IQR) is the difference between the first and third quartiles, i.e. IQR = Q3- Q1

  34. Example The ordered blood pressure data is: 113 124 124 132 146 151 170 Q1 Q3 Inter Quartile Range (IQR) is 151-124 = 27

  35. 60% of slides complete!

  36. 5. Box-plots • A box-plot is a visual description of the distribution based on • Minimum • Q1 • Median • Q3 • Maximum • Useful for comparing large sets of data

  37. Example 1 The pulse rates of 12 individuals arranged in increasing order are: 62, 64, 68, 70, 70, 74, 74, 76, 76, 78, 78, 80 Q1=(68+70)2 = 69, Q3=(76+78)2 = 77 IQR = (77 – 69) = 8

  38. Example 1: Box-plot

  39. Example 2: Box-plots of intensities from 11 gene expression arrays

  40. Outliers • An outlier is an observation which does not appear to belong with the other data • Outliers can arise because of a measurement or recording error or because of equipment failure during an experiment, etc. • An outlier might be indicative of a sub-population, e.g. an abnormally low or high value in a medical test could indicate presence of an illness in the patient.

  41. Outlier Boxplot • Re-define the upper and lower limits of the boxplots (the whisker lines) as: Lower limit = Q1-1.5IQR, and Upper limit = Q3+1.5IQR • Note that the lines may not go as far as these limits • If a data point is < lower limit or > upper limit, the data point is considered to be an outlier.

  42. Example – CK data outliers

  43. 6. Scatter-plot • Displays the relationship between two continuous variables • Useful in the early stage of analysis when exploring data and determining is a linear regression analysis is appropriate • May show outliers in your data

  44. Example 1: Age versus Systolic Blood Pressure in a Clinical Trial

  45. Example 2: Up-regulation/Down-regulation of gene expression across an array (Control Cy5 versus Disease Cy3)

  46. Example of a Scatter-plot matrix (multiple pair-wise plots)

  47. Other graphical representations • Dot-Plots, Stem-and-leaf plots • Not visually appealing • Pie-chart • Visually appealing, but hard to compare two datasets. Best for 3 to 7 categories. A total must be specified. • Violin-plots • =boxplot+smooth density • Nice visual of data shape

  48. Multivariate Data • Clustering is useful for visualising multivariate data and uncovering patterns, often reducing its complexity • Clustering is especially useful for high-dimensional data (p>>n): hundreds or perhaps thousands of variables • An obvious areas of application are gel electrophoresis and microarray experiments where the variables are protein abundances or gene expression ratios

  49. 7. Clustering • Aim: Find groups of samples or variables sharing similiarity • Clustering requires a definition of distance between objects, quantifying a notion of (dis)similarity • Points are grouped on the basis on minimum distance apart (distance measures) • Once a pair are grouped, they are combined into a single point (using a linkage method) e.g. take their average. The process is then repeated.

  50. Clustering • Clustering can be applied to rows or columns of a data set (matrix) i.e. to the samples or variables • A tree can be constructed with branch length proportional to distances between linked clusters, called a Dendrogram • Clustering is an example of unsupervisedlearning: No use is made of sample annotations i.e. treatment groups, diagnosis groups

More Related