1 / 89

# HCI 510 : HCI Methods I - PowerPoint PPT Presentation

HCI 510 : HCI Methods I. Statistics. HCI 510: HCI Methods I. Descriptive Statistics Inferential Statistics Significance T-Test. HCI 510: HCI Methods I. Descriptive Statistics Inferential Statistics Significance T-Test. Descriptive Statistics. Descriptive vs. Inferential statistics.

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.

HCI 510 : HCI Methods I

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

## HCI 510 : HCI Methods I

Statistics

### HCI 510: HCI Methods I

Descriptive Statistics

Inferential Statistics

Significance

T-Test

### HCI 510: HCI Methods I

Descriptive Statistics

Inferential Statistics

Significance

T-Test

• Descriptive Statistics

Descriptive vs. Inferential statistics

• Descriptive Statistics

Descriptive statistics are used to describe the basic features of the data in a study.

They provide simple summaries about the sample and the measures.

Together with simple graphics analysis, they form the basis of virtually every quantitative analysis of data

• Descriptive Statistics

Descriptive statistics are typically distinguished from inferential statistics.

With descriptive statistics you are simply describing what is or what the data shows.

With inferential statistics, you are trying to reach conclusions that extend beyond the immediate data alone.

• Descriptive Statistics

For instance, we use inferential statistics to try to infer from the sample data what the population might think.

Or, we use inferential statistics to make judgments of the probability that an observed difference between groups is a dependable one or one that might have happened by chance in this study.

Thus, we use inferential statistics to make inferences from our data to more general conditions; we use descriptive statistics simply to describe what's going on in our data.

• Descriptive Statistics

Descriptive Statistics are used to present quantitative descriptions in a manageable form.

In a research study we may have lots of measures.

Or we may measure a large number of people on any measure.

Descriptive statistics help us to simply large amounts of data in a sensible way.

Each descriptive statistic reduces lots of data into a simpler summary.

• Descriptive Statistics

For instance, consider a simple number used to summarize how well a batter is performing in baseball, the batting average.

This single number is simply the number of hits divided by the number of times at bat.

A batter who is hitting .333 is getting a hit one time in every three at bats.

One batting .250 is hitting one time in four.

The single number describes a large number of discrete events.

• Descriptive Statistics

Or, consider the scourge of many students, the Grade Point Average (GPA).

This single number describes the general performance of a student across a potentially wide range of course experiences.

• Descriptive Statistics

Every time you try to describe a large set of observations with a single indicator you run the risk of distorting the original data or losing detail.

The batting average doesn't tell you whether the batter is hitting home runs or singles. It doesn't tell whether she's been in a slump or on a streak.

The GPA doesn't tell you whether the student was in difficult courses or easy ones, or whether they were courses in their major field or in other disciplines.

Even given these limitations, descriptive statistics provide a powerful summary that may enable comparisons across people or other units.

• Descriptive Statistics

Univariate Analysis

Univariate analysis involves the examination across cases of one variable at a time. There are three major characteristics of a single variable that we tend to look at:

•the distribution

•the central tendency

•the dispersion

In most situations, we would describe all three of these characteristics for each of the variables in our study.

• Descriptive Statistics

The Distribution.

The distribution is a summary of the frequency of individual values or ranges of values for a variable.

The simplest distribution would list every value of a variable and the number of persons who had each value.

• Descriptive Statistics

The Distribution.

For instance, a typical way to describe the

distribution of college students is by year

in college, listing the number or percent

of students at each of the four years.

Or, we describe gender by listing the number or percent of males and females.

In these cases, the variable has few enough values that we can list each one and summarize how many sample cases had the value.

• Descriptive Statistics

The Distribution.

But what do we do for a variable like income or GPA?

With these variables there can be a large number of possible values, with relatively few people having each one.

In this case, we group the raw scores into categories according to ranges of values.

For instance, we might look at GPA according to the letter grade ranges. Or, we might group income into four or five ranges of income values.

• Descriptive Statistics

The Distribution.

One of the most common ways to describe a single variable is with a frequency distribution.

Depending on the particular variable, all of the data values may be represented, or you may group the values into categories first (e.g., with age, price, or temperature variables, it would usually not be sensible to determine the frequencies for each value.

Rather, the value are grouped into ranges and the frequencies determined.).

• Descriptive Statistics

The Distribution.

Frequency distributions can be depicted in two ways, as a table or as a graph. The table shows an age frequency distribution with five categories of age ranges defined.

• Descriptive Statistics

The Distribution.

The same frequency distribution can be depicted in a graph as shown in the figure. This type of graph is often referred to as a histogram or bar chart.

• Descriptive Statistics

Central Tendency.

The central tendency of a distribution is an estimate of the "center" of a distribution of values.

There are three major types of estimates of central tendency:

•Mean

•Median

•Mode

• Descriptive Statistics

The Mean or average is probably the most commonly used method of describing central tendency.

To compute the mean all you do is add up all the values and divide by the number of values.

For example, the mean or average quiz score is determined by summing all the scores and dividing by the number of students taking the exam.

• Descriptive Statistics

The Mean or average is probably the most commonly used method of describing central tendency.

For example, consider the test score values:

15, 20, 21, 20, 36, 15, 25, 15

The sum of these 8 values is 167,

so the mean is 167/8 = 20.875.

• Descriptive Statistics

The Median is the score found at the exact middle of the set of values.

One way to compute the median is to list all scores in numerical order, and then locate the score in the center of the sample.

• Descriptive Statistics

The Median

For example, if there are 500 scores in the list, score #250 would be the median. If we order 8 scores, we would get:

15,15,15,20,20,21,25,36

There are 8 scores and score #4 and #5 represent the halfway point. Since both of these scores are 20, the median is 20.

If the two middle scores had different values, you would have to interpolate to determine the median.

• Descriptive Statistics

The mode is the most frequently occurring value in the set of scores.

To determine the mode, you might again order the scores as shown above, and then count each one. The most frequently occurring value is the mode.

In our example, (15, 20, 21, 20, 36, 15, 25, 15) the value 15 occurs three times and is the model.

In some distributions there is more than one modal value. For instance, in a bimodal distribution there are two values that occur most frequently.

• Descriptive Statistics

Notice that for the same set of 8 scores we got three different values

-- 20.875, 20, and 15 –

for the mean, median and mode respectively.

If the distribution is truly normal (i.e., bell-shaped), the mean, median and mode are all equal to each other.

• Descriptive Statistics

Dispersion.

Dispersion refers to the spread of the values around the central tendency.

There are two common measures of dispersion,

the range and the standard deviation.

• Descriptive Statistics

Dispersion.

Dispersion refers to the spread of the values around the central tendency.

There are two common measures of dispersion,

the range and the standard deviation.

The range is simply the highest value minus the lowest value. In our example distribution, the high value is 36 and the low is 15, so the range is 36 - 15 = 21.

• Descriptive Statistics

The Standard Deviation is a more accurate and detailed estimate of dispersion because an outlier can greatly exaggerate the range.

Look at the set of scores: 15,20,21,20,36,15,25,15

the single outlier value of 36 stands apart from the rest of the values.

The Standard Deviation shows the relation that set of scores has to the mean of the sample.

• Descriptive Statistics

To compute the standard deviation, we first find the distance between each value and the mean.

We know from above that the mean is 20.875. So, the differences from the mean are:

15 - 20.875 = -5.875

20 - 20.875 = -0.875

21 - 20.875 = +0.125

20 - 20.875 = -0.875

36 - 20.875 = 15.125

15 - 20.875 = -5.875

25 - 20.875 = +4.125

15 - 20.875 = -5.875

• Descriptive Statistics

Notice that values that are below the mean have negative discrepancies and values above it have positive ones.

Next, we square each discrepancy:

-5.875 * -5.875 = 34.515625

-0.875 * -0.875 = 0.765625

+0.125 * +0.125 = 0.015625

-0.875 * -0.875 = 0.765625

15.125 * 15.125 = 228.765625

-5.875 * -5.875 = 34.515625

+4.125 * +4.125 = 17.015625

-5.875 * -5.875 = 34.515625

• Descriptive Statistics

Now, we take these "squares" and sum them to get the Sum of Squares (SS) value. Here, the sum is 350.875.

Next, we divide this sum by the number of scores minus 1. Here, the result is 350.875 / 7 = 50.125.

This value is known as the variance.

To get the standard deviation, we take the square root of the variance (remember that we squared the deviations earlier).

This would be SQRT(50.125) = 7.079901129253.

• Descriptive Statistics

Although this computation may seem convoluted, it's actually quite simple.

To see this, consider the formula for the standard deviation:

• Descriptive Statistics

In the top part of the ratio, the numerator,

we see that each score has the mean

subtracted from it, the difference is

squared, and the squares are summed.

In the bottom part, we take the number of scores minus 1.

The ratio is the variance and the square root is the standard deviation. We can describe the standard deviation as:

The square root of the sum of the squared deviations from the mean divided by the number of scores minus one

• Descriptive Statistics

Although we can calculate these univariate statistics by hand, it gets quite tedious when you have more than a few values and variables.

Every statistics program and calculator is capable of calculating them easily for you.

• Descriptive Statistics

• The standard deviation allows us to reach some conclusions about specific scores in our distribution.

• Assuming that the distribution of scores is normal or bell-shaped (or close to it!), the following conclusions can be reached:

• approximately 68% of the scores in the sample fall within one standard deviation of the mean

• approximately 95% of the scores in the sample fall within two standard deviations of the mean

• approximately 99% of the scores in the sample fall within three standard deviations of the mean

• Descriptive Statistics

For instance, since the mean in our example is 20.875 and the standard deviation is 7.0799,

we can from the above statement estimate that approximately 95% of the scores will fall in the range of 20.875-(2*7.0799) to 20.875+(2*7.0799)

or between 6.7152 and 35.0348.

This kind of information is a critical stepping stone to enabling us to compare the performance of an individual on one variable with their performance on another, even when the variables are measured on entirely different scales.

• Descriptive Statistics

Worksheet 01

While performing a usability test on a new computer interface the following data was collected :

Create a frequency distribution for each variable.

Calculate the mean, median and mode for each data set.

Calculate the range, variance and standard deviation for each data set.

4. What conclusions can be drawn, if any from these statistics.

### HCI 510: HCI Methods I

Descriptive Statistics

Inferential Statistics

Significance

T-Test

• Inferential Statistics

Inferential Statistics

With inferential statistics, you are trying to reach conclusions that extend beyond the immediate data alone.

For instance, we use inferential statistics to try to infer from the sample data what the population might think.

Or, we use inferential statistics to make judgments of the probability that an observed difference between groups is a dependable one or one that might have happened by chance in this study.

• Inferential Statistics

Inferential Statistics

Thus,

we use inferential statistics to make inferences from our data to more general conditions;

we use descriptive statistics simply to describe what's going on in our data.

• Inferential Statistics

Inferential Statistics

Inferential statistics are useful in experimental and quasi-experimental research design or in program outcome evaluation.

• Inferential Statistics

Inferential Statistics

Perhaps one of the simplest inferential test is used when you want to compare the average performance of two groups on a single measure to see if there is a difference.

You might want to know whether eighth-grade boys and girls differ in math test scores or whether a program group differs on the outcome measure from a control group.

Whenever you wish to compare the average performance between two groups you should consider the t-test for differences between groups.

• Inferential Statistics

Inferential Statistics

Most of the major inferential statistics come from a general family of statistical models known as the General Linear Model.

This includes the t-test, Analysis of Variance (ANOVA), Analysis of Covariance (ANCOVA), regression analysis, and many of the multivariate methods like factor analysis, multidimensional scaling, cluster analysis, discriminant function analysis, and so on.

• Significance

Significance

"Significance level" is a misleading term that many people do not fully understand. In normal English, "significant" means important, while in Statistics "significant" means probably true (not due to chance).

A research finding may be true without being important. When statisticians say a result is "highly significant" they mean it is very probably true.

They do not (necessarily) mean it is highly important.

• Significance

Take a look at the table below.

• Significance

The chi squares at the bottom of the table show two rows of numbers. The top row numbers of 0.07 and 24.4 are the chi square statistics themselves. The second row contains values .795 and .001. These are the significance levels.

### HCI 510: HCI Methods I

Descriptive Statistics

Inferential Statistics

Significance

T-Test

• Significance

Significance levels show you how likely a result is due to chance.

The most common level, used to mean something is good enough to be believed, is .95.

This means that the finding has a 95% chance of being true.

However, this value is also used in a misleading way.

• Significance

No statistical package will show you "95%" or ".95" to indicate this level.

Instead it will show you ".05," meaning that the finding has a five percent (.05) chance of not being true, which is the converse of a 95% chance of being true.

To find the significance level, subtract the number shown from one.

For example, a value of ".01" means that there is a 99% (1-.01=.99) chance of it being true.

• Significance

In this table, there is probably no difference in purchases of gasoline X by people in the city center and the suburbs, because the probability is .795 (i.e., there is only a 20.5% chance that the difference is true).

.

• Significance

In contrast the high significance level for type of vehicle (.001 or 99.9%) indicates there is almost certainly a true difference in purchases of Brand X by owners of different vehicles in the population from which the sample was drawn.

• Significance

In all cases of calculating statistical significance, the p value tells you how likely something is to be not true.

If a chi square test shows probability of .04, it means that there is a 96% (1-.04=.96) chance that the answers given by different groups really are different.

If a t-test reports a probability of .07, it means that there is a 93% chance that the two means being compared would be truly different if you looked at the entire population.

• Significance

Significance is a statistical term that tells how sure you are that a difference or relationship exists.

To say that a significant difference or relationship exists only tells half the story. We might be very sure that a relationship exists, but is it a strong, moderate, or weak relationship?

After finding a significant relationship, it is important to evaluate its strength.

Significant relationships can be strong or weak. Significant differences can be large or small. It just depends on your sample size.

• Significance

For example, suppose we give 1,000 people an IQ test, and we ask if there is a significant difference between male and female scores.

The mean score for males is 98 and the mean score for females is 100.

We use an independent groups t-test and find that the difference is significant at the .001 level.

The big question is, "So what?". The difference between 98 and 100 on an IQ test is a very small difference...so small, in fact, that its not even important.

• Significance

Then why did the t-statistic come out significant?

Because there was a large sample size.

When you have a large sample size, very small differences will be detected as significant.

This means that you are very sure that the difference is real (i.e., it didn't happen by fluke). It doesn't mean that the difference is large or important.

If we had only given the IQ test to 25 people instead of 1,000, the two-point difference between males and females would not have been significant.

• Significance

People sometimes think that the 95% level is sacred.

If a test shows a .06 probability, it means that it has a 94% chance of being true. You can't be quite as sure about it as if it had a 95% chance of being be true, but the odds still are that it is true.

The 95% level comes from academic publications, where a theory usually has to have at least a 95% chance of being true to be considered worth telling people about.

In the ‘real’ world if something has a 90% chance of being true (probability =.1), it can't be considered proven, but it is probably better to act as if it were true rather than false.

• Significance

Many researchers use the word "significant" to describe a finding that may have decision-making utility to a client.

From a statistician's viewpoint, this is an incorrect use of the word.

However, the word "significant" has virtually universal meaning to the public.

Thus, many researchers use the word "significant" to describe a difference or relationship that may be strategically important to a client (regardless of any statistical tests).

• Significance

In these situations, the word "significant" is used to advise a client to take note of a particular difference or relationship because it may be relevant to the company's strategic plan.

The word "significant" is not the exclusive domain of statisticians and either use is correct in the business world.

Thus, for the HCI expert, it may be wise to adopt a policy of always referring to "statistical significance" rather than simply "significance" when communicating with the public.

• Significance

One important concept in significance testing is whether you use a one-tailed or two-tailed test of significance.

• Significance

• When your research hypothesis states the direction of the difference or relationship, then you use a one-tailed probability.

• For example, a one-tailed test would be used to test these null hypotheses:

• Females will not score significantly higher than males on an IQ test.

• Blue collar workers are will not buy significantly more product than white collar workers.

• Superman is not significantly stronger than the average person.

• In each case, the null hypothesis (indirectly) predicts the direction of the difference.

• Significance

• A two-tailed test would be used to test these null hypotheses:

• There will be no significant difference in IQ scores between males and females.

• There will be no significant difference in the amount of product purchased between blue collar and white collar workers.

• There is no significant difference in strength between Superman and the average person.

• The one-tailed probability is exactly half the value of the two-tailed probability.

• Significance

• A two-tailed test would be used to test these null hypotheses:

• There will be no significant difference in IQ scores between males and females.

• There will be no significant difference in the amount of product purchased between blue collar and white collar workers.

• There is no significant difference in strength between Superman and the average person.

• The one-tailed probability is exactly half the value of the two-tailed probability.

• Significance

Whenever we perform a significance test, it involves comparing a test value that we have calculated to some critical value for the statistic.

It doesn't matter what type of statistic we are calculating (e.g., a t-statistic, a chi-square statistic, an F-statistic, etc.), the procedure to test for significance is the same.

1. Decide on the critical alpha level you will use (i.e., the error rate you are willing to accept).

2. Conduct the research.

3. Calculate the statistic.

4. Compare the statistic to a critical value obtained from a table.

• Significance

• If your statistic is higher than the critical value from the table:

• You reject the null hypothesis.

• The probability is small that the difference or relationship happened by chance, and p is less than the critical alpha level (p < alpha ).

• Significance

• If your statistic is lower than the critical value from the table:

• Your finding is not significant.

• You fail to reject the null hypothesis.

• The probability is high that the difference or relationship happened by chance, and p is greater than the critical alpha level (p > alpha ).

### HCI 510: HCI Methods I

Descriptive Statistics

Inferential Statistics

Significance

T-Test

• T-Test

The t-test assesses whether the means of two groups are statistically different from each other.

This analysis is appropriate whenever you want to compare the means of two groups, and especially appropriate as the analysis for the posttest-only two-group randomized experimental design.

• T-Test

The figure shows the distributions for the treated (orange) and control (purple) groups in a study.

• T-Test

The figure indicates where the control and treatment group means are located.

The question the t-test addresses is whether the means are statistically different.

• T-Test

What does it mean to say that the averages for two groups are statistically different?

• T-Test

What does it mean to say that the averages for two groups are statistically different?

• T-Test

The first thing to notice about the three situations is that the difference between the means is the same in all three.

But, you should also notice that the three situations don't look the same -- they tell very different stories.

• T-Test

The top example shows a case with moderate variability of scores within each group. The second situation shows the high variability case. the third shows the case with low variability.

• T-Test

Clearly, we would conclude that the two groups appear most different or distinct in the bottom or low-variability case.

Why? Because there is relatively little overlap between the two bell-shaped curves.

In the high variability case, the group difference appears least striking because the two bell-shaped distributions overlap so much.

• T-Test

This leads us to a very important conclusion: when we are looking at the differences between scores for two groups, we have to judge the difference between their means relative to the spread or variability of their scores.

The t-test does just this..

• T-Test

Statistical Analysis of the t-test

The formula for the t-test is a ratio.

The top part of the ratio is just the difference between the two means or averages.

The bottom part is a measure of the variability or dispersion of the scores.

• T-Test

Statistical Analysis of the t-test

This formula is essentially another example of the signal-to-noise metaphor in research:

the difference between the means is the signal that, in this case, we think our program or treatment introduced into the data;

the bottom part of the formula is a measure of variability that is essentially noise that may make it harder to see the group difference.

• T-Test

The figure below shows the formula for the t-test and how the numerator and denominator are related to the distributions.

• T-Test

The top part of the formula is easy to compute -- just find the difference between the means.

The bottom part is called the standard error of the difference.

To compute it, we take the variance for each group and divide it by the number of people in that group.

We add these two values and then take their square root.

• T-Test

The final formula for the t-test is shown below:

Remember, that the variance is simply the square of the standard deviation.

• T-Test

The t-value will be positive if the first mean is larger than the second and negative if it is smaller.

Once you compute the t-value you have to look it up in a table of significance to test whether the ratio is large enough to say that the difference between the groups is not likely to have been a chance finding.

To test the significance,

you need to set a risk level

(called the alpha level).

• T-Test

You also need to determine the degrees of freedom (df) for the test.

In the t-test, the degrees of freedom is the sum of the persons in both groups minus 2.

• T-Test

Given the alpha level, the df, and the t-value, you can look the t-value up in a standard table of significance to determine whether the t-value is large enough to be significant.

If it is, you can conclude that the difference between the means for the two groups is different (even given the variability).

• T-Test

Worksheet 2

Rosenthal and Jacobson (1968) informed classroom teachers that some of their students showed unusual potential for intellectual gains.

Eight months later the students identified to teachers as having potentional for unusual intellectual gains showed significantly greater gains performance on a test said to measure IQ than did children who were not so identified.

• T-Test

Conclusions

If you do a large number of tests, falsely significant results are a problem.

Remember that a 95% chance of something being true means there is a 5% chance of it being false. This means that of every 100 tests that show results significant at the 95% level, the odds are that five of them do so falsely.

If you took a random, meaningless set of data and did 100 significance tests, the odds are that five tests would be falsely reported significant.

As you can see, the more tests you do, the more of a problem these false positives are. You cannot tell which the false results are - you just know they are there.

• T-Test

Conclusions

Limiting the number of tests to a small group chosen before the data is collected is one way to reduce the problem.

If this isn't practical, there are other ways of solving this problem.

The best approach from a statistical point of view is to repeat the study and see if you get the same results.

If something is statistically significant in two separate studies, it is probably true.

• T-Test

Conclusions

In real life it is not usually practical to repeat a survey, but you can use the "split halves" technique of dividing your sample randomly into two halves and do the tests on each.

If something is significant in both halves, it is probably true.

The main problem with this technique is that when you halve the sample size, a difference has to be larger to be statistically significant.

• T-Test

Conclusions

The last common error is also important.

Most significance tests assume you have a truly random sample.

If your sample is not truly random, a significance test may overstate the accuracy of the results, because it only considers random error.

The test cannot consider biases resulting from non-random error (for example a badly selected sample).

• T-Test

• Conclusions

• To summarize:

• In statistical terms, significant does not necessarily mean important.

• Probability values should be read in reverse (1 - p).

• Too many significance tests will turn up some falsely significant relationships.

• Check your sampling procedure to avoid bias.