1 / 36

# PPA 501 – Analytical Methods in Administration - PowerPoint PPT Presentation

PPA 501 – Analytical Methods in Administration. Lecture 9 – Bivariate Association. Statistical Significance and Theoretical Significance. Tests of significance detect nonrandom relationships.

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.

## PowerPoint Slideshow about ' PPA 501 – Analytical Methods in Administration' - honorato-cunningham

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

### PPA 501 – Analytical Methods in Administration

Lecture 9 – Bivariate Association

• Tests of significance detect nonrandom relationships.

• Measures of association go one step farther and provide information on the strength and direction of relationships.

• Measures of association allow us to achieve two goals:

• Trace causal relationships among variables.

• But, they cannot prove causality.

• Predict scores on the dependent variable based on information from the independent variable.

• Two variables are said to be associated if the distribution of one of them changes under the various categories or scores of the other.

• Liberals are more likely to vote for Democratic candidates than for Republican Candidates.

• Presidents are more likely to grant disaster assistance when deaths are involved than when there are no deaths.

• Bivariate tables are devices for displaying the scores of cases on two different variables.

• Independent or X variable in the columns.

• Dependent variable or Y variable in the rows.

• Each column represents a category on the independent variable. Each row represents a category on the dependent variable. Each cell represents those cases that fall into each combination of categories.

Each column’s frequency distribution is called a conditional distribu-

tion of Y.

• Often you will calculate a chi-square when you generate table. If chi-square is zero, then there is no association.

• Usually, however, chi-square is positive to some degree.

• Statistical significance is not the same as association. It is, however, usually the case that significance is the first step in determining the strength of association.

• Does an association exist?

• Statistical significance the first step.

• Calculate column percentages and compare across conditional distributions.

• If there is an association, the largest cell will change from column to column.

• If there is no association, the conditional percentages will not change.

• How strong is the association?

• Once we establish the existence of an association, we need to determine how strong it is?

• A matter of measuring the changes across conditional distributions.

• No association – no change in column percentages.

• Perfect association – each value of the independent variable is associated with one and only one value of the dependent variable.

• The huge majority of relationships fall in between.

• How strong is the association (contd.)?

• Virtually all statistics of association are designed to vary between 0 for no association and +1 for perfect association (±1 for ordinal and interval data).

• The meaning of the statistics varies a little from statistic to statistics, but 0 signifies no association and 1 signifies perfect association in all cases.

• What is the pattern and/or direction of the association?

• Pattern is determined by examining which categories of X are associated with which categories of Y.

• Direction only matters for ordinal and interval-ratio data. In positive association, low values on one variable are associated with low values on the other and high with high.

• On negative association, low values on variable are associated with high values on the other and vice versa.

The five-step model is calculated using chi-square.

• The logic of proportional reduction in error (PRE) involves first attempting to guess or predict the category into which each case will fall on the dependent variable without using information from the independent variable.

• The second step involves using the information on the conditional distribution of the dependent variable within categories of the independent variable to reduce errors in prediction.

• If there is a strong association, there will be a substantial reduction in error from knowing the joint distribution of X and Y.

• If there is no association, there will be no reduction in error.

Prediction rule: Predict that all cases fall the largest category.

Where E1 is the number of prediction errors without knowing X.

And E2 is the number of prediction errors knowing the distribution

On X.

• E1 is calculated by subtracting the cases in the largest category of the row marginals from the total number of cases.

• E2 is calculated by subtracting the largest category in each column from the column total and summing across columns of the independent variable.

• The key problem with lambda occurs when the distribution on one of the variables is lopsided (many cases in one category and few cases in the others). Under those circumstances, lambda can equal zero, even when there is a relationship.

• The five-step test of significance for both the chi-square-based statistics and the PRE statistics is chi-square.

• Some ordinal variables have many categories and look like interval variables. These can be called continuous ordinal variables. The appropriate ordinal association statistic is Spearman’s rho (ignored in this lecture)

• Some ordinal variables have only a few categories and can be called collapsed ordinal variables. The appropriate ordinal association statistic is gamma and the other tau-class statistics.

• These measures of association compare each case to every other case. The total number of pairs of cases is equal to N(N-1)/2.

• There are five classes of pairs:

• C or Ns: Pairs where one case is higher than the other on both variables.

• D or Nd: Pairs where one case is higher on one variable and lower on the other.

• Ty: Pairs tied on the dependent variable but not the independent variable.

• Tx: Pairs tied on the independent variable but not the dependent variable.

• Txy:Pairs tied on both variables.

• To calculate C, start in the upper left cell and multiply the frequency in each cell by the total of all frequencies to right and below the cell in the table.

• To calculate D, start in the upper right cell and multiply the frequency in each cell by the total of all frequencies to the left and below the cell in the table.

• To calculate Tx, start in the upper left cell and multiply the frequency in each cell by the total of all frequencies directly below the cell.

• To calculate Ty, start in the upper left cell and multiple the frequency in each cell by the total of all frequencies directly to the right of the cell.

• To calculate Txy, start in the upper left cell and multiply N(N-1)/2 for each cell.

Gamma and its associated statistics can have a PRE interpretation.

• Step 1. Making assumptions.

• Random sampling.

• Ordinal measurement.

• Normal sampling distribution.

• Step 2 – Stating the null hypothesis.

• H0: γ=0.0

• H1: γ 0.0

• Step 3. Selecting the sampling distribution and establishing the critical region.

• Sampling distribution = Z distribution.

• Alpha = .05.

• Z (critical) = ±1.96.

• Step 4. Computing the test statistic.

• Step 5. Making a decision.

• Z(obtained) is less than Z(critical). Fail to reject the null hypothesis that gamma is zero in the population. There is no relationship between race and the feeling of safety at home based on the JCHA 2000 sample.