Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

One-way ANOVA: Comparing Group Means with Homogeneity of Variance Assumption, Study notes of Design

Analysis of VarianceMultiple Comparison ProceduresHypothesis TestingStatistical Analysis

The One-way ANOVA statistical method used to compare the means of multiple groups while maintaining the Type I error rate at a pre-established level. It covers the problems with using multiple t-tests, the assumptions of normality, homogeneity of variance, and independence, and testing these assumptions using Bartlett's test and Levene's test. The document also discusses the Kruskal-Wallis test as an alternative when assumptions are not met.

What you will learn

  • What are the problems with using multiple t-tests?
  • What is the Levene's test used for in the One-way ANOVA?
  • When should you use the Kruskal-Wallis test instead of the One-way ANOVA?
  • What is the One-way ANOVA statistical method used for?
  • How do you test the assumption of homogeneity of variance using Bartlett's test?

Typology: Study notes

2021/2022

Uploaded on 08/05/2022

char_s67
char_s67 🇱🇺

4.5

(109)

1.9K documents

1 / 13

Toggle sidebar

Related documents


Partial preview of the text

Download One-way ANOVA: Comparing Group Means with Homogeneity of Variance Assumption and more Study notes Design in PDF only on Docsity! UNDERSTANDING THE ONE-WAY ANOVA The One-way Analysis of Variance (ANOVA) is a procedure for testing the hypothesis that K population means are equal, where K > 2. The One-way ANOVA compares the means of the samples or groups in order to make inferences about the population means. The One-way ANOVA is also called a single factor analysis of variance because there is only one independent variable or factor. The independent variable has nominal levels or a few ordered levels. PROBLEMS WITH USING MULTIPLE t TESTS To test whether pairs of sample means differ by more than would be expected due to chance, we might initially conduct a series of t tests on the K sample means – however, this approach has a major problem (i.e., inflating the Type I Error rate). The number of t tests needed to compare all possible pairs of means would be K(K – 1)/2, where K = number of means. When more than one t test is run, each at a specified level of significance (such as α = .05), the probability of making one or more Type I errors in the series of t tests is greater than α. The Type I Error Rate is determined as c)1(1 α−− Where α = level of significance for each separate t test c = number of independent t tests Sir Ronald A. Fisher developed the procedure known as analysis of variance (ANOVA), which allows us to test the hypothesis of equality of K population means while maintaining the Type I error rate at the pre-established (a priori) α level for the entire set of comparisons. THE VARIABLES IN THE ONE-WAY ANOVA In an ANOVA, there are two kinds of variables: independent and dependent. The independent variable is controlled or manipulated by the researcher. It is a categorical (discrete) variable used to form the groupings of observations. However, do not confuse the independent variable with the “levels of an independent variable.” In the One-way ANOVA, only one independent variable is considered, but there are two or more (theoretically any finite number) levels of the independent variable. The independent variable is typically a categorical variable. The independent variable (or factor) divides individuals into two or more groups or levels. The procedure is a One-way ANOVA, since there is only one independent variable. There are two types of independent variables: active and attribute. If the independent variable is an active variable then we manipulate the values of the variable to study its affect on another variable. For example, anxiety level is an active independent variable. An attribute independent variable is a variable where we do not alter the variable during the study. For example, we might want to study the effect of age on weight. We cannot change a person’s age, but we can study people of different ages and weights. THE ONE-WAY ANOVA PAGE 2 The (continuous) dependent variable is defined as the variable that is, or is presumed to be, the result of manipulating the independent variable. In the One-way ANOVA, there is only one dependent variable – and hypotheses are formulated about the means of the groups on that dependent variable. The dependent variable differentiates individuals on some quantitative (continuous) dimension. The ANOVA F test (named after Sir Ronald A. Fisher) evaluates whether the group means on the dependent variable differ significantly from each other. That is, an overall analysis-of-variance test is conducted to assess whether means on a dependent variable are significantly different among the groups. MODELS IN THE ONE-WAY ANOVA In an ANOVA, there are two specific types of models that describe how we choose the levels of our independent variable. We can obtain the levels of the treatment (independent) variable in at least two different ways: We could, and most often do, deliberately select them or we could sample them at random. The way in which the levels are derived has important implications for the generalization we might draw from our study. For a one-way analysis of variance, the distinction is not particularly critical, but it can become quite important when working with more complex designs such as the factorial analysis of variance. If the levels of an independent variable (factor) were selected by the researcher because they were of particular interest and/or were all possible levels, it is a fixed-model (fixed-factor or effect). In other words, the levels did not constitute random samples from some larger population of levels. The treatment levels are deliberately selected and will remain constant from one replication to another. Generalization of such a model can be made only to the levels tested. Although most designs used in behavioral science research are fixed, there is another model we can use in single-factor designs. If the levels of an independent variable (factor) are randomly selected from a larger population of levels, that variable is called a random- model (random-factor or effect). The treatment levels are randomly selected and if we replicated the study we would again choose the levels randomly and would most likely have a whole new set of levels. Results can be generalized to the population levels from which the levels of the independent variable were randomly selected. HYPOTHESES FOR THE ONE-WAY ANOVA The null hypothesis (H0) tested in the One-way ANOVA is that the population means from which the K samples are selected are equal. Or that each of the group means is equal. H0: Kµµµ === ...21 Where K is the number of levels of the independent variable. For example: If the independent variable has three levels – we would write… H0: 321 µµµ == If the independent variable has five levels – we would write… H0: 54321 µµµµµ ==== THE ONE-WAY ANOVA PAGE 5 histogram graph (non-normal shape is a concern), an examination of box plots (extreme outliers is a concern), and an examination of the Normal Q-Q Plots (a non-linear relationship is a concern). These procedures are typically done during data screening. In examining skewness and kurtosis, we divide the skewness (kurtosis) statistic by its standard error. We want to know if this standard score value significantly departs from normality. Concern arises when the skewness (kurtosis) statistic divided by its standard error is greater than z +3.29 (p < .001, two-tailed test) (Tabachnick & Fidell, 2007). We have several options for handling non-normal data, such as deletion and data transformation (based on the type and degree of violation as well as the randomness of the missing data points). Any adjustment to the data should be justified (i.e., referenced) based on solid resources (e.g., prior research or statistical references). As a first step, data should be thoroughly screened to ensure that any issues are not a factor of missing data or data entry errors. Such errors should be resolved prior to any data analyses using acceptable procedures (see for example Howell, 2007 or Tabachnick & Fidell, 2007). TESTING THE ASSUMPTION OF HOMOGENEITY-OF-VARIANCE Another one of the first steps in the One-way ANOVA test is to test the assumption of homogeneity of variance, where the null hypothesis assumes no difference between the K group’s variances. One method is the Bartlett’s test for homogeneity of variance (this test is very sensitive to non-normality). The Levene’s F Test for Equality of Variances, which is the most commonly used statistic (and is provided in SPSS), is used to test the assumption of homogeneity of variance. Levene’s test uses the level of significance set a priori for the ANOVA (e.g., α = .05) to test the assumption of homogeneity of variance. Test of Homogeneity of Variances SCORE 1.457 2 42 .244 Levene Statistic df1 df2 Sig. For Example: For the SCORE variable (shown above), the F value for Levene’s test is 1.457 with a Sig. (p) value of .244. Because the Sig. value is greater than our alpha of .05 (p > .05), we retain the null hypothesis (no difference) for the assumption of homogeneity of variance and conclude that there is not a significant difference between the three group’s variances. That is, the assumption of homogeneity of variance is met. Test of Homogeneity of Variances VISUAL 17.570 1 498 .000 Levene Statistic df1 df2 Sig. THE ONE-WAY ANOVA PAGE 6 For Example: For the VISUAL variable (shown above), the F value for Levene’s test is 17.570 with a Sig. (p) value of .000 (< .001). Because the Sig. value is less than our alpha of .05 (p < .05), we reject the null hypothesis (no difference) for the assumption of homogeneity of variance and conclude that there is a significant difference between the two group’s variances. That is, the assumption of homogeneity of variance is not met. VIOLATION OF THE ASSUMPTIONS OF THE ONE-WAY ANALYSIS OF VARIANCE If a statistical procedure is little affected by violating an assumption, the procedure is said to be robust with respect to that assumption. The One-way ANOVA is robust with respect to violations of the assumptions, except in the case of unequal variances with unequal sample sizes. That is, the ANOVA can be used when variances are only approximately equal if the number of subjects in each group is equal (where equal can be defined as the larger group size not being more then 1½ times the size of the smaller group). ANOVA is also robust if the dependent variable data are even approximately normally distributed. Thus, if the assumption of homogeneity of variance (where the larger group variance is not more than 4 or 5 times that of the smaller group variance), or even more so, the assumption of normality is not fully met, you may still use the One-way ANOVA. Generally, failure to meet these assumptions changes the Type I error rate. Instead of operating at the designated level of significance, the actual Type I error rate may be greater or less than α, depending on which assumptions are violated. When the population sampled are not normal, the effect on the Type I error rate is minimal. If the population variances differ, there may be a serious problem when sample sizes are unequal. If the larger variance is associated with the larger sample, the F test will be too conservative. If the smaller variance is associated with the larger sample, the F test will be too liberal. (If the α level is .05, “conservative” means that the actual rate is less than .05.) If the sample sizes are equal, the effect of heterogeneity of variances (i.e., violating the assumption of homogeneity of variance) on the Type I error is minimal. In other words, the effects of violating the assumptions vary somewhat with the specific assumptions violated. If there are extreme violations of these assumptions – with respect to normality and homogeneity of variance – an alternate test such as the Kruskal-Wallis test should be used instead of the one-way analysis of variance test. The Kruskal-Wallis test is a nonparametric test that is used with an independent groups design employing K groups. It is used as a substitute for the parametric one-way ANOVA, when the assumptions of that test are seriously violated. The Kruskal-Wallis test does not assume population normality nor homogeneity of variance, as does the parametric ANOVA, and only requires ordinal scaling of the dependent variable. It is used when violations of population normality and/or homogeneity of variance are extreme or when interval or ratio scaling are required and not met by the data. THE ONE-WAY ANOVA PAGE 7 THE WELCH AND BROWN-FORSYTHE STATISTIC If the equal variance assumption has been violated (e.g., if the significance for the Levene’s test is less than .05), we can use an adjusted F statistic. Two such types of adjustments are provided by the Welch statistic and the Brown-Forsythe statistic. The Welch test is more powerful and more conservative than the Brown-Forsythe test. If the F ratio is found to be significant with either the Welch statistic or the Brown- Forsythe statistic, an appropriate post hoc test would be required. The Games-Howell post hoc test, for example, is appropriate when the equal variances assumption has been violated. This test is not appropriate if equal variance were assumed. Robust Tests of Equality of Means Number of words recalled 9.037 4 21.814 .000 6.095 4 18.706 .003 Welch Brown-Forsythe Statistica df1 df2 Sig. Asymptotically F distributed.a. The output from the above table is only valid if the equal variance assumption has been violated. From this example, using the Welch statistic, we find that F(4, 21.814) = 9.037, p < .001. If for example, our a priori alpha level were set at .05, we would conclude that the adjusted F ratio is significant. Since the p value is smaller than α we would reject the null hypothesis and we would have permission to proceed and compare the group means. The difference between the adjusted F ratio (devised by Welch and Brown and Forsythe) and the ordinary F ratio is quite similar to that of the adjusted t and ordinary t found in the independent-samples t test. In both cases it is only the denominator (i.e., error term) of the formula that changes. SUMMARY TABLE FOR THE ONE-WAY ANOVA Summary ANOVA Source Sum of Squares Degrees of Freedom Variance Estimate (Mean Square) F Ratio Between SSB K – 1 MSB = 1−K SSB W B MS MS Within SSW N – K MSW = KN SSW − Total SST = SSB + SSW N – 1 THE ONE-WAY ANOVA PAGE 10 EFFECT SIZE Effect size, broadly, is any of several measures of association or of the strength of a relation, such as Pearson’s r or eta (). Effect size is thought of as a measure of practical significance. It is defined as the degree to which a phenomenon exists. Keep in mind that there are several acceptable measures of effect size. The choice should be made based on solid references based on the specific analysis being conducted. So why bother with effect size at all? Any observed difference between two sample means can be found to be statistically significant when the sample sizes are sufficiently large. In such a case, a small difference with little practical importance can be statistically significant. On the other hand, a large difference with apparent practical importance can be nonsignificant when the sample sizes are small. Effect sizes provide another measure of the magnitude of the difference expressed in standard deviation units in the original measurement. Thus, with the test of statistical significance (e.g., the F statistic) and the interpretation of the effect size (ES), the researcher can address issues of both statistical significance and practical importance. When we find significant pairwise differences – we will need to calculate an effect size for each of the significant pairs, which will need to be calculated by hand. An examination of the group means will tell us which group performed significantly higher than the other did. For example, using the following formula: W ji MS XX ES −= Note that ji XX − (which can also be written as ki XX − ) is the mean difference of the two groups (pairs) under consideration. This value can be calculated by hand or found in the Mean Difference (I-J) column on the Multiple Comparison table in SPSS. MSW is the Within Group’s Mean Square value (a.k.a. Mean Square Within or ERROR), which is found on the ANOVA Summary Table. Suppose that the mean for the Red Group = 16.60 and the mean for the Green Group = 11.10, and the Mean Square Within (MSW) found in the ANOVA table = 16.136, we would find that the ES = 1.37. That is, the mean difference of 5.50 is 1.37 standard deviation units away from the hypothesized mean difference of 0. Recall that H0: µ1 - µ2 = 0. For Red / Green, we find ES = 3691933.1 016964.4 50.5 136.16 10.1160.16 ==− = 1.37 THE ONE-WAY ANOVA PAGE 11 TESTING THE NULL HYPOTHESIS FOR THE ONE-WAY ANOVA In testing the null hypothesis for the One-way ANOVA – we follow these steps: 1. STATE THE HYPOTHESES • State the Null Hypothesis and Alternative Hypothesis 2. SET THE CRITERION FOR REJECTING H0 • Set the alpha level, which in turn identifies the critical values 3. TEST THE ASSUMPTIONS FOR THE ONE-WAY ANOVA • Assumption of Independence (e.g., Research Design) • Assumption of Normality (e.g., Shaprio-Wilks) • Assumption of Homogeneity of Variance (e.g., Levene’s Test of Homogeneity) 4. COMPUTE THE TEST STATISTIC • Calculate the F ratio (or adjusted F ratio using for example the Welch statistic) 5. DECIDE WHETHER TO RETAIN OR REJECT H0 • Using the obtained probability value (compared to the alpha level) • If p > α  Retain the null hypothesis • If p < α  Reject the null hypothesis • Using the obtained statistic value (compared to the critical value) • If Fobt < Fcrit  Retain the null hypothesis • If Fobt > Fcrit  Reject the null hypothesis 6. CALCULATE MEASURE OF ASSOCIATION (ω2, OMEGA SQUARED) • If the null hypothesis is rejected, calculate omega squared to determine strength of the association between the independent variable and the dependent variable. 7. DECIDE WHETHER TO RETAIN OR REJECT H0 FOR EACH OF THE PAIRWISE COMPARISONS (I.E., CONDUCT POST HOC PROCEDURES) • If the null hypothesis is rejected, use the appropriate post hoc procedure to determine whether unique pairwise comparisons are significant. • Choice of post hoc procedures is based on whether the assumption of homogeneity of variance was met (e.g., Tukey HSD) or not (e.g, Games-Howell). 8. CALCULATE EFFECT SIZE • Calculate an effect size for each significant pairwise comparison 9. INTERPRET THE RESULTS 10. WRITE A RESULTS SECTION BASED ON THE FINDINGS THE ONE-WAY ANOVA PAGE 12 SAMPLE APA RESULTS A One-way Analysis of Variance (ANOVA) was used to examine whether students’ scores on a standardized test is a function of the teaching method they received. The independent variable represented the three different types of teaching methods: 1) lecture only; 2) hands-on only; and 3) lecture and hands-on. The dependent variable was the students’ score on a standardized test. See Table 1 for the means and standard deviations for each of the three groups. Table 1 Means and Standard Deviations of Standardized Test Scores Method n Mean SD Lecture Only 15 82.80 9.59 Hands-On Only 15 88.53 8.73 Lecture and Hands-On 15 92.67 6.22 Total Group 45 88.00 9.09 An alpha level of .05 was used for all analyses. The test for homogeneity of variance was not significant [Levene F(2, 42) = 1.46, p > .05] indicating that this assumption underlying the application of ANOVA was met. The one-way ANOVA of standardized test score (see Table 2) revealed a statistically significant main effect [F(2, 42) = 5.34, p < .01] indicating that not all three groups of the teaching methods resulted in the same standardized test score. The ω2 = .162 indicated that approximately 16% of the variation in standardized test score is attributable to differences between the three groups of teaching methods. Table 2 Analysis of Variance for Standardized Test Scores Source SS df MS F p Between 736.53 2 368.27 5.34 .009 Within 2895.47 42 68.94 Total 3632.00 44
Docsity logo



Copyright © 2024 Ladybird Srl - Via Leonardo da Vinci 16, 10126, Torino, Italy - VAT 10816460017 - All rights reserved