Factorial designs - summary of chapter 14 of statistics by A. Field (5th edition)

Statistics
Chapter 14
Factorial designs

Factorial designs

Factorial design: when an experiment has two or more independent variables.
There are several types of factorial designs:

  • Independent factorial design: there are several independent variables or predictors and each has been measured using different entities (between groups).
  • Repeated-measures (related) factorial design: several independent variables or predictors have been measured, but the same entities have been used in all conditions.
  • Mixed design: several independent variables or predictors have been measured: some have been measured with different entities, whereas others used the same entities.

We can still fit a linear model to the design.
Factorial ANOVA: the linear model with two or more categorical predictors that represent experimental independent variables.

Independent factorial designs and the linear model

The general linear model takes the following general form:

Yi =b0 + b1X1i+b2X2i+... +bnXnii

We can code participant’s category membership on variables with zeros and ones.

For example:

Attractivenessi = b0+b1Ai+b2Bi+b3ABii

b3AB is the interaction variable. It is A dummy multiplied by B dummy variable.

Behind the scenes of factorial designs

Calculating the F-statistic with two categorical predictors is very similar to when we had only one.

  • We still find the total sum of squared errors (SST) and break this variance down into variance that can be explained by the model/experiment (SSM) and variance that cannot be explained (SSR)
  • The main difference is that with factorial designs, the variance explained by the model/experiment is made up of not one predictor, but two.

Therefore, the sum of squares gets further subdivided into

  • variance explained by the first predictor/independent variable (SSA)
  • variance explained by the second predictor/independent variable (SSB)
  • variance explained by the interaction of these two predictors (SSAxB)

Total sum of squares (SST)

We start of with calculating how much variability there is between scores when the ignore the experimental condition from which they came.

The grand variance: the variance of all scores when we ignore the group to which they belong.
We treat the data as one big group.
The degrees of freedom are: N-1

SST = s2Grand(N-1)

The model sum of squares (SSM)

The model sum of squares is broken down into the variance attributable to the first independent variable, the variance attributable to the second independent variable, and the variance attributable to the interaction of those two.

The model sum of squares: the difference between what the model predicts and the overall mean of the outcome variable.
What the model predicts is the group mean.
We work out the model sum of squares by looking at the difference between each group mean and the overall mean.

SSM = Σkg=1ng(ẍg-ẍgrand)2

n = the number of scores in each group

Grand mean = the mean of all scores

The degrees of freedom are k-1

The main effect of A or B

To work out the variance accounted for by the first predictor/independent variable (for example, type of cat) we group Y ratings according to which type of cat was being rated.
So, we ignore the other independent variable, and place all the ratings of (for example) cat type 1 in one group, and all the ratings of (for example) cat type 2 into another.

We apply the same equation for the model sum of squares as above.

The interaction effect, SSAxB

The final stage is to calculate how much variance is explained by the interaction of the two variables.

The SSM is made up of three components (SSA, SSB, and SSAxB)
Given that we know SSA and SSB, we can calculate the interaction term by subtraction

SSAxB = SSM – SSA – SSB

The degrees of freedom can be calculated in the same way, but are also the product of the degrees of freedom for the main effects. Two methods:

  • dfAxB = dfM – dfA-dfB
  • dfAxB = dfA x dfB

The residual sum of squares, SSR

The residual sum of squares represents errors in prediction from the model. In experimental designs, it also reflects individual differences in performance or variance that can’t be explained by factors that were systematically manipulated.

The value is calculated by taking the squared error between each data point and its corresponding group mean.

SSR = Σkg=2s2g(ng-1)

s2group1 (n1-1)+s2group2 (n2-1)+….+ s2groupn (nn-1)+

Degrees of freedom for each group will be one less than the number of scores per group.
We add the degrees of freedom for each group to get a total.

The F-statistics

Each effect in a factorial design has its own F-statistc.
In a two-way design this means we compute F for the two main effects and the interaction.
To calculate these, we first calculate the mean squares of each effect by taking the sum of squares and dividing by the respective degrees of freedom.

MSA = SSA/dfA

MSB = SSB/dfB

MSAxB = SSAxB/dfAxB

MSR = SSR/dfR

The F-statistic for each effect is calculated by dividing its mean squares by the residual mean squares.

FA = MSA/MSR

FB = MSB/MSR

FAxB = MSAxB/MSR

Model assumptions in factorial design

When using the linear model to analyse a factorial design the sources of potential bias discussed in chapter 6 apply.

  • and homogeneity of variance

Output from factorial design

The main effect of A

“Tests of between-subjects effects” tells us whether any of the independent variables had a significant effect on the outcome.

The interaction effect

“Tests of between-subjects effects” also tells us about the interaction effect.

You should not interpret a main effect in the presence of a significant interaction involving that main effect.

Contrasts

Look at ‘contrast results (K Matrix)
The top of the table shows the contrast for Level 1 vs Later. The control vs experimental.
And then Level 2 vs Level 3.

Simple effects analysis

A particular effective way to break down interactions is simple effects analysis.
Looks at the effect of one independent variable at individual levels of the other independent variable .

Post hoc analysis

The Bonferroni post hoc test break down the main effect A and can be interpreted as if A was the only predictor in the model.

Factorial ANOVA

  • Two-way independent designs compare several means when there are two independent variables and different entities have been used in all experimental conditions.
  • In the table called ‘tests of between-subject effects’, look at the column called sig for all main effects and interactions.
  • To interpret a significant interaction, plot an interaction graph and conduct simple effects analysis.
  • You don’t need to interpret main effects if an interaction effect involving that variable is significant.
  • If significant main effects are not qualified by an interaction then consult post hoc tests to see which groups differ. Significance is shown by values smaller than 0.05 in the columns labelled sig. And bootstrap confidence intervals that do not contain zero.
  • Test the same assumptions as for any linear model.

Interpreting interaction graphs

  • Non-parallel lines on an interaction graph indicate some degree of interaction, but how strong an whether the interaction is significant depends on how non-parallel the lines are
  • Lines on an interaction graph that cross are very non-parallel, which hints at a possible significant interaction. But, crossing lines don’t always reflect a significant interaction.

Calculating effect sizes

r = square root ((F(1, dfR))/(F(1,dfR) +dfR))

Reporting results of factorial design

We report the details of the F-statistic and the degrees of freedom for each effect. Also p and effect size.

Like: F(1, 8) = 6,06, p = 0.04, r = 0.4.

Image

Access: 
Public

Image

Join WorldSupporter!

Join with a free account for more service, or become a member for full access to exclusives and extra support of WorldSupporter >>

Check: concept of JoHo WorldSupporter

Concept of JoHo WorldSupporter

JoHo WorldSupporter mission and vision:

  • JoHo wants to enable people and organizations to develop and work better together, and thereby contribute to a tolerant tolerant and sustainable world. Through physical and online platforms, it support personal development and promote international cooperation is encouraged.

JoHo concept:

  • As a JoHo donor, member or insured, you provide support to the JoHo objectives. JoHo then supports you with tools, coaching and benefits in the areas of personal development and international activities.
  • JoHo's core services include: study support, competence development, coaching and insurance mediation when departure abroad.

Join JoHo WorldSupporter!

for a modest and sustainable investment in yourself, and a valued contribution to what JoHo stands for

Check more: this content is used in

Summary of Discovering statistics using IBM SPSS statistics by Field - 5th edition

Image

 

 

Contributions: posts

Help other WorldSupporters with additions, improvements and tips

Add new contribution

CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Image CAPTCHA
Enter the characters shown in the image.

Image

Check more: related and most recent topics and summaries
Check more: institutions, jobs and organizations
Check more: countries and regions
Check more: WorldSupporter and development goals
Check more: this content is also used in

Image

Follow the author: SanneA
Share this page!
Statistics
5150
Submenu & Search

Search only via club, country, goal, study, topic or sector