ECOL 425

    Cards (71)

    • There are 2 types of studies:
      • Experimental
      • the researcher assigns treatments
      • can introduce artefacts
      • bias in measurements produced by unintended consequences of procedures
      • Observational
      • no influence over treatments
      • used for detecting large scale patterns
    • Experiments examine the casual relationship between a predictor (x) and response (y) variable
    • Strength is the effect of predictor (x) when isolated from the effects of confounding variables
    • Good experiments are designed "a priori" and:
      • reduce bias
      • reduce sampling error
      • difference between sample and population result
    • Bias is reduced through:
      • control groups
      • similar conditions as sample, no treatment
      • randomization
      • of individuals receiving treatment
      • cannot occur in observational studies
      • blinding
      • concealing information about treatment assigned
      • Single-blind = subjects unaware
      • Double-blind = researcher + subjects unaware
    • Sampling error is reduced by:
      • replication
      • necessary due to unique individuals
      • increased sample sizes decrease error and provide more information
      • to decide sample size:
      • predetermine level of precision OR power
      • pseudoreplication = measurements are not independent, but are recorded as so
      • balance
      • equal sample sizes in treatments
      • blocking
      • reduces variance by dividing individuals into groups and randomizing within a block
      • Ensures each group is representative of population
    • Confounding variables introduce bias.
      Can be minimised by:
      • pairing individuals with a control of similar characteristics
      • adjustment - categorising data based on the confounding variable and analysing the relationship between x and y
    • ANOVA compares group means by comparing variance between groups.
      • Individual treatment means are fitted and their distance from their treatment mean is analysed
      • Variation within groups is compared with variation among groups
      • If residuals (SSE) < individual treatment mean (SSA) = means are different
    • ANOVA is used when explanatory variables (x) are categorical
    • SSA measures the among group variation and has a df of k-1
    • SSE measures the within group variation has a df of N-k
    • The total variation in ANOVA has a df of N-1
    • Assumption of ANOVA:
      • Random sampling
      • Equal variance
      • Independence of errors
      • Normal distribution of errors
    • Factorial ANOVA tests the effects of 2+ factors and their interaction on a response (y) variable.
      • Reduces Type I error and accounts for variation from crossing variables
    • Factorial ANOVA compares variance of each effect to error variance using the mean square
    • Factorial ANOVA table:
      A) SSA
      B) SSB
      C) SSAB
      D) SSE
      E) ab - 1
      F) a - 1
      G) b - 1
      H) (a - 1)(b - 1)
      I) ab(n - 1)
      J) N - 1
      K) SSA / df
      L) MSA / MSE
      M) MSB / MSE
      N) MSAB / MSE
    • An interaction means that the effect of one factor on a response variable (y) is not constant and depends on the other factor
    • A contrast is an interpretation of a significant Multi-way ANOVA result.
      • compares groups of means (single df comparisons)
    • Contrast significance is judged by an F-test:
      F=F =SScontrastk(n1) \frac{SS_{contrast}}{k(n-1)}
      • "a priori" means "before"
      • "a posteriori" means "after the fact"
    • There can only be k-1 orthogonal contrasts.
      • statistically independent comparisons = compared only once
      • Product of contrast coefficient = 0
    • Contrast Coefficient: numerical description of hypothesis tested.
      Rules:
      • grouped levels get same sign
      • contrasting levels get opposite sign
      • Excluded levels get 0
      • All coefficients in contrast must sum to 0
    • Fixed effects influence the mean of y
    • Random effects influence the variance of y
      • Include numeric and factor levels
    • Nested sampling reduces random effects by accounting for variation contributed by each factor.
      Used for:
      • studies conducted at different spatial scales
      • repeated measurements from same individual
    • Split-plot analysis reduces fixed effects by splitting a sample into plots of different sizes and applying different treatments.
      • Each plot has own error variance
      • Ordered from largest plot with lowest replication to smallest plot with high replication
      Error term = error(largest/medium/smallest plot)
    • Difference between PCA and RDA:
      A) Variable reduction
      B) data visualisation
      C) Regression analysis
      D) relationship exploration
      E) only x variables
      F) x and y variables
      G) Unconstrained
      H) Constrained ordination analysis
      I) Captures overall data variation
      J) Explains variation in y by looking at variation in x
      K) PCs
      L) Significance test
    • Similarities between PCA and RDA:
      • Use loading systems
      • multi-variate
      • useful for large datasets
    • Linear regression is a measure of how steeply the response (y) variable changes with a change in explanatory variable (x)
      • uses least squares regression (line of best fit)
      • both variables are continuous
      • applied mostly in observational studies
    • Maximum Likelihood is applied for parameter estimation to increase the probability of observed data appearing
    • Regression line is calculated by: Y=Y =a+ a +bX bX
    • The regression slope is calculated by:
      b=b =(XiX)(YiY)(XiX)2 \frac{\sum (X_{i} - \overline{X})(Y_i - \overline{Y})}{\sum (X_i - \overline{X})^{2}}
    • The least squares regression line always goes through the means of x and y
      a=a =YbX \overline{Y} - b\overline{X}
    • The assumptions of linear regression:
      1. at each X, the mean of Y lies on the regression line
      2. at each X, the distribution of Y is normal
      3. at each X, Y variance is the same
      4. at each X, Y is a random sample from all possible Ys
    • Variance of residuals (MSresidual) quantifies the spread of the scatter above and below a regression line
      • df = n - 2
      • estimate slope and intercept
    • Outliers create non-normal distributions and affect estimates
      • Can affect slope calculations
      • Can cause equal variance assumption violations
    • Residuals plots can show normality and variance of the data
    • Uncertainty of estimation of the slope is measured with
      SEb=SE_b =MSresidual(XiX)2 \sqrt{\frac{MS_{residual}}{\sum (X_i - \overline{X})^2}}
    • Hypothesis testing with regression is used to evaluate whether the slope is equal to null slope, or β0.
      • under null, the df = n-2
      • t=t =bβ0SEb \frac{b - \beta_0}{SE_b}
    • Regression takes the deviation between an observation Y and mean Y and breaks it into a:
      • Residual component
      • YiY^Y_i - \hat{Y}
      • Regression component
      • Y^Yˉ\hat{Y} - \bar{Y}
      • if H0 true, both MS will be equal
      • if H0 not true, regression MS > residual MS
    See similar decks