Biostats Exam 4

Subdecks (1)

Cards (62)

  • Correlation analysis
    Analyses the relationship between two variables
  • Regression analysis
    Analyses the relationship between a dependent variable and one or more independent variables
  • Correlation and regression analyses
    Differentiate between the two
  • Correlation coefficient
    Measures the strength and direction of the linear relationship between two variables
  • Types of correlation coefficients
    • Pearson's correlation coefficient
    • Spearman's rank correlation coefficient
    • Kendall's tau
  • Interpretation of correlation coefficients
    • Positive, negative, or no correlation
    • Strength of correlation (weak, moderate, strong)
  • Scatterplots
    Visualize the relationship between variables
  • Correlation analysis has limitations and assumptions
  • Regression model
    Consists of predictor variables (independent variables), outcome variable (dependent variable), and a regression equation
  • Types of regression models
    • Simple linear regression
    • Multiple linear regression
    • Logistic regression
  • Slope
    Indicates the change in the dependent variable for a one-unit change in the independent variable
  • Intercept
    The value of the dependent variable when the independent variable is zero
  • Residuals
    The difference between the observed and predicted values of the dependent variable
  • Regression analysis has assumptions: linearity, independence of errors, homoscedasticity, normality of residuals
  • Model fit and goodness of fit measures
    1. squared, adjusted R-squared, AIC, BIC
  • Statistical software is used to conduct correlation and regression analyses
  • Statistical software output should be interpreted to understand the results of correlation and regression analyses
    • Parametric tests make assumptions about the underlying data distribution and are suitable for interval or ratio data.
    • Nonparametric tests do not make distributional assumptions and are suitable for ordinal, nominal, or non-normally distributed data.
    • Parametric tests are more powerful and efficient when the assumptions are met, while nonparametric tests are more flexible and robust but may have lower power in certain situations.
  • When assumptions of normality are violated, the underlying data does not follow a normal distribution
  • Violations of normality can lead to inaccurate results and reduced statistical power in parametric tests
  • Presence of outliers can greatly influence the results of statistical analyses
  • Outliers
    Data points that are significantly different from the rest of the data
  • Outliers can skew the distribution and impact the estimates of central tendency and variability
  • Outliers may distort the results of both parametric and nonparametric tests, leading to biased estimates and inaccurate conclusions
  • Small sample size
    Having a limited number of observations in the dataset
  • With small sample sizes, statistical tests may lack power to detect true effects or differences between groups
  • Small sample sizes can also increase the risk of Type I and Type II errors, making it challenging to draw reliable conclusions from the data
  • Data measurement inaccuracy
    Errors or inaccuracies in the measurement process
  • Measurement inaccuracy can introduce noise and bias into the data, affecting the validity and reliability of statistical analyses
  • Alpha (α)

    The significance level chosen by the researcher before conducting a hypothesis test
  • Type I error
    The rejection of a true null hypothesis
  • Commonly used significance levels
    • 0.05 (5%)
    • 0.01 (1%)
  • Alpha
    Serves as the threshold for determining whether the observed results are statistically significant
  • Smaller alpha value
    Indicates a more stringent criterion for rejecting the null hypothesis
    1. value
    The probability of observing the test statistic (or more extreme) under the assumption that the null hypothesis is true
  • Smaller p-value
    Suggests stronger evidence against the null hypothesis and greater support for the alternative hypothesis
  • If p-value is less than or equal to alpha (p ≤ α)

    The null hypothesis is rejected in favor of the alternative hypothesis
  • If p-value is greater than alpha (p > α)

    The null hypothesis is not rejected
  • Statistical significance decision rule
    A guideline for making decisions based on the comparison of the p-value to the chosen significance level (alpha)
  • If p-value is less than or equal to alpha (p ≤ α)

    The results are considered statistically significant, and the null hypothesis is rejected