Reliability= the consistency of the results, indicates whether results are trustworthy enough to draw firm conclusions. Found out through repetition of studies
Validity= we are measuring what we set out to measure, affects whether we can generalise the results
Extraneous variable= any variable that isn't the IV, but could affect the DV. Need to control, as they can reduce our validity
Extraneous variables: Situational variables= connected to research situation, ie temperature, time of day, researcher's behaviour
Extraneous variables become confounding variables if they are not controlled for, and therefore affect one of the conditions of your IV
Control participant variables by:
randomly assigning people to each condition- should hopefully "spread out" participant variables
Repeated measures/ matched pairs design
Situational variables: demand characteristics
is a change to participant's behaviour as a result of being in the study- behaviour is no longer natural, so validity is compromised.
cues in the research environment that lead pps to work out the aim
control by: single blind technique- pps don't know which condition they're in, they are unable to guess how the researcher is expecting them to behave
Situational variables: investigator effects
investigator behaves in a way to increase the chance of their hypothesis being correct- through the way instructions are given or gestures etc
Control by: double blind- both researcher and pps unaware of the condition each is in (reduces both demand characteristics & investigator effects)
Control by: standardisation- all instructions given to pps are standardised (made identical using a script etc) reducing investigator effects
Situational variables: order effects
only relevant when repeated measures design used
pps get better between condition 1 & 2 (practice), or get worse (fatigue).
Control by: counterbalancing- may get half of the pps to do condition 1 first and the other half do condition 2 first
if they do get better/ worse, the two different groups will cancel each other out
Control by: randomisation- changing order the conditions are completed every time
Situational variables: observer effect
when a pp is aware they're being observed, which therefore affects their behaviour & makes it less natural
Control by: double blind- neither pp or researcher know which condition the pp is in- reduces the likelihood of behaviour changes
Assessing/ measuring reliability:
Inter-observer reliability- how closely do the ratings of the two people administering the test correlate, look for strong positive correlation between observers.
Test-retest reliability- whether the same test administered on 2 different occasions gives highly positively correlated scores
Improving reliability:
Pilot studies- a small scale test of methods/ procedures being used on a larger scale. Allows you to check the proposed method of measurement for issues that might compromise reliability
Average of multiple results- find an average from more than one measurement from each pp to reduce the impact of anomalous results
Internal validity is the extent to which:
Findings can be attributed to the effect of the IV, ie minimised extraneous variables rather than demand characteristics
The DV accurately measures the behaviour we are interested in- down to operationalisation of the variables, ie the way you have decided to measure your variable
External validity= the extent to which results can be generalised outside of the experiment
Ecological validity- to other settings/ real life & influenced by how realistic the setting or task is to the behaviour being measured
Temporal validity- to other time frames & influenced by whether aspects such as specific cultural pressures are considered
Population validity- to other people & influenced by how representative and unbiased the sample is
Assessing validity:
Face validity- does it look correct/ valid in the eyes of experts
Concurrent validity- involves assessing how closely the scores on one test correlate with another already established assessment of the same behaviour
Investigator effects- control by double blind/ standardisation
Demand characteristics- control by single blind
Order effects- control by counterbalancing
Choosing experimental methods carefully; environments which allow you to control the variables can improve internal validity, but may compromise external validity.
Using face validity or concurrent validity assessments to make alterations (amending questions or putting controls in place).
Reliability refers to how much we can depend on a measurement- is all about consistency. We can only draw firm conclusions if results are reliable, also reliability does not mean accuracy.
Observation:
Form of measurement- a researcher is recording what people/ animals are doing.
Assessing reliability: assessed by the observer. Repeating the observations for a second time. Looking for strong positive correlation.
Improving reliability: operationalising categories more, observers practising using the categories.
Self-report:
Research method where the individual tells a researcher what they think/ feel.
Assessing reliability: test-retest reliability, measure should be reliable.
Improving reliability: reduce ambiguity- test items need to be re-examined and possibly re-written.
Experiment:
The dependent variable in an experiment is often measured using a rating scale or behavioural categories.
Assessing reliability: testing if the method used to measure the DV is consistent, eg test-retest.
Improving reliability: standardisation- same procedures exactly each time when repeating experiments.
Validity refers to the truthfulness or legitimacy- does the results measure what they claim to measure (ie are they a true reflection of the behaviour or are they down to the set up of the study). Validity is important as it affects our ability to generalise the results.