Reliability Flashcards

(34 cards)

1
Q

Reliability definition?

A

the extent to which a measurement of behav is consistent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

internal reliability definition?

A

consistency within a study or measure (tool and/or researcher)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

across researchers def?

A

more than 1 researcher - is there consistency in how they are interpreting behav or carrying out the study

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

inter-rater reliability def ?

A
  • internal reliability
  • 2+ observers watching same behaviour should have same data
  • both can agree on
  • mainly used in observations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

across measures reliability def?

A
  • MORE THAN 1 MEASURE - if there is more than 1 test (such as used in a repeated measures design) then is there consistency of challenge in each test?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

split-half reliability def ?

A
  • used to test IR
  • answers to test are halved and compared - should be similar
  • questionnaire/ test of behav
  • ppts/ measures ect..
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

External reliability definition def?

A

consistency across time (different occassions)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

across time or different occasions def?

A
  • Behaviour is only measured once - the test may not measure behav consistenly at other times e.g. a personality test may only indicate ppt’s personality traits at the time the study was carried out
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

test - retest reliability def?

A
  • assess ER
  • test scorers compared, should be same
  • retested again (after a while)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Dealing w issues of IR - more than one researcher?

A
More than 1 researcher: 
      - clearly operationalise variable so less room for 
        interpretation 
      - standardise procedures
      - detailed training of researchers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Dealing w issues of IR - multiple measures?

A

multiple measures:

- standardize measures - make sure they are the same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Dealing w issues of IR - within a measure?

A

within a measure:

- standardise each question or item

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

method: Lab environment, reliability:?

A

High R - high control of the envir - control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Method: field experiment, Reliability:?

A

Low R - experiment in real life conditions - low control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Method: quasi experiment - Natural, R:?

A

Low R - naturally occurring IV -low control over EVs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Questionnaires - Reliability?

A

High R - ppts can repeat questions ro check for consistent results

17
Q

Interview - unstructured: R?

A

Low reliability as each ppt is asked diiff questions - diff to repeat

18
Q

Interview structured: R?

A

High R - each ppt is asked the same questions so it can be repeated to check for consistent results

19
Q

Observation - controlled: R?

A

High R - behav is observed in a controlled envir - control over EVs

20
Q

Observation - naturalistic: R?

A

Low R - behav is observed in real life envir - low control over EVs

21
Q

Case study: R?

A

Low R - unique cases - impossible to repeat in same conditions

22
Q

Content analysis: R?

A

High R - behaviour categories - can be repeated/ checked by other researchers

23
Q

Assessing IR: Inter- rater reliability?

A
  • all of the researchers carry out a pilot on the same ppts, if they are 80% or higher in agreement then they are reliable
  • Pilot the different measures on the same ppts w/o changing the conditions and then compare their scores to ensure consistency
24
Q

Assessing IR: Split half?

A

Measure behaviour using 1st half of the test and then the second half of the test using same ppts, if scores are similar, then the test is consistent throughout

25
Assessing ER?
- Test - retest | - The same ppts repeat the same test at another time e.g. 4 weeks later, to show a consistency over time
26
How to deal with ER?
Ensure tests ect. Are designed in relation to measuring behaviours, attitudes ect beyond one setting or time
27
What is assessing reliability?
Measures used to check the consistency of a set of results
28
DEFINITION FOR EXAM - inter rater reliability?
- Method of assessing IR - Where 2 or more psychologists in the same study produce consistent results by using a standardised procedure , agreed coding system, or correlation of their data. - This can be checked for by comparing each psychologists results if there is 80% + agreement then = IR
29
DEFINITION FOR EXAM - split half?
- Method of assessing IR - Involves splitting a pp’s test answers in half and seeing whether s/he got the same or similar scores on the 2 halves. If so, IR is high
30
DEFINITION FOR EXAM - test re-test?
- Method of assessing external R - Involves testing and retesting the same ppts over time, with the same test, and comparing their scorers. If the scores are the same, the test has ER
31
What are the issues of internal reliability?
1- more than 1 researcher 2 - more than 1 measure 3 - within a measure
32
What are the issues of external reliability?
Behaviour is only measured once
33
How to deal with issues of IR?
1- inter rater reliability | 2 - split half
34
How to deal w issues of ER?
1- test retest