Measures of Reliablity and Validity Flashcards

(39 cards)

1
Q

requires constant collection, evaluation, analysis, and use of quantitative and qualitative data.

A

Clinical Medicine

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Error

A
  • Mistakes in the diagnosis and treatment of patients
  • Mistakes due to clear negligence
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Goal

A

minimize error in data so as to guide, not mislead

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

PROMOTING

A

Accuracy and Precision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What errors do we need to reduce

A

Differential and Indifferential Erros

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

We need to reduce what variability?

A

intraobserver and interobserver

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

closer to the true value

A

Accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Also known as “reproducibility” or “reliability”

•Ability of a test to give the same result or a similar result with repeated measurement of the same factor

A

Precision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Differential error

A

information errors differ between groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

information is incorrect, but is the same across groups

A

Nondifferential error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

refers to any systematic error that may occur during the collection of baseline or follow-up data

A

Measurement bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Examples of Measurement Bias

A

• Blood pressure values
• measuring height with shoes on
• laboratories and the use of different methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

• Variability and unpredictability
• results in lack of precision
• some observations are too high and some are too low

A

Random error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

one observer examining the same results more than once

A

Intraobserver variability (within the observer)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Interobserver variability (between observers)

A

2 or more observers examining the same material

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

measure of the consistency of a metric or a method

17
Q

• Overall percent agreement
• Paired observation
• Multiple variables
• Kappa test ratio

A

MEASURES OF RELIABILITY

18
Q

Common way to measure agreement

A

Overall Percent Agreement (OPA)

19
Q
  • does not include prevalence
  • does not show how disagreement occurred
  • Agreement might be due to chance alone
A

Drawbacks of OPA

20
Q

Percent Agreement Formula

A

PA = a+d/a+b+c+d(100)

21
Q

Measures the extent to which agreement exceeds that expected by chance

A

KAPPA TEST RATIO

22
Q

Kappa =
(Percent Agreement Observed)-(Percent agreement expected by chance alone)/100% -(Percent agreement expected by chance alone)

A

FORMULA FOR KAPPA TEST RATIO

23
Q

INTERPRETATION OF KAPPA

A

<0 = Less than Chance agreement
0.01-0.20 = Slight
0.21-0.40 = Fair
0.41-0.60 = Moderate
0.61-0.80 = Substantial
0.81-0.99 = Almost perfect agreement

24
Q

Ability of a test to distinguish between
WHO HAS a disease and WHO DOES NOT

25
Screening tests
is performed as a preventative measure
26
- able to correctly identify who has the disease - Reliably finding a disease when it is present - Avoids false negative results
Sensitivity
27
Specificity
– correctly identifies who does not have the disease - Reliably excluding a disease when it is absent - Avoids false positive results
28
Type 1 error / false-positive error / alpha error
Finding a positive result in a patient in whom disease is absent
29
Finding a negative result in a patient whom disease is present
Type II error / false-negative error / beta error
30
Sensitivity Formula
TP/TP+FN
31
Specificity Formula
TN/TN+FP
32
False Positive error rate formula
b/(b+d)
33
False Negative Error rate formula
c/(a+c)
34
Predictive Values
Describes the probability of having actual disease given the results of a test
35
Positive predictive value (PPV)
Indicates what proportion of the subjects with positive test results actually have the disease
36
Negative predictive value (NPV)
Indicates what proportion of the subjects with negative test results actually do not have the disease
37
Formula for PPV
PPV = number of people with gold-standard evidence of disease who test positive (a)/ number of people who test positive (a+b)
38
NPV FORMULA
NPV = number of people with gold-standard absence of disease who test negative (d)/ number of people who test negative (c+d)
39
Sensitivity & specificity versus predictive value
Sensitivity and specificity are characteristics of a test. Positive predictive value (PPV) and negative predictive value (NPV) are best thought of as the clinical relevance of a test.