Study Guide Exam 1 (Assessment and Diagnosis) Flashcards

(128 cards)

1
Q

Earliest forms of testing

A

China: interview tests for civil servants

Han dynasty: written tests for people wanting to work in government

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Physiognomy

A

Practiced by ancient Greeks
Belief that internal characteristics are shown in external features
Example: person with upturned nose is arrogant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Phrenology

A

Germany in 1700s
Franz Joseph Gall
Examining bumps on the head: “organs” in the brain that had been exercised, contributing to personality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Psychophysics

A

Theorized that there is a difference between the physical world and the world that people experience
Tested absolute threshold and just noticeable difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Absolute threshold

A

Related to psychophysiology

Amount of stimulus needed to detect as present 50% of the time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Just noticeable difference

A

Related to psychophysiology

Amount of change in stimulus needed to detect as present 50% of the time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Darwin

A

Contributed to psychology: focusing on individual differences between individuals (intra-species variability)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Galton

A

Key person in “Brass instruments era”
Tested individual differences between people to determine intelligence
Sensory and motor, questionnaires, and physical attributes were tested

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Cattell

A

Performed work similar to that of Galton (“Brass instruments era”)
Coined the term “mental tests”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Wissler

A

Brought about the end of the “Brass instruments era”

Discovered that physical attributes (particularly sensory and motor) didn’t actually correlate with intelligence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Wundt

A

Father of modern psychology

Ran first psychology lab

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Kraeplin

A

First person to classify mental illness

Tested for emotional handicaps (interview-based test of emotional regulation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Esquirol

A

Developed test to determine degrees of mental retardation

Mental ability was classified according to verbal ability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Binet

A

First person to develop intelligence test, which was used to determine placement of children in school

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Army alpha/beta

A

Used to place people as officers or non-officers
Alpha- verbal test
Beta- non-verbal test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Wechsler scales

A

Intelligence test used the most in modern times

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Woodworth

A

Created first personality test (Personal Data Sheet)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Examples of personality tests

A

Rorschach inkblot test
Thematic Apperception Test
Minnesota Multiphasic Personality Inventory

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Test

A

Procedure in which a sample of an individual’s behavior is obtained, evaluated, and scored using standardized procedures

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Measurement

A

A set of rules for assigning numbers to represent objects, traits, attributes, or behaviors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Assessment

A

Systematic procedure for collecting information that can be used to make inferences about the characteristics of people or objects

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Reliability

A

Consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Validity

A

Accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Maximum performance tests

A

Goal: best performance
Examples: classroom tests, intelligence tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Achievement tests
Test specific skills
26
Aptitude tests
Assess abilities
27
Objective tests
Specified scoring (clear right and wrong answer)
28
Subjective tests
Require judgment to evaluate (no clear right and wrong answers)
29
Power tests
Unlimited time | Give best performance
30
Speed tests
Timed tests | Usually fairly easy
31
Typical response tests
Survey | No right or wrong answer
32
Objective typical response tests
Answers can be calculated without subjectivity (like surveys)
33
Subjective typical response tests
Project something about self onto stimulus
34
Norm referenced scores
``` Your score is dependent on others' scores Percentile rank (ex- ACT, SAT, GRE) Tests that have been curved fall into this category ```
35
Criterion referenced scores
There is a set criterion for success Typical of classroom tests Doesn't depend on others' performance
36
Norm samples: what they need to be
Representative of the population taking the test Consistent with that population Current (must match current generation) Large enough sample size
37
Types of norm samples
Nationally representative sample (reflects society as a whole) Local sample Clinical sample (compare to people with given diagnosis) Criminal sample (utilizing criminals) Employee sample (used in hiring decisions)
38
Flynn effect
Intelligence increases over successive generations | In order to stay accurate, intelligence tests must be renormed every couple of years
39
Raw scores
Number of questions answered correctly on a test | Only used to calculate other scores
40
Mean and standard deviation for z scores
M=0 | SD=1
41
Mean and standard deviation for t scores
M=50 | SD=10
42
Mean and standard deviation for IQ scores
M=100 | SD=15
43
Example of age equivalents
13 and performing at an 11 year-old level
44
Example of grade equivalents
In the 8th grade and performing at a 6.5 grade level
45
3 types of criterion-referenced interpretations
``` Percentage correct Mastery testing (pass/fail) Standard-based interpretation (assigning letter grade) ```
46
Classical test theory equation
Xi=T+E Xi- obtained score T- true score E- error
47
Content sampling error
Difference between sample of items on test and total domain of items
48
Time sampling error
Random fluctuations in performance over time | Can be due to examinee (fatigue, illness, anxiety, maturation) or due to environment (distractions, temperature)
49
Inter-rater differences
When scoring is subjective, different scorers may score answers differently
50
Clerical error
Adding up points incorrectly
51
Test-retest reliability
Administer the same test on 2 occasions Correlate the scores from both administrations Sensitive to sampling error
52
Alternate-form reliability
Develop two parallel forms of test Administer both forms (simultaneously or delayed) Correlate the scores of the different forms Sensitive to content sampling error (simultaneous and delayed) and time sampling error (delayed only)
53
Split-half reliability
Administer the test Divide it into 2 equivalent halves Correlate the scores for the half tests Sensitive to content sampling error
54
Kuder-Richardson and coefficient (Cronbach's) alpha
Administer test Compare each item to all other items Use KR-20 for dichotomous answers and Cronbach's alpha for any type of variable Sensitive to content sampling error and item heterogeneity Measures internal consistency
55
Inter-rater reliability
Administer test 2 individuals score test Calculate agreement between scores Sensitive to differences between raters
56
Composite scores
Scores that are combined to form a combined score | Reliability of these is usually better than their individual parts
57
Difference scores
Calculated difference between 2 scores Reliability of these is usually lower than their individual parts (information is lost: only can see change, not initial baseline)
58
High-stake decision tests: reliability coefficient used
Greater than 0.9 or 0.95
59
General clinical use: reliability coefficient used
Greater than 0.8
60
Class tests and screening tests: reliability coefficient used
Greater than 0.7
61
How to improve reliability
Increase number of test items Use composite scores Develop better items Standardize administration
62
Standard error of measurement (SEM)
Standard deviation of test administered to the same individual an infinite number of times Useful when interpreting test scores When reliability increases, this decreases
63
What are used to calculate confidence intervals?
Use SEM and SD
64
Generalizability theory
Shows how much variance is associated with different sources of error
65
Construct underrepresentation
Test doesn't measure important aspects of the specified construct Similar to content sampling error
66
Construct-irrelevant variance
Test measures features that are unrelated to the specified construct
67
External threats to validity
Examinee characteristics (ex- anxiety, which hinders examinee) Deviation from standard test administration and scoring Instruction and coaching Standardization sample isn't representative of population taking test
68
Content validity
Degree to which the items on the test are representative of the behavior the test was designed to sample
69
How content validity is determined
Expert judges systematically review the test content | Evaluate item relevance and content coverage
70
Criterion-related validity
Degree to which the test is effective in estimating performance on an outcome measure
71
Predictive validity (form of criterion-related validity)
Time interval between test and criterion | Example: ACT and college performance
72
Concurrent validity (form of criterion-related validity)
Test and criterion are measured at same time | Example: language test and GPA
73
Construct validity
Degree to which test measures what it is designed to measure
74
Convergent validity
Correlate test scores with tests of same or similar construct: look for convergence
75
Divergent/discriminant validity
Correlate test scores with tests of dissimilar construct: look for divergence
76
Factor analysis
Used to determine if test is measuring factors related to the given construct Assign factor loadings (similar to correlation coefficients): variables should have high loadings on only 1 factor
77
Evidence based on internal structure
Examine internal structure to determine if it matches the construct being measured
78
Evidence based on response processes
Is the manner of responses consistent with the construct being assessed?
79
Evidence based on consequences of testing
If the test is thought to result in benefits, are those benefits being achieved?
80
Incremental validity
Determines if the test provides a gain over another test
81
Face validity
Determines if the test appears to measure what it is designed to measure Not a true form of validity Problem with tests high in these: can fake them
82
Internal vs. external validity
Internal: Does the measure work in ideal conditions? External: Does it work in the real world?
83
Multitrait-multimethod approach to determining construct validity
Use multiple measures for same constructs to check for convergence as well as measures for other constructs to check for divergence
84
Contrasted group study approach to determining construct validity
Create 2 separate and different groups: administer test and look for differences between them
85
6 steps of test construction
1. Define the test (what are we testing and why) 2. Select item format 3. Construct test items 4. Test the items (determine reliability and validity) 5. Revise the test 6. Publish the test
86
Answer choice formats: selected-response vs. constructed-response items
Selected items: pick from a number of answers (multiple choice, true/false, matching) Constructed items: generate your own answers (short answer, essay)
87
Strengths of selected-response items
Can include more items (each question takes less time to answer) Increased content sampling as well as reliability and validity Reduction of construct-irrelevant factors Scoring is efficient and reliable
88
Weaknesses of selected-response items
Developing items is time consuming (easier to write constructed items) Unable to assess all abilities Subject to random guessing (make it look like examinee knows more than he/she actually does)
89
Strengths of constructed-response items
Questions are relatively easy to write Can assess higher-order cognitive abilities (have to show reasoning) No random guessing
90
Weaknesses of constructed-response items
Test can include relatively few items (takes longer to answer each one) Difficult to score reliably (even with good rubric, still hard) Subject to misinterpretation (examinee might misconstrue question) Construct-irrelevant factors can sneak in (ex- bad handwriting makes answers hard to read)
91
3 things on a test that should be clear
``` Clear directions (examinee should know how to answer question) Clear questions (questions should only ask 1 thing; answering should be able to be done in a decisive manner) Clear print (should be easy to read) ```
92
5 things that should not be included on a test
Cues to answers (ex- including answer in a different question) Items that cross pages (increases likelihood of examinee error) Construct-irrelevant factors Exact phrasing from materials (encourages rote memorization over understanding of concept) Biased language and content
93
2 things to consider surrounding placement of items on a test
Item arrangement: placement should make sense | Number of items: if using a power test, should be able to complete questions in given time limit
94
Type of material that should be used on a matching test
Homogenous material (all items should relate to a common theme)
95
Multiple choice tests: what kinds of stems should not be included?
Negatively-stated ones | Unclear ones
96
Multiple choice tests: how many alternatives should be given?
3-5
97
Multiple choice tests: what makes a bad alternative?
Long Grammatically incorrect in question Implausible
98
Multiple choice tests: how many best/correct answers per question?
1
99
Multiple choice tests: how should placement of correct answer be determined?
Random (otherwise, examinees can detect pattern)
100
Multiple choice tests, true/false tests, and typical response tests: what kind of wording should be avoided?
"Never" or "always" for all 3 "Usually" for true/false "All of the above" or "none of the above" for multiple choice
101
True/false tests: how many ideas per item?
1
102
True/false tests: what should be the ratio of true to false answers?
1:1
103
Matching tests: ratio of responses to stems?
More responses than stems (make it possible to get only 1 wrong)
104
Matching tests: how long should responses and lists be?
Brief
105
Essay tests and short answer tests: what needs to be created?
Scoring rubric
106
Essay tests: what kinds of material should be covered?
Objectives that can't be easily measured with selected-response items
107
Essay tests: how should grading be done?
Blindly
108
Short answer tests: how long should answers be?
Questions should be able to be answered in only a few words
109
Short answer tests: how many correct responses?
1
110
Short answer tests: for quantitative items, what should be specified?
Desired level of precision
111
Short answer tests: how many blanks should be included? How long should they be?
Only 1 blank included Should be long enough to write out answer Otherwise, becomes dead giveaway
112
Short answer tests: where should blanks be included?
At the end of the sentence
113
Typical response tests: what should be covered?
Focus items on experiences (thoughts, feelings, behaviors) | Limit items to a single experience
114
Typical response tests: what kinds of questions should be avoided?
Items that will be answered universally the same | Leading questions
115
Typical response tests: how should response scales be constructed?
If neutral option is desired, have odd numbered scale High numbers shouldn't always represent the same thing Options should be labeled as Likert-type scale (rating from 0-7, etc.)
116
Pilot testing
Test on a few people Get feedback Practice scoring Assess problem areas
117
Large scale testing
Develop norm sets | Evaluate reliability, validity, factors
118
What hypotheses for planning assessment are based on
Referral question Presenting concerns Intake interview results Behavioral observations
119
Typical intake interview
Presenting concerns (must start with what client was asking for) Case history Diagnostic questions Mental status exam
120
2 sources of information gained from interviews
``` Content (what is said; thoughts and feelings) Behavioral observations (what is displayed) ```
121
Things examined in behavioral observations
General appearance and behavior Mood and affect Sensorium (awareness of situation) Perception (vision, hearing, etc.: influence what tests are administered) General intelligence Higher cognitive functions (speech and form of thought, insight and judgment, memory, attention and concentration)
122
How to build rapport with a client
Comfortable atmosphere Collaborative stance Acceptance, understanding, empathy, respect
123
2 types of questions in an interview
Close-ended (produce 1 or 2- word answers; used to gather specific information) Open-ended (require longer answers; gather lots of information)
124
Clarification
Questioning client to gain additional understanding from an ambiguous answer or confirm accuracy of clinician's perception "Are you saying that..." "Could you describe for me..." "Say what you mean by..."
125
Reflection
Describing feelings of client's message to encourage the client to continue to express their feelings, have the client feel the emotion more intensely, and help the client become more aware of and discriminate between their feelings
126
Paraphrasing
Describing the content of the client's message to provide opportunity for client to clarify, encourage client to expand on thoughts, and provide an opportunity to redirect client to central topic
127
Summarizing
Two or more paraphrases/reflections that condense the client's message to tie together multiple elements in a common theme, interrupt excessive talking, and review progress
128
Affirmations
Directly affirming and supporting the client through the interview process to acknowledge the client's struggles and build rapport Must be careful not to overuse (can sound disingenuous)