Chapter 6 Flashcards
(45 cards)
Major Steps in Test Development
- Define purpose
- Preliminary design issues
- Item preparation
- Item analysis
- Standardization/research
- Final materials and publication
Statement of Purpose
- Just one sentence
- Focus on trait and scores/interpretation
Mode of Administration
- Group versus individual issue to consider
- Understand processes as person is taking
Length of Exam
- Short limits reliability
- Long makes it more reliable
Item Format
- Type of question to consider when addressing issues
- Some subjectivity in essay responses, etc.
Number of Scores
- Issue to consider when designing
- Bigger test developers with wider application
- Improves marketability
Administrator Training
- How much to be considered when designing exam
Background Research
- Literature search issue
- Discussions with practitioners if used by clinicians
Item Preparation
- Stimulus/item stem (question, apparatus, etc.)
- Response format (M-C, T-F)
- Scoring procedures (correct/incorrect, partial credit, etc.)
- Conditions governing response (time limit, probing of responses, etc.)
Selected Response Items (Fixed Response)
- T/F, M/C, Likert, etc.
- Objectively scored
- Assigning of points
- Keep content right, simple, and don’t be too obvious
Constructed Response Items (Free Response)
Fill in the blank style test items
Inter-Rater Reliability
Scoring requires judgment and a certain degree of agreement is crucial so items are evaluated in the same way or similar way
Holistic
- Scoring scheme in which a single judgment about quality
- Overall impression of what the paper is
Analytic
- Scoring scheme in which it is rated on several different dimensions
- Grammar, organization, vocabulary, criteria, etc.
Point System
Scoring scheme in which certain points must be included for a perfect answer or full credit
Automated Scoring
- Scoring scheme by computer programs that simulate human judgment
- Comes from the “and now” period
- Used for scoring essays too
Suggestions for Writing Constructed Response Items
- Clear directions
- Avoid optional items (chose to answer 3 out of 5 essays)
- Be specific about scoring procedure when preparing questions
- Score anonymously
- Use sufficient number of items to maximize reliability and validity
Three Selected-Response Advantages
- Scoring reliability
- Scoring efficiency
- Temporal efficiency
Two Constructed Response Advantages
- Easier observation of behavior and processes
- Exploring unusual areas such as covering materials multiple choice cannot address
Item Analysis
Involves the statistical analysis of data obtained from an item tryout
Three Phases of Item Tryout
- Item tryout
- Statistical analysis
- Item selection
Informal Item Tryout
5-10 people similar to those for whom the test is intended comment on the items and directions
Formal Item Tryout
Administration of test items to samples of examinees who are representative of the target population for the test
Independent Study
- Conducting a study exclusively for the purpose of item analysis
- Most common practice
- Formal practice for item tryout, subjects often paid