Test Development Flashcards

1
Q

What are the first six steps to developing a measure?

A

1) overall goal and pre-planning
2) content definition
3) test specifications
4) item development
5) test design and assembly
6) test production

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the last six steps to developing a measure?

A

7) test administration
8) scoring responses
9) establishing passing scores
10) reporting results
11) item banking
12) test technical report

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the first step for developing a measure and what does it aim to do?

A

overall goal and pre-planning: provides a systematic framework for the project

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what happens during the overall goal and pre-planning stage of developing a measure?

A

(1) what is the aim or purpose of the test
(2) what construct will be measured
(3) what will the test format be
(4) how will it be administered. decide on a timeline for developing the test, how we will check the quality of the test, who will produce the test or publish or print it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is an aim?

A

In the first step we have to be clear to specify the aim of the test/measure - specify what you want to achieve with the test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are the different assessments?

A

screening or in-depth assessment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

describe screening tests

A

include fewer items and cover less content, quick test, easy to administer.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

describe in-depth assessments

A

more items, cover more content of the construct. A detailed test is generally more reliable, but can be much more time-consuming. Often detailed tests may also need special training on how to administer and interpret the test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what are the modes for interpretating outcomes?

A

Normative, Ipsative and Criterion-referenced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what is the normative mode of interpretation?

A

compares scores to a norm group, score is compared to the average score of the rest of the sample of test-takers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is the Ipsative mode of interpretation?

A

test that makes it possible to compare different aspects - so a test that allows for intra-individual comparisons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what is the criterion-referenced mode of interpretation?

A

the performance is compared to a pre-defined status - what constitutes a ‘pass’ or ‘fail’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the second step of developing a measure and what does it aim to do?

A

content definition: aims operationally define the construct you are measuring

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what does the word ‘operationalization’ mean?

A

refers to the act of defining or making a ‘fuzzy’ concept measurable. Define what you mean by the concept you are measuring. There are many ways in which you can define a specific concept, eg; The construct of intelligence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the second part content definition when developing a measure (step two)?

A

defining the purpose of the measure - what will you use the scores on the test for?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is step three of developing a measure and what does it aim to do?

A

test specifications: the test blueprint

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

what are some factors in test specification?

A

1) test/response format
2) item format
3) test length
4) content areas of the constructs tested
5) whether items will contain visual stimuli
6) how test scores will be interpreted
7) time limits

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

what are the types of response formats?

A

selected response, constructed response, performance response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

define the selective response format

A

A selected response, is something like a questionnaire with a Likert scale where a test-taker must choose to what extent they agree or disagree with a statement, selected one of the options, hence selected response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

define the constructive response format

A

A constructed response is one where the test-taker must construct the answer, and generate it from their own knowledge. eg; where you are expected to respond by writing an essay, or filling-in-the-blank

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

define the performance response format

A

a performance response is one where the test-taker needs to perform a task, such as where you are asked to build a puzzle, or build a design.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

what are the types of formats for response formats?

A

objective or subjective

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

what is an objective response format?

A

objective response format would be very structured, where you pick one response – there is a right or wrong, or a definite answer such as with an MCQ test or a Likert scale for a questionnaire assessing personality (pick to what extent something resonates with you/applies to you)

24
Q

what is a subjective response format?

A

subjective format would be where the interpretation of the response depends on the examiner’s judgement, so more unstructured.

25
Q

give examples of psychological tests that use subjective response formats

A

Rorschach or TAT.

26
Q

what are the types of item formats?

A

open-ended items, forced-choice items, sentence-completion items, performance-based items

27
Q

define an open-ended item format

A

where there are no limitations on the test-taker.

28
Q

define a forced-choice item format

A

limited to how you can answer. You have to select within the options you are given

29
Q

define a sentence-completion item format

A

asked to complete a sentence - some open-endedness

30
Q

define a performance-based item format

A

items require you to perform a task, like write an essay or present an oral

31
Q

what is important when considering test length?

A

the amount of administration time available, what the purpose of the measure is, test fatigue, enough items to measure in enough detail

32
Q

what does ‘test content areas’ refer to?

A

ensure that all domains of the construct are tested

33
Q

how do we ensure all domains of a construct are tested?

A

a test structure (blueprint) outlining content areas (different categories/subcategories of the construct) and manifestations (How does the construct manifest)
content areas in table column, manifestations in table rows.

34
Q

what is the fourth step for developing a measure and what does it aim to do?

A

item development: involves setting the items in the test

35
Q

what are the guidelines for item development?

A

1) use clear wording to avoid ambiguity
2) use appropriate vocabulary
3) avoid double negatives
4) if writing a test for children consider a different format
5) don’t make questions too obvious

36
Q

what is the fifth step for developing a measure and what does it aim to do?

A

Test design and assembly: assembling the psychological test

37
Q

what things need to be considered when assembling/designing a test?

A

1) placement of correct items
2) check for errors
3) manual or computer assembly
4) how are people going to answer the test?
5) does the test look aesthetically pleasing?

38
Q

what is pre-testing?

A

where you administer the test to a representative sample from the target population to see how well it is working or if it is working in the way that you want it to.

39
Q

what does pre-testing aim to do?

A

Running a pilot assessment in this way, can tell us also whether people answer how we expect them to, if there are any problems or issues with the test, it the items are too easy or difficult, whether the time limit set for the test is enough.

40
Q

what is the sixth step for developing a measure and what does it aim to do?

A

test production: everything that goes into putting together the final version of the test, which includes finalizing which items are included, the sequence they are presented in, and the necessary visual stimuli such as images you’d like to include - It also involves quality control

41
Q

what is the seventh step for developing a measure and what does it aim to do?

A

test administration: Test administration is the most public and visible aspect of testing, when you administer your test to a sample.

42
Q

what are important factors of test administration?

A

security and standardization of testing conditions

43
Q

name methods for standardizing test administration

A

control extraneous variables, make conditions identical for all examinees, standard instructions, same time limits

44
Q

why is standardizing test administration important?

A

comparability between samples, fairness when interpreting test scores, increases validity of findings, increases reliability of findings

45
Q

what are steps eight and nine in developing a measure and what do they entail?

A

scoring responses and establishing a passing score: deciding how we will score a test or questionnaire

46
Q

what aspects go into scoring criteria for a psychological test?

A

developing a scoring key, decide when we drop participants from the study sample, investigate whether there is a response bias, evaluate how good or bad items are through item analysis

47
Q

what is item analysis?

A

analyze whether the test items are good or bad to decide whether to keep or omit their data

48
Q

what are the considerations when doing item analysis?

A

whether the items are congruent with the test objective, and whether they are valid – so questions surrounding validity, questions surrounding reliability, poor performing items that need to be discarded, which items are easy/difficult, how long it takes an examinee to complete each item

49
Q

what is step nine in developing a measure and what does it aim to do?

A

establishing passing scores: set up norms – what the distribution of scores look like in our target population, set performance standards, or decide on cut-off scores

50
Q

what do you have to achieve before establishing scores can take place?

A

ensure the test is reliable and valid

51
Q

what is step ten in developing a measure and what does it entail?

A

reporting results: all examinees have a right to accurate, timely and useful reports of their performance, in understandable language.

52
Q

what is step eleven in developing a measure and what does it entail?

A

Item banking: refers to storing items for future use.
NB! security

53
Q

what is step twelve in developing a measure, and what does it aim to do?

A

publishing and refinement: a technical report detailing the how one would go about administering the test and scoring it and much of the information detailed in the previous steps. You would also then submit the measure to be classified as a psychological test or not and sort out publishing and marketing should you want to charge for it

54
Q

what factors need to be considered when developing a multi-cultural test?

A

education level and language

55
Q

why is it important to develop multi-cultural tests?

A

South African samples are often multi-cultural. If you are targeting South Africans in general, you then have to be especially mindful during the planning phase of several things that might influence performance on tests.