Final Flashcards
(176 cards)
An evaluation that uses the data collected before a program begins. Examples include pilot projects, baseline surveys, and feasibility studies.
Design evaluation
Coefficient of stability: Evaluator administers the same instrument twice, separated by a short period of time. Results are compared, using a statistic such as a correlation coefficient
Reliability coefficient/ test-retest reliability
Evaluator administers two equivalent versions of the same instrument (parallel forms) to the same group of people. Results are compared, using a coefficient of stability.
Alternate-form coefficient
A nonexperimental research method relying on questionnaires or interview protocols
Survey
The extent to which researchers can make this statement with confidence
Internal validity
Study conducted at a single time period, and data are collected from multiple groups
Cross-sectional study
Data are collected at two or more points in time
Longitudinal study
Design that combines cross-sectional and longitudinal elements by following two or more age groups over time
Cohort-sequential design
Two observers’ data are compared to see whether they are consistently recording the same behaviors when they view the same events. Statistical procedures such as correlation or percentage of agreement can be used to establish this type of reliability.
Interrater reliability
Interpretive research approach relying on multiple types of subjective data and investigation of people in particular situations in their natural environment
Qualitative research
Use of multiple data sources, research methods, investigators, and/or theories/ perspectives to cross-check and corroborate research data and conclusions
Triangulation
used to determine whether a single observer is consistently recording data over a period of time.
Intrarater reliability
To what degree does all accumulated evidence support the intended interpretation of scores for the proposed purpose?
construct validity
commonly used data collection instruments or procedures designed to measure personality, aptitude, achievement, and performance.
Tests
Items on the test represent content covered in the program (e.g., did the teacher teach the children that blue on the map means water?). Evaluators can work with content specialists to list the content that is part of the program, and can compare the test items to see whether they correspond.
Content-related evidence
a self-report data collection instrument that is filled out by research participants.
Questionnaire
The researcher takes independent samples from a general population over time and the same questions are asked.
Trend study
a situation where the interviewer asks the interviewee a series of questions
Interview
indicates that the measure actually reflects current or future behaviors or dispositions.
Criterion-related evidence
a situation where a focus group moderator keeps a small and homogeneous group (of 6–12 people) focused on the discussion of a research topic or issue.
Focus group
Evaluators need to be aware of the consequences of using data, especially with regard to the potential to worsen inequities.
Consequential evidence
Researcher watches and records events or
behavioral patterns of people
Observation
Observation conducted in real-world situations
Naturalistic observation
Observation conducted in lab setting set up by the researcher
Laboratory observation