This site is 100% ad supported. Please add an exception to adblock for this site.

Program Evaluation

Terms

undefined, object
copy deck
Reactive
distorted by presence of evaluator
Typical case sampling
program subjects that typify norm
reliable
consistently and dependably measures concept with accuracy
Saturation
when data reveal repetition and redundancy, confirm existing findings.
internal consistency
tool to measure reliability, see how well each item in scale correlate with scale as whole
shadow price
attempt to reflect real cost of goods and sevices
Patterns of use
client utilization data: rates-under-treatment approach: records, reports, files coming directly from agency
Comparative need
inferred measure of need from finding characteristics in population similar to those receiving services
Deviant case sampling
to learn about outliers
client satisfaction
relevance, response, relationship, results
Snowball sampling
needed when interest population is isolated, hard to reach tapping into existing social networks
emic codes
indigenous categories of meaning, categories constructed from participant classification
subscales
to measure multiple dimensions of functioning
index
creating new variable as sum of other variables ie. 25 items to measure client's anxiety
criterion validity
instrument can be validated by external criterion
construct validity
testing presumed relationships of scale to other variables
Mixed methods
mixing quantitative and qualitative methods
Reflexivity
self-monitoring to note personal biases
split-half
dividing scale in half to see how well each half correlates with another
test-retest
reliability when scale is adminstered to same group on repeated occasions
external validity
extent to which finding is generalizable
validity
when instrument corresponds to concept designed to measure
triangulation
relying on more than 1 type of data to compile findings and enhance accuracy
Normative need
condition/situation defined by an expert
Evaluation research
investigation to determine success of program
service ticket
initial application data file
Operational definition
way concept or variable is defined for evaluation purposes
confidence level
how likely to expect similar results on repeat study
Expressed need
demand/requests for service
known-groups technique
adminstering instrument to 2 different groups to find differences in responses
Social indicators
variables used to gauge extent of social problem
etic codes
view from outside, researcher constructed codes
face validity
when collegue's look over instrument and agree it will be useful to measure concept
scale
item that measures solitary concept ie. anxiety
Service philosophy
no one is turned away", or certain eligibility guidelines
Ecological fallacy
using social indicators to estimate conditions of persons in geographic area.
predictive validity
type of criterion validity of future behavior
negative case anaylsis
finding cases that refute findings
group research designs
tool to examine program outcomes
Felt need
perceptions of need determined by consulting actual clients
Program evaluation
provide information to be used to improve social programs
efficiency
measuring cost-effectiveness and cost-benefit analysis
inter-rater reliability
improve observer matching through training and role playing
content validity
if concept is represtned in items used in scale
concurrent validity
type of criterion validity to predict current status
alpha
reliability coefficient

Deck Info

46

permalink