The following definitions have been drawn from numerous sources and are tailored specifically for content on the NREPP Web site. The terms defined here may have slightly different meanings in other settings.
View All | A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z
- In scientific studies, a number representing the statistical probability that an outcome is due simply to chance alone rather than as a result of the intervention. Typically, if the p-value is less than or equal to 1 in 20 (i.e., ≤ 5 in 100, or p ≤ .05), researchers can conclude that the outcome of the study is not due to chance alone.
- A study design in which (1) there are no control or comparison conditions and (2) data are collected at pretest or posttest only; includes simple observational or case studies. The preexperimental study design provides the most limited scientific rigor of the three types of designs (experimental, quasi-experimental, and preexperimental).
- Process evaluation
- An evaluation to determine whether an intervention has been implemented as intended.
- Program drift
- A threat to fidelity due to compromises made during implementation.
- Program fit
- The degree to which a program matches a community’s needs, resources, and implementation capacity.
- Promising Program
- A few Promising Programs were re-reviewed for NREPP using updated criteria in 2006-2007 and can now be found by searching for the program on the Find an Intervention page.
- The construction of instruments and procedures for measurement.