Practicality
AssessmentTest PracticalityFeasibility
Practicality is the feasibility of a test given the available resources — time, money, personnel, facilities, and expertise. A test can be theoretically excellent (high Validity, high Reliability) but utterly impractical to implement. Practicality is the reality check on assessment design.
Dimensions of Practicality
| Dimension | Questions to ask |
|---|---|
| Administration time | How long does the test take? Can it fit within available class/session time? |
| Preparation time | How long does it take to create, pilot, and revise the test? |
| Scoring time | How quickly can results be produced? Can scoring be done in-house? |
| Scorer expertise | Does scoring require trained raters, or can any teacher score it? |
| Cost | Materials, printing, technology, examiner fees, venue hire? |
| Technology requirements | Does it need computers, internet, recording equipment, specialized software? |
| Logistics | Can it be administered to the required number of learners in the available space and time? |
| Score interpretation | Can results be understood and used by the intended audience (teachers, administrators, learners)? |
The Three-Way Trade-Off
Assessment design always involves balancing three competing demands:
[Validity](/terms/validity)
/ \
/ \
[Reliability](/terms/reliability) --- Practicality
- Maximizing validity (e.g., extended writing tasks, live speaking interviews, portfolio assessment) typically reduces practicality (expensive, slow to score, needs expert raters) and may reduce reliability (subjective scoring).
- Maximizing reliability (e.g., many objective items, machine scoring, standardized conditions) may reduce validity (multiple-choice cannot assess productive skills) and has mixed effects on practicality.
- Maximizing practicality (e.g., short, machine-scored, cheap) often reduces validity (limited construct coverage) and may reduce reliability (fewer items = less stable scores).
No test can maximize all three. The art of assessment design is finding the right balance for the context.
Context Determines the Balance
| Context | Priority |
|---|---|
| Low-stakes classroom quiz | Practicality and validity over reliability. Quick, relevant, informative — perfect reliability is not needed. |
| End-of-course exam | Balance all three. Must be valid enough to be fair, reliable enough to be defensible, practical enough to administer. |
| High-stakes standardized test | Validity and reliability over practicality. Worth investing significant resources to get it right. |
| Large-scale placement | Practicality is paramount. Must process many learners quickly. Accept some validity trade-offs. |
Practical Strategies
- Use rubrics. Even simple analytic rubrics improve both reliability and practicality — faster, more consistent scoring than impressionistic marking.
- Pilot before deploying. A quick pilot with a small group reveals timing issues, ambiguous items, and scoring difficulties before they become problems at scale.
- Mix formats. Combine quick objective items (for breadth and reliability) with a smaller number of productive tasks (for validity). This is more practical than all-productive and more valid than all-objective.
- Leverage technology where it helps. Online testing, automated scoring of objective items, and speech recording for later assessment all improve practicality without necessarily sacrificing validity.
- Accept "good enough." Especially in classroom contexts. A reasonably valid, reasonably reliable, practical test administered regularly is more useful than a perfect test that never gets created.