Assessment Theory
How to read this page: This article maps the topic from beginner to expert across six levels � Remembering, Understanding, Applying, Analyzing, Evaluating, and Creating. Scan the headings to see the full scope, then read from wherever your knowledge starts to feel uncertain. Learn more about how BloomWiki works ?
Assessment Theory is the study of how we "Measure" the human mind. While most people think of assessment as just "The Big Test at the end of the year," theory argues that assessment is actually the "Heartbeat" of the learning process. It is the art of gathering "Evidence" to answer three questions: "Where is the learner now?", "Where are they going?", and "How do they get there?" From the high-pressure "Summative" exams to the gentle "Formative" feedback of a daily check-in, assessment theory explores how to measure growth fairly, accurately, and without destroying the student's motivation to learn.
Remembering
- Assessment — The systematic process of documenting and using empirical data on the knowledge, skill, attitudes, and beliefs to refine programs and improve student learning.
- Formative Assessment — Assessment "For" learning; small, daily check-ins used to adjust teaching (e.g., a quiz or a conversation).
- Summative Assessment — Assessment "Of" learning; big tests at the end of a unit used to give a grade (e.g., a Final Exam).
- Validity — The "Truth" of a test: does it actually measure what it says it measures? (e.g., a math test that is too "Wordy" might accidentally be measuring reading skill instead).
- Reliability — The "Consistency" of a test: would the student get the same score if they took it tomorrow?
- Rubric — A scoring guide used to evaluate "Subjective" work (like an essay or a project) using specific criteria.
- Norm-Referenced — Comparing a student to their "Peers" (e.g., "You are in the 90th percentile").
- Criterion-Referenced — Comparing a student to a "Standard" (e.g., "You got 8 out of 10 right").
- Authentic Assessment — Testing a student on a "Real-world" task (e.g., "Build a working bridge" rather than "Answer a question about bridges").
- Feedback — The "Information" given to a student to help them close the gap between their current state and the goal.
Understanding
Assessment theory is understood through Purpose and Accuracy.
1. Assessment for, of, and as Learning:
- For Learning (Formative): Like a "Coach" giving tips during practice. The goal is to get better.
- Of Learning (Summative): Like the "Final Score" of the game. The goal is to judge.
- As Learning (Metacognition): When the student "Assesses themselves" to see what they know. The goal is to become a better learner.
2. The "Proxy" Problem: We can't "See" inside a student's brain.
- We use a "Test" as a "Proxy"—a substitute for the real thing.
- If a student is "Anxious" or "Tired," the proxy might be "Broken"—their score might be low even if they know the material perfectly.
- Assessment theory is about finding ways to make the "Proxy" as close to the "Reality" as possible.
3. The "Washback" Effect: "What gets tested gets taught."
- If a high-stakes test only measures "Remembering," then the whole school will focus on "Memorizing" rather than "Thinking."
- Assessment theory aims to design tests that "Force" good teaching.
The 'Pygmalion Effect': The psychological finding that a student will perform "Up" or "Down" based on the teacher's expectation. If an assessment "Labels" a student as "Struggling," they often become even more struggling.
Applying
Modeling 'The Reliability Check' (Seeing if a test is 'Broken'): <syntaxhighlight lang="python"> def check_test_reliability(scores_set_a, scores_set_b):
"""
If the same students get wildly different scores on two
identical tests, the test is unreliable.
"""
diffs = [abs(a - b) for a, b in zip(scores_set_a, scores_set_b)]
avg_diff = sum(diffs) / len(diffs)
if avg_diff > 15: # Out of 100
return "UNRELIABLE: The test results are too 'Random'."
else:
return "RELIABLE: The test measures consistently."
- Test results for 5 students across two days
print(check_test_reliability([80, 85, 90, 70, 75], [78, 88, 92, 72, 74])) </syntaxhighlight>
- Assessment Landmarks
- The SAT (1926) → The birth of "Mass Standardized Testing," which changed who got into university but also created a massive debate about "Inherent Bias" and "Test-Prep" inequality.
- PISA (The Program for International Student Assessment) → A global test that ranks entire countries (like Finland vs. China), turning education into a "Geopolitical Competition."
- The 'Portfolio' Movement → A shift in the 1990s away from "Tests" and toward a "Collection of Work" that shows a student's growth over time.
- No Child Left Behind (2001) → A US law that made "Test Scores" the only way to judge a school's success, leading to the "Teaching to the Test" controversy.
Analyzing
| Feature | Formative (The Coach) | Summative (The Judge) |
|---|---|---|
| Timing | During the learning | At the end of the learning |
| Stakes | Low (No 'Grade' usually) | High (Determines the 'Grade') |
| Goal | To improve the next step | To judge the previous steps |
| Feedback | Specific and actionable | Final and descriptive |
The Concept of "Construct Under-representation": Analyzing when a test is "Too small." If you want to test "Being a Scientist," but your test only asks "Can you name the parts of a cell?", you are "Under-representing" the real goal of being a scientist (like curiosity and experimenting).
Evaluating
Evaluating assessment theory:
- The "Anxiety" Factor: Is it "Ethical" to judge a person's future based on a 3-hour test?
- Bias: Do "Standardized Tests" favor students who have more money, better health, and a more "Standard" culture?
- Cheating: In the age of AI (ChatGPT), is it even "Possible" to have a take-home assessment anymore?
- The "Grade" Trap: Does giving a "Letter Grade" (A, B, C) actually "Stop" learning because students only care about the grade rather than the knowledge?
Creating
Future Frontiers:
- Stealth Assessment: Using "Learning Games" that track a student's every click to assess their skill *without them even knowing they are being tested*.
- AI Feedback Loops: An AI that provides "Instant, Personalized Feedback" on every sentence a student writes, 24 hours a day.
- Blockchain Credentials: Moving away from "Diplomas" and toward a "Digital Ledger" of thousands of verified "Micro-skills" that a person can prove to an employer.
- Dynamic Assessment: A test that "Changes its difficulty" in real-time based on the student's previous answers to find the exact "Limit" of their knowledge.