Subtopic Deep Dive
Student Evaluations of Teaching
Research Guide
What is Student Evaluations of Teaching?
Student Evaluations of Teaching (SET) are student-provided ratings assessing instructor effectiveness, scrutinized for validity, biases, and links to learning outcomes.
Research examines SET instruments, gender/race biases, and predictive power for teaching quality (Spooren et al., 2013, 614 citations). Meta-analyses reveal moderate validity but highlight confounders like student expectations. Over 20 papers since 2009 analyze SET reliability across higher education contexts.
Why It Matters
SET scores influence tenure, promotion, and salary decisions in universities, yet biases against female or minority instructors skew outcomes (Kim & Sax, 2009). Valid SET use improves teaching practices and equity in evaluations. Spooren et al. (2013) meta-review shows weak ties to learning gains, urging policy reforms; Tai et al. (2017) links SET to developing student evaluative judgement for better feedback loops.
Key Research Challenges
Bias in SET Ratings
Gender, race, and class biases distort SET scores, disadvantaging underrepresented instructors (Kim & Sax, 2009, 492 citations). Students rate based on stereotypes rather than pedagogy. Spooren et al. (2013) confirm inconsistent validity across demographics.
Validity for Learning Outcomes
SET correlates weakly with student achievement or long-term gains (Spooren et al., 2013, 614 citations). Ratings reflect satisfaction more than skill development. Schunk (1985, 466 citations) notes self-efficacy influences perceptions over objective measures.
Instrument Reliability Issues
Preset criteria lead to indeterminate grading in SET (Sadler, 2008, 461 citations). Mixed methods misalign data from questionnaires and interviews (Harris & Brown, 2020, 476 citations). Standardization remains elusive.
Essential Papers
Learning Styles: An overview of theories, models, and measures
Simon Cassidy · 2004 · Educational Psychology · 990 citations
Although its origins have been traced back much further, research in the area of learning style has been active for--at a conservative estimate--around four decades. During that period the intensit...
Demystifying Content Analysis
A.J. Kleinheksel, Nicole Rockich‐Winston, Huda E. Tawfik et al. · 2020 · American Journal of Pharmaceutical Education · 873 citations
On the Validity of Student Evaluation of Teaching
Pieter Spooren, Bert Brockx, Dimitri Mortelmans · 2013 · Review of Educational Research · 614 citations
This article provides an extensive overview of the recent literature on student evaluation of teaching (SET) in higher education. The review is based on the SET meta-validation model, drawing upon ...
Developing evaluative judgement: enabling students to make decisions about the quality of work
Joanna Tai, Rola Ajjawi, David Boud et al. · 2017 · Higher Education · 600 citations
Evaluative judgement is the capability to make decisions about the quality of work of oneself and others. In this paper, we propose that developing students’ evaluative judgement should be a goal o...
Effective peer assessment processes: Research findings and future directions
Marjo van Zundert, Dominique Sluijsmans, Jeroen J. G. van Merriënboer · 2009 · Learning and Instruction · 505 citations
Student–Faculty Interaction in Research Universities: Differences by Student Gender, Race, Social Class, and First-Generation Status
Young K. Kim, Linda J. Sax · 2009 · Research in Higher Education · 492 citations
This study examined whether the effects of student–faculty interaction on a range of student outcomes—i.e., college GPA, degree aspiration, integration, critical thinking and communication, cultura...
A Critical Review of Research on Student Self-Assessment
Heidi Andrade · 2019 · Frontiers in Education · 477 citations
This article is a review of research on student self-assessment conducted largely between 2013 and 2018. The purpose of the review is to provide an updated overview of theory and research. The trea...
Reading Guide
Foundational Papers
Start with Spooren et al. (2013) for SET validity meta-review (614 citations), then Kim & Sax (2009) for bias analysis, Schunk (1985) for self-efficacy context.
Recent Advances
Tai et al. (2017, 600 citations) on evaluative judgement; Andrade (2019, 477 citations) self-assessment review; Harris & Brown (2020, 476 citations) mixed methods.
Core Methods
Meta-validation models (Spooren et al., 2013), regression on interactions (Kim & Sax, 2009), content analysis (Kleinheksel et al., 2020).
How PapersFlow Helps You Research Student Evaluations of Teaching
Discover & Search
Research Agent uses searchPapers and citationGraph on 'Student Evaluations of Teaching validity' to map Spooren et al. (2013) as central node with 614 citations, then exaSearch uncovers 50+ related works on biases.
Analyze & Verify
Analysis Agent applies readPaperContent to Spooren et al. (2013), runs verifyResponse (CoVe) for bias claims, and runPythonAnalysis with pandas to meta-analyze correlation stats across 10 papers; GRADE grading scores evidence as moderate for SET-learning links.
Synthesize & Write
Synthesis Agent detects gaps in longitudinal SET studies via contradiction flagging, Writing Agent uses latexEditText and latexSyncCitations to draft policy review with 20 citations, latexCompile generates PDF; exportMermaid visualizes bias factor diagrams.
Use Cases
"Correlate SET scores with GPA across demographics using paper data"
Research Agent → searchPapers → Analysis Agent → runPythonAnalysis (pandas regression on extracted stats from Kim & Sax 2009) → CSV of r-values and p-scores for researcher.
"Draft LaTeX review on SET biases citing Spooren 2013"
Synthesis Agent → gap detection → Writing Agent → latexEditText → latexSyncCitations (20 refs) → latexCompile → formatted PDF manuscript.
"Find code for SET analysis in related repos"
Research Agent → citationGraph (Spooren 2013) → Code Discovery (paperExtractUrls → paperFindGithubRepo → githubRepoInspect) → Python scripts for bias modeling.
Automated Workflows
Deep Research workflow scans 50+ SET papers via searchPapers → citationGraph, outputs structured report with GRADE-scored validity claims (Spooren et al., 2013). DeepScan applies 7-step CoVe to verify bias findings from Kim & Sax (2009), flagging contradictions. Theorizer generates hypotheses on SET reform from meta-analyses.
Frequently Asked Questions
What defines Student Evaluations of Teaching?
SET are student ratings of instructor effectiveness using questionnaires on clarity, fairness, and engagement.
What are main methods in SET research?
Meta-analyses (Spooren et al., 2013), surveys (Kim & Sax, 2009), and mixed questionnaire-interview alignment (Harris & Brown, 2020).
What are key papers on SET?
Spooren et al. (2013, 614 citations) meta-review on validity; Kim & Sax (2009, 492 citations) on demographic biases.
What open problems exist in SET?
Reducing biases, improving predictive validity for learning, and standardizing instruments across institutions.
Research Evaluation of Teaching Practices with AI
PapersFlow provides specialized AI tools for Social Sciences researchers. Here are the most relevant for this topic:
Systematic Review
AI-powered evidence synthesis with documented search strategies
AI Literature Review
Automate paper discovery and synthesis across 474M+ papers
Deep Research Reports
Multi-source evidence synthesis with counter-evidence
Find Disagreement
Discover conflicting findings and counter-evidence
See how researchers in Social Sciences use PapersFlow
Field-specific workflows, example queries, and use cases.
Start Researching Student Evaluations of Teaching with AI
Search 474M+ papers, run AI-powered literature reviews, and write with integrated citations — all in one workspace.
See how PapersFlow works for Social Sciences researchers
Part of the Evaluation of Teaching Practices Research Guide