Skip to content

Peer Evaluation of Teaching

Before doing any evaluation, pause and consider what you mean by “effective teaching,” the standard by which you’ll be evaluating your peers. 

Explore the selected frameworks to the right for examples and inspiration:

One way to monitor the influence of your experiences, assumptions, and biases in evaluating a peer’s teaching is to use tools that focus and guide your observation according to your definition of effective teaching.

The Classroom Observation Process

Peer observations of teaching for review, tenure, and promotion processes typically follow a standard set of good practices divided into three stages: before the observation, during the observation, and after the observation. A key part of that process is considering and mitigating the observer’s potential biases.  

Classroom Observation Instruments

During the observation, the observer can either use an existing, ready-to-use instrument, or customize an observation protocol to match specific aspects of “effective teaching” and types of classes. Resources for both options are below.

Option 1: Use Ready-to-Use Instruments

Below is a selection of existing instruments, ranging from simple to complex, but all effective:

 ​​​​​​Option 2: Customize a Class Observation Protocol

Fair and effective peer observations are guided by decisions about what aspects of teaching will be evaluated, what behaviors will demonstrate those elements, and how to document the observation. Rather than using an existing instrument, you can make your own following the 2 steps below: ​​​​​​​

1. WHAT do you want to evaluate through classroom observation?​​​​​​​  Explore these collections of observable teaching behaviors for different aspects of teaching, compiled from existing instruments:

2. HOW will you focus and document your observation as it happens?​​​​​​​ Below are two ways to provide structure and help prevent bias during your observation by guiding how you record examples of learning or engagement, and how you note analysis and recommendations (Gleason & Sanger):

The teaching narrative or statement is not an intuitive piece of writing. It’s typically written directly in response to the institution’s instructions and expectations. Before you evaluate this document, review this explicit guidance that has been communicated to the candidate. This guidance for the candidate should also be the guidance — the tools — for your evaluation.

Often, candidates are expected to “reflect on their teaching” within this narrative. This typically means the following:

Component of ReflectionHow to Assess
Self-awareness of their own strengths and weaknessesHow does their self-assessment of their strengths and weaknesses align with the other evidence provided? With your definition of effective teaching?
Responsiveness to feedback from students and peersIn their discussions of student evaluations, classroom observations, and previous review letters, does the narrative address specific concerns and recommendations with comprehension, and with EITHER acknowledgement and a plan for improvement OR an effective counter-perspective?
Demonstration of growth over timeHow does the narrative illustrate the candidate has learned something about their teaching? See sample tools for assessing growth or development here.

Request artifacts that will provide evidence of specific aspects of your definition of effective teaching (Step 1 above). Especially when annotated, these artifacts can document key issues like course design, illustrations of student learning expectations, compliance with department/institution guidelines, constructive alignment (learning outcomes-assessments-activities), disciplinarity, varied assessments, approach to student feedback, patterns in student performance, rigor, attention to diversity-equity-inclusion, and more.

Some examples of these materials, along with sample tools for evaluation, are below. As you review any artifacts, ask What does this particular material provide evidence of? (And what does it not provide evidence of?)

Artifacts of Teaching Performance
* syllabi from previous, new, or redesigned courses (sample evaluation tools)
* assignments and accompanying rubrics (sample evaluation tool​​​​​​​)
* instructor’s feedback to students on anonymized samples of student work
* evidence of reasonable time to return graded work
* plans for a class session, lecture notes
* ​​​​​​​instructor’s curated highlights of student evaluations for a particular course, a course type, or a single question on the evaluations across time–with brief annotation about strengths, growth, and/or continued challenges

Artifacts of Student Performance
* anonymized samples of above average, average, & below average student work, possibly with grade distribution of assignment (sample evaluation tool)
* anonymized samples of work from the same student(s) to show evidence of growth (sample evaluation tool)
* grade distribution of all courses taught, or specific course types 

What else? Think about what traits you’re looking for, and where you’d find them.

–> The “Critical Teaching Behaviors” (Barbeau & Happel) instrument has a column indicating where different teaching traits and behaviors are documented.

If you’re looking for tools to focus on evidence of the instructor’s attention to diversity, equity, and inclusion (DEI) across materials, see this collection of DEI evaluation tools. (They’re also included in the relevant samples throughout this document.)

Some institutions invite relevant experts from outside the institution to provide letters of evaluation.

In An Inclusive Academy, Abigail Stewart and Virginia Valian recommend “a high degree of skepticism about the weight [these letters] should have in the evaluation process” (p. 356) and provide guidance on how to request and evaluate them effectively (pp. 355-360; [direct link to pages in Olin Library’s electronic copy]).

In addition to formal end-of-semester student evaluations of teaching administered by the institution, you may consider multiple ways of capturing different moments of students’ experiences and perspectives.  For instance, what would each of the following tell you?  And while some by themselves are certainly insufficient, what do they contribute to a more holistic picture of teaching effectiveness?

  • responses to anonymous mid-semester or end-of-semester surveys tailored to the course (if truly anonymous, can be administered by the colleague or by a third party)
  • notes from current students (solicited or unsolicited)
  • letters from former students about longer-term impact, solicited and collected by a third party
  • teaching awards adjudicated by students

A Caution
There is plenty of evidence challenging the practice of relying too heavily on students’ assessments of teaching quality (e.g., the “Statement on Student Evaluations of Teaching” from the American Sociological Association, Rollins’s white paper on student evaluations, or this review from Auburn University’s Teaching Effectiveness Committee).

While these statements focus on formal student evaluations, the biases that affect students’ perspectives influence informal assessments as well. This is one of the reasons why using multiple measures to evaluate teaching is essential.

This site is licensed under a CC-BY-NC (Creative Commons Attribution-NonCommercial) 4.0 International License.
Nancy Chick, Endeavor Foundation Center for Faculty Development at Rollins College, 2021