Skip to content
360 SELF-EVALUATION

360 self-evaluation: how to run one that produces useful data

The self-evaluation is the riskiest part of a 360. Self-ratings cluster half a level above observed ability on average, and when the self-rater takes the exercise as a chance to defend themselves instead of reflect, the whole report becomes a negotiation. This subtopic covers how to run a self-eval that surfaces honest data and produces a real development conversation.

By Simon CarviPublished April 20267 min read

On this page

What a self-evaluation is inside a 360

The self-evaluation is the assessee rating themselves on the same competencies, at the same proficiency scale, that every other rater will rate them on. In a 360, the self-rating is never the answer; it is one data point triangulated against manager, peer, and direct-report ratings. What the self-eval gives you is not 'how good is this person' but 'where does their self-perception diverge from the observed reality'. That delta is the single most useful signal in the entire report.

For the full methodology behind the 360 cycle, the 360 assessment guide covers the rater stack, report structure, and common failure modes. This page goes deep on the self-rating layer specifically.

The 360 rater stack

Five perspectives on the same competencies. Each group sees behaviours the others cannot.

Self

Own view of capability and intent

1 rater

Manager

Upward behaviours: how the assessee presents and decides

1 rater

Peers

Horizontal behaviours: collaboration, conflict, cross-team work

3 raters minimum

Direct reports

Downward behaviours: management, coaching, delegation

3 raters minimum

External

Cross-org view: customers, partners, cross-functional peers

2-3 raters, optional

Groups below 3 raters merge with peers for anonymity. Same competencies, same scale, across every group.

Why self-ratings overstate by half a level

Self-ratings in a structured 360 cluster about 0.5 levels above the converged observer average on most competencies, and closer to 1 full level above for high-stakes competencies where the assessee feels they should score well (leadership, strategic thinking, communication). This is not dishonesty, it is the Dunning-Kruger effect plus ego protection plus a lack of calibration against a shared behavioural anchor.

  • 1No calibration against anchors Without reading the behavioural anchor for each level, assessees default to a subjective interpretation of 'I am good at this'. Observers rate against the same anchor every time and converge.
  • 2Identity-protective rating When the rating feels like an evaluation of who you are rather than what you do, most people anchor a half-level above neutral to protect identity. This is why pure numeric self-ratings are unreliable.
  • 3Asymmetric observation Peers and direct reports see the assessee across many situations. The assessee only sees their own intent. Intent and behaviour diverge, and the self-rating often reflects intent.
  • 4Over-estimation is the signal When self-rating is consistently half a level above observer average, that is the default and manageable. When it is 1.5 or 2 levels above on a specific competency, that is a blind spot the debrief must address.

How to write self-rating items that work

The self-rating item is the same item every other rater sees. The only difference is the perspective anchor ('I', 'you', 'they'). Three rules separate items that produce signal from items that produce survey noise.

  • 1One behaviour per item 'I present data to executives with a clear recommendation' is one behaviour. 'I communicate strategically' is three behaviours and a value judgment. Short, concrete, observable.
  • 2Every item reads off the same anchor For each proficiency level, the assessee and every rater see the same one-sentence behavioural anchor. This is what calibrates the rating. Hiding the anchor from the assessee is the single most common reason self-ratings inflate.
  • 3Open-text only where it earns its keep Mandatory text on every competency doubles survey time and the last two answers are always thinner than the first. Reserve open-text for the 2 or 3 highest-leverage competencies the debrief will focus on.

Self-eval calibrated against observable anchors

Assessees see the same behavioural anchor every rater sees. Over-estimation and under-estimation flagged automatically in the report blind-spots section.

See how it works

Preparing the assessee (before the self-eval)

The assessee should receive a 1-page brief before opening the self-evaluation. The brief should state the purpose (development, not performance), the anonymity model of the other rater groups, and the expected time to complete (15 to 25 minutes). Without that frame, assessees default to one of two failure modes: rate themselves down to seem humble, or rate themselves up to defend against what the other raters might say.

After the campaign closes, the self-rating feeds directly into the gap analysis that seeds the IDP. The gap is against the role target, not against the self-rating. The self-rating is context that shapes the debrief, not the basis for the development priorities.

Self-evaluation on Huneety

On Huneety, the self-evaluation runs on the same campaign as the other rater groups. The assessee sees the behavioural anchor for each level (calibrated rating), uses the same scale (Dreyfus 1 to 5 by default, or your custom scale), and submits in under 20 minutes. The report flags over-estimation (self > observer by 1.0 or more) and under-estimation (observer > self by 1.0 or more) explicitly in a blind-spots section, so the debrief starts with the right data. See how Huneety runs a 360.

QUICK ANSWERS

Quick answers

Is self-only a valid 360?
No. Self-only is an opinion survey, not a 360. It is useful as a pre-reflection before a real 360, or as a check-in between full cycles, but it should not be used as the basis for development decisions. The value of the 360 is in the triangulation with other rater groups.
Should the assessee see the behavioural anchors during the self-eval?
Yes. Hiding the anchor from the self-rater is the single most common reason self-ratings inflate. Every rater, including self, rates against the same anchor for the same level. That is what makes the scale consistent across groups.
What if the assessee under-rates themselves?
Under-estimation (others > self by 1.0 or more) is common for early-career individual contributors and first-time managers. The debrief should name it explicitly: the person is further along than they think. Under-estimation left unaddressed depresses career ambition and slows internal promotion.

Run a calibrated 360 on Huneety

Assessees, managers, peers, and direct reports rate against the same behavioural anchors. Blind spots flagged automatically. Reports generated in minutes.