flowchart LR P[1. Plan<br/>Set goals · Define behaviours] --> A[2. Act<br/>Manager observes · Coaches] A --> R[3. Review<br/>Mid-cycle and final review] R --> RW[4. Reward<br/>Pay · Promotion · Development plan] RW -. feeds back .-> P style P fill:#E3F2FD,stroke:#1565C0 style A fill:#FFF3E0,stroke:#EF6C00 style R fill:#FCE4EC,stroke:#AD1457 style RW fill:#E8F5E9,stroke:#2E7D32
29 Performance Management and Appraisal
29.1 Performance Management vs Performance Appraisal
Two terms are often confused — performance appraisal and performance management. The distinction matters in the NTA stems and in practice.
Performance appraisal is the periodic, formal evaluation of an individual employee’s job performance. Performance management is the broader, continuous process that includes goal-setting, ongoing feedback, development, recognition and reward — of which appraisal is one event.
| Feature | Performance Appraisal | Performance Management |
|---|---|---|
| Frequency | Periodic — typically annual | Continuous |
| Focus | Evaluating past performance | Improving future performance |
| Approach | Backward-looking | Forward-looking |
| Scope | Single event | Linked system: goals → feedback → development → reward |
| Owner | HR-driven | Manager-driven |
| Outcome | Rating | Ratings + plans + development + capability building |
Aswathappa: “Performance appraisal is part of performance management; the two are not synonymous” (aswathappa2020?). Dessler treats performance management as a strategic activity that aligns individual performance with organisational goals (dessler2020?).
29.2 The Performance Management Cycle
Modern performance management runs as a continuous cycle, not an annual event.
The cycle is the textbook PM design — plan, act, review, reward, repeat. Goals set in stage 1 are typically SMART (Specific, Measurable, Achievable, Relevant, Time-bound), drawn from Locke’s goal-setting theory (locke1968?).
29.3 Methods of Performance Appraisal
The discipline distinguishes traditional and modern methods.
| Method | How it works | Strength | Weakness |
|---|---|---|---|
| Ranking | Employees ranked from best to worst | Simple | Not informative for development |
| Paired comparison | Each employee compared with every other | Forces choices | Time-consuming for large groups |
| Forced distribution | Employees forced into a distribution (e.g., 20% top, 70% middle, 10% bottom) | Counters rating inflation | Demoralising; arbitrary in small teams |
| Graphic rating scale | Each trait rated on a numerical scale | Simple | Subjective; halo effect |
| Checklist | Yes/no on a list of behaviours | Structured | Limited insight |
| Critical incidents | Specific behaviours catalogued | Behavioural focus | Coverage uneven |
| Field review | HR specialist interviews supervisors | Reduces line-manager bias | Distant from work |
| Confidential report | Open-ended write-up by supervisor | Detail | Subjective; legacy of Indian government |
| Method | How it works | Originator |
|---|---|---|
| Behaviourally Anchored Rating Scales (BARS) | Each rating point anchored to a specific behaviour | Smith & Kendall (1963) |
| Behavioural Observation Scale (BOS) | Frequency of specific behaviours rated | Latham & Wexley |
| Management by Objectives (MBO) | Mutual goal-setting + periodic review | Peter Drucker (1954) |
| 360-Degree feedback | Multi-source feedback — supervisor, peers, subordinates, customers, self | Various |
| Assessment centre | Battery of exercises rated by trained assessors | British military, post-war |
| Psychological appraisal | Test-based forecast of future potential | I/O psychologists |
| Human Resource Accounting (HRA) | Treats people as assets; values acquisition + investment cost | Likert; Flamholtz |
| Balanced Scorecard | Four perspectives: financial, customer, process, L&G | Kaplan & Norton (1992) |
| OKRs | Quarterly Objectives + Key Results, ambitious and transparent | Andy Grove → John Doerr; popularised by Google |
29.3.1 BARS — the gold-standard rating method
Behaviourally Anchored Rating Scales (BARS) combine the structure of graphic rating scales with the specificity of critical incidents. Each performance dimension has a 5-to-9 point scale where each point is anchored to a concrete behavioural example. BARS reduces rating ambiguity and is widely used in research-grade appraisal.
29.3.2 360-Degree Feedback
A 360 collects feedback from multiple raters — supervisor, peers, subordinates, customers and self — typically against a competency framework. It is more often used for development than for ratings (because peers can be less honest when their feedback affects pay). A common variant — 180-degree feedback — drops some sources; 270 / 540 / 720 are sometimes encountered in older texts.
flowchart LR E[Employee] --- S[Supervisor] E --- P[Peers] E --- SR[Subordinates] E --- C[Customers] E --- SE[Self] style E fill:#FCE4EC,stroke:#AD1457
29.3.3 MBO and OKRs
MBO (Drucker, 1954) — mutually agreed objectives, periodic review, reward linked to objective achievement. OKRs (Andy Grove at Intel; brought to Google by John Doerr) — quarterly Objectives + measurable Key Results, deliberately ambitious (target ~70% achievement), transparent across the firm.
29.3.4 Continuous performance management
The 2010s saw a wave of large firms (Adobe, GE, Microsoft) abolish annual ratings in favour of continuous check-ins, real-time feedback, and quarterly conversations — though most firms have since re-introduced some annual rating element.
29.4 Rating Errors
A perennial NTA topic — the typical cognitive errors that distort appraisals.
| Error | Definition | Remedy |
|---|---|---|
| Halo | One positive trait colours the overall rating | BARS, multi-source |
| Horns | One negative trait dominates | BARS, multi-source |
| Leniency | Rater rates everyone high | Forced distribution; calibration |
| Strictness | Rater rates everyone low | Calibration; rater training |
| Central tendency | Rater clusters everyone in the middle | Force discrimination |
| Recency | Recent events dominate the year’s view | Continuous record-keeping |
| Primacy | First impression dominates | Multiple data points |
| Contrast | Comparison to a recent reference distorts judgment | Standardised criteria |
| Similar-to-me | Bias toward those like the rater | Diverse rater panel |
| Stereotype | Group-based judgment | Awareness training |
| Spillover | Past ratings carry into the current cycle | Anchored evidence |
29.5 Linking Performance to Other HR Systems
A robust performance system feeds:
- Compensation — bonus and increments aligned with rating.
- Career development — IDPs targeted at gaps surfaced in review.
- Succession planning — identifying high-potentials.
- Training — directing investment to where it pays off.
- Termination — defensible documentation when separation is necessary.
A poorly designed system is worse than none — it produces gaming, demotivation, and litigation.
29.6 Practice Questions
Performance management differs from performance appraisal primarily in that:
View solution
In a BARS appraisal system, each rating point on the scale is:
View solution
A 360-degree feedback system collects feedback from:
View solution
OKRs (Objectives and Key Results) — popularised by Google — were originally developed at:
View solution
A manager who rates a recent presenter highly across all dimensions because the presentation went well in the last week of the cycle is showing:
View solution
Management by Objectives (MBO), as a method of appraisal, was popularised by:
View solution
A "forced distribution" appraisal system primarily controls for:
View solution
SMART, the standard test for a usable performance objective, stands for:
View solution
- Performance appraisal = periodic event. Performance management = continuous system: plan → act → review → reward.
- Goals are SMART (Locke). Frequencies: annual, half-yearly, quarterly, continuous.
- Traditional methods: ranking, paired comparison, forced distribution, graphic rating, checklist, critical incidents, field review, confidential report.
- Modern methods: BARS (Smith & Kendall), BOS, MBO (Drucker), 360°, assessment centre, HRA, BSC, OKRs (Grove → Doerr).
- 360° = supervisor + peers + subordinates + customers + self.
- Common rating errors: halo, horns, leniency, strictness, central tendency, recency, primacy, contrast, similar-to-me, stereotype, spillover.
- Modern trend: continuous performance management — replace annual ratings with quarterly check-ins (Adobe, GE, Microsoft).