← Vault Index
Source: frameworks/kit-interview-scorecard-design/07-process-agent.md

Process Agent — AI-Assisted Scorecard Workflow

What This File Covers

This file defines where and how AI tools can be used in the scorecard design and evaluation process. It is a supplement to the consultant process — not a replacement. Every AI-generated output requires practitioner review before it becomes a deliverable.

The agent process has two phases:

  1. Design phase — AI assists in building the scorecard (generating questions, structuring templates, developing focus area descriptions)
  2. Evaluation phase — AI assists in processing completed scorecards (aggregating scores, summarizing debriefs, producing candidate write-ups)

Design Phase — Building the Scorecard

Generating Behavior-Based Questions

When to use AI: After focus areas have been defined and described by the practitioner. AI generates candidate questions for the practitioner to review and select from.

What to provide:

Expected output: 6-10 candidate questions per focus area. The practitioner selects 3-5 and may edit them.

Quality check: Every question must be reviewed by the practitioner for:

Common AI failure modes:

Structuring the Scorecard Template

When to use AI: After all design decisions have been made (focus areas, assignments, scoring scale, questions). AI assembles the document.

What to provide:

Expected output: A formatted scorecard document ready for practitioner review.

Quality check: Verify all content is present, correctly assigned, and properly formatted. Run Gate 2 QC from 04-quality.md.

Developing Focus Area Descriptions

When to use AI: After focus areas have been named and the practitioner has confirmed what each one covers. AI drafts the "what good looks like" and "what risk looks like" descriptions.

What to provide:

Expected output: Draft descriptions for practitioner review.

Quality check: Descriptions must be specific to the role, not generic competency language. "Demonstrates strategic thinking" is not a description. "Articulates a coherent organizational strategy informed by market dynamics, financial constraints, and stakeholder priorities" is a description.


Evaluation Phase — Processing Completed Scorecards

Pre-Debrief Score Aggregation

When to use AI: After all interviewers have submitted scorecards and before the debrief. AI aggregates the scores and produces a facilitator summary.

What to provide:

Expected output: A summary showing:

Quality check: The facilitator reviews the summary for accuracy. Verify that the aggregation doesn't flatten important nuance — two interviewers giving a 3 for different reasons is not the same as agreement.

Critical rule: The summary goes to the facilitator only. It is not shared with the interview team before the debrief. It is a facilitation tool, not a pre-debrief anchor.

Debrief Transcript Summary

When to use AI: After the debrief is recorded and transcribed. AI produces a structured summary for distribution to decision makers who weren't in the debrief, or as a record of the discussion.

What to provide:

Expected output: A structured summary containing:

Quality check: The practitioner reviews the summary for:

Candidate Write-Up from Interview Transcript

When to use AI: After an interview is recorded and transcribed. AI produces a structured write-up for the client using the practitioner's template format.

What to provide:

Expected output: A structured write-up containing:

Quality check: The practitioner reviews for:

Critical distinction: The write-up presents the candidate fairly. It does not sell them. The practitioner works on behalf of the client, not the candidate. This is a fundamental difference from executive search firms where the recruiter's compensation increases with the candidate's salary. The write-up should help the client make an informed decision — not persuade them to advance a candidate.


AI Tool Selection

This kit does not prescribe a specific AI tool. The practitioner uses whatever tool they are comfortable with and that produces acceptable quality output. Common options include Claude, ChatGPT, and Gemini.

Tool selection considerations:


What AI Does Not Do

Make evaluation decisions. AI can aggregate scores but cannot recommend whether a candidate should advance.

Replace the practitioner's judgment on focus areas. AI can generate focus area options, but determining what matters for a specific role in a specific organization requires human judgment informed by organizational context.

Facilitate the debrief. The debrief is a live conversation requiring interpersonal skill, the ability to read the room, and the judgment to challenge vague assessments. AI produces the inputs; the practitioner runs the discussion.

Guarantee defensibility. AI can help produce consistent documentation, but defensibility depends on the process being followed consistently by humans. A well-documented scorecard that wasn't used consistently is not defensible.

Serve as the final quality gate. Every AI output is a draft. The practitioner converts drafts into deliverables through review, editing, and professional judgment. Gate 2 QC is always performed by a human.