Dear Negotiation Explorer,

Welcome to issue 27 of our NegoAI series.

Over the past two weeks, we've covered how to prompt AI for negotiation preparation (the metaprompt in issue 25) and why that approach works (the 5 elements in issue 26).

But there's a step most people skip: checking if the output is actually good.

AI sounds confident. It structures responses cleanly. It gives you what looks like solid advice. But confident delivery isn't the same as quality content.

This week: how to assess AI output before you trust it—and how to make AI improve its own work until it meets your standard.

The Problem

I've seen it repeatedly in courses and workshops:

Someone prompts AI, gets a detailed response, and moves forward. Then in the negotiation, something doesn't hold up:

  • A statistic was hallucinated

  • The advice was generic—could apply to any deal

  • A key stakeholder's interest was missed

  • The "specific talking point" wasn't actually usable

The problem isn't bad prompting. It's accepting output without verification.

And here's the thing: you can make AI verify its own work. You just need to tell it what to check for.

The 4 Factors

Quality AI output for negotiation comes down to four factors:

1. Accuracy (Weight: 30)

Can every claim be verified from the information you provided?

AI can hallucinate facts, statistics, precedents, and even quotes. It presents everything with the same confident tone—whether it's drawing from your context or making things up.

What to check:

  • Numbers and statistics

  • Claims about the counterparty

  • References to past interactions or history

  • Any "typically" or "usually" statements (often signals guessing)

2. Relevance (Weight: 30)

Is this specific to your situation? Or could this advice apply to any negotiation?

Generic advice is the most common AI failure. "Build rapport before negotiating" is true but useless. "Given their Q3 deadline pressure, open with timeline flexibility" is specific and useful.

What to check:

  • Does it reference your specific constraints, goals, and context?

  • Could a competitor get the same advice with different inputs?

  • Is it telling you what you already know?

3. Completeness (Weight: 25)

Are there gaps? What's missing? What didn't the AI consider that you know matters?

AI works with what you give it. It can't know what it doesn't know. But it also sometimes ignores information you did provide, or fails to connect dots that matter.

What to check:

  • Key stakeholders not addressed

  • Risks or downsides not considered

  • Missing alternatives or Plan B

  • Information you provided that wasn't used

4. Actionability (Weight: 15)

Can you actually use this?

Vague recommendations aren't actionable. "Consider the relationship dynamics" tells you nothing. "Open with acknowledgment of the delayed payment before asking for the contract extension" is something you can do.

What to check:

  • Are there specific next steps?

  • Could you act on this tomorrow?

  • Is the language something you'd actually say?

Two Frameworks for Assessment

I'll give you two methods—one quick, one comprehensive. Both follow the same principle: make AI evaluate its own work, then improve it.

Framework 1: Traffic Light (Quick Check)

Use when: Time pressure, lower stakes, or initial screening.

Time required: 30 seconds to 2 minutes per iteration.

The Prompt:

Evaluate your previous output using the Traffic Light method.

Base your evaluation ONLY on the context and information I have provided.
Do not assume facts or context beyond what was given.

For each section of your output, answer these 4 questions:

1. ACCURACY: Can every claim be verified from the information provided?
2. RELEVANCE: Is this specific to my situation, not generic advice?
3. COMPLETENESS: Are there gaps or missing considerations?
4. ACTIONABILITY: Can I use this directly in my negotiation?

Assign:
- GREEN: Yes to all 4
- YELLOW: 1-2 questions answered "No" or "Partially"
- RED: 3+ questions answered "No", or any critical failure on Accuracy

For each section:
1. Answer the 4 questions (Yes / No / Partially)
2. Assign Green, Yellow, or Red
3. If Yellow or Red: explain why in one sentence

After scoring, provide:
- Summary of what needs improvement (if anything is Yellow or Red)
- Specific recommendations to reach all Green

Then STOP and wait for my instruction before making changes.

After AI responds, you decide:

  • "Apply your recommendations and re-evaluate" → AI improves and checks again

  • "Good enough, proceed" → You're done

Threshold: All Green to trust the output.

Framework 2: Weighted Assessment (Comprehensive)

Use when: High-stakes negotiation, important preparation, when you need confidence in the output.

Time required: 5-15 minutes per iteration.

The Prompt:

Evaluate your previous output using these 4 factors.

Base your evaluation ONLY on the context and information I have provided.
Do not assume facts or context beyond what was given.

| Factor | Weight | Score 1-5 |
|--------|--------|-----------|
| Accuracy (verifiable facts, no hallucinations) | 30 | ? |
| Relevance (specific to my situation, not generic) | 30 | ? |
| Completeness (covers what I need, no major gaps) | 25 | ? |
| Actionability (I can use this in my negotiation) | 15 | ? |

1. Score each factor 1-5
2. Calculate weighted total (max 500)
3. For any factor below 4: explain the weakness and specific recommendation

Present:
- Scores and weighted total
- Recommendations for improvement (if below 450)

Then STOP and wait for my instruction before making changes.

Target: 450+ to proceed.

Free Live Session: February 12th

Build Your AI Negotiation System — No Code, No Experience

Thursday, February 12th at 6:00 PM CET | 30 minutes | Free

In 30 minutes, you'll see:

  • How to build your first AI negotiation assistant from zero (no coding needed)

  • How to prepare for a deal in less than 25 minutes

  • The 3 questions to ask before trusting any AI with your negotiation strategy

  • How to spot when AI makes things up—and structure prompts for reliable outputs

The TOAST Method: Loop Until Quality

Both frameworks follow the same loop:

┌─────────────────────────────────────────┐
│  TEST: AI generates output              │
└─────────────────┬───────────────────────┘
                  ↓
┌─────────────────────────────────────────┐
│  OBSERVE: AI evaluates its own work     │
│  (Traffic Light or Weighted Assessment) │
└─────────────────┬───────────────────────┘
                  ↓
┌─────────────────────────────────────────┐
│  ANALYZE: AI identifies weaknesses      │
│  and recommends improvements            │
└─────────────────┬───────────────────────┘
                  ↓
┌─────────────────────────────────────────┐
│  [STOP] — Human reviews and decides     │
│  "Improve" or "Good enough"             │
└─────────────────┬───────────────────────┘
                  ↓
┌─────────────────────────────────────────┐
│  SOLVE: AI applies improvements         │
└─────────────────┬───────────────────────┘
                  ↓
┌─────────────────────────────────────────┐
│  TEST AGAIN: AI re-evaluates            │
│  Loop until threshold met               │
└─────────────────────────────────────────┘

Test → Observe → Analyze → Solve → Test again

The key: you stay in control at every iteration. AI doesn't automatically improve and move on. It stops, shows you the assessment, and waits for your decision.

This matters because:

  • You might see something AI missed

  • You might decide "good enough" before perfection

  • You might want to add context that changes the evaluation

  • You maintain ownership of the final output

When to Use Which

Situation

Framework

Why

Quick prep, time pressure

Traffic Light

Fast feedback loop, sufficient for lower stakes

Important negotiation

Weighted Assessment

Thorough evaluation, catches subtle issues

Initial screening

Traffic Light

Quick filter before investing more time

Final check before meeting

Weighted Assessment

Confidence that nothing was missed

Iterating on a draft

Traffic Light first, Weighted at the end

Efficient improvement cycle

The Complete Workflow

Putting it all together with what we've covered:

  1. Prompt using the metaprompt structure (issue 25)

  2. Build your prompt using the 5 elements (issue 26)

  3. Assess output using Traffic Light or Weighted Assessment (this issue)

  4. Iterate using TOAST until threshold met

  5. Use the verified output in your negotiation

This Week's Exercise (15-20 minutes)

Step 1: Take output from a recent AI negotiation prep (or run the metaprompt from issue 25).

Step 2: Run the Traffic Light prompt. See what comes back Yellow or Red.

Step 3: Tell AI to apply improvements. Re-run the evaluation.

Step 4: Once all Green, run the Weighted Assessment. See if it hits 450.

Step 5: Notice the difference between your initial output and final output.

What's Next

Next week: How to give AI the context it actually needs.

We've covered prompting, the elements that matter, and output assessment. But the quality of everything depends on what you feed in.

I'll share a framework for preparing negotiation context—what to include, what to leave out, and how to structure it so AI can use it effectively.

Prompting is half the job. Verifying is the other half.

Reply

Avatar

or to participate

Keep Reading

No posts found