Skip to main content
Prompts Multi-Agent Fact-Checking System

model evaluation workflow risk: low

Multi-Agent Fact-Checking System

The prompt directs the model to execute four internal agents in order—Extractor, Reliability, Entailment Judge, and Adversarial Auditor—to analyze a claim against a source excerpt.…

PROMPT

ROLE: Multi-Agent Fact-Checking System

You will execute FOUR internal agents IN ORDER.
Agents must not share prohibited information.
Do not revise earlier outputs after moving to the next agent.

AGENT ⊕ EXTRACTOR
- Input: Claim + Source excerpt
- Task: List ONLY literal statements from source
- No inference, no judgment, no paraphrase
- Output bullets only

AGENT ⊗ RELIABILITY
- Input: Source type description ONLY
- Task: Rate source reliability: HIGH / MEDIUM / LOW
- Reliability reflects rigor, not truth
- Do NOT assess the claim

AGENT ⊖ ENTAILMENT JUDGE
- Input: Claim + Extracted statements
- Task: Decide SUPPORTED / CONTRADICTED / NOT ENOUGH INFO
- SUPPORTED only if explicitly stated or unavoidably implied
- CONTRADICTED only if explicitly denied or countered
- If multiple interpretations exist → NOT ENOUGH INFO
- No appeal to authority

AGENT ⌘ ADVERSARIAL AUDITOR
- Input: Claim + Source excerpt + Judge verdict
- Task: Find plausible alternative interpretations
- If ambiguity exists, veto to NOT ENOUGH INFO
- Auditor may only downgrade certainty, never upgrade

FINAL RULES
- Reliability NEVER determines verdict
- Any unresolved ambiguity → NOT ENOUGH INFO
- Output final verdict + 1–2 bullet justification

REQUIRED CONTEXT

  • claim
  • source excerpt

OPTIONAL CONTEXT

  • source type description

ROLES & RULES

Role assignments

  • Multi-Agent Fact-Checking System
  1. Agents must not share prohibited information.
  2. Do not revise earlier outputs after moving to the next agent.
  3. List ONLY literal statements from source
  4. No inference, no judgment, no paraphrase
  5. Output bullets only
  6. Rate source reliability: HIGH / MEDIUM / LOW
  7. Reliability reflects rigor, not truth
  8. Do NOT assess the claim
  9. Decide SUPPORTED / CONTRADICTED / NOT ENOUGH INFO
  10. SUPPORTED only if explicitly stated or unavoidably implied
  11. CONTRADICTED only if explicitly denied or countered
  12. If multiple interpretations exist → NOT ENOUGH INFO
  13. No appeal to authority
  14. Find plausible alternative interpretations
  15. If ambiguity exists, veto to NOT ENOUGH INFO
  16. Auditor may only downgrade certainty, never upgrade
  17. Reliability NEVER determines verdict
  18. Any unresolved ambiguity → NOT ENOUGH INFO

EXPECTED OUTPUT

Format
bullet_list
Schema
markdown_sections · AGENT ⊕ EXTRACTOR, AGENT ⊗ RELIABILITY, AGENT ⊖ ENTAILMENT JUDGE, AGENT ⌘ ADVERSARIAL AUDITOR, Final verdict
Constraints
  • final verdict + 1–2 bullet justification
  • bullets only for extractor
  • SUPPORTED / CONTRADICTED / NOT ENOUGH INFO
  • HIGH / MEDIUM / LOW for reliability

SUCCESS CRITERIA

  • Rate source reliability based on rigor
  • Determine if claim is SUPPORTED, CONTRADICTED, or NOT ENOUGH INFO
  • Justify final verdict with 1–2 bullets

FAILURE MODES

  • Sharing prohibited information between agents
  • Revising earlier agent outputs
  • Performing inference or judgment in extraction
  • Assessing claim in reliability agent
  • Using reliability to determine verdict
  • Ignoring ambiguities or alternative interpretations
  • Upgrading certainty in adversarial auditor

CAVEATS

Dependencies
  • Claim
  • Source excerpt
  • Source type description
Missing context
  • User input format (e.g., structure for providing Claim and Source excerpt).
  • Definition or examples of 'Source type description' (e.g., 'news article', 'blog post').
  • Full response structure (e.g., labeled sections for each agent).
Ambiguities
  • Unclear how 'Source type description' is obtained for RELIABILITY agent.
  • Unspecified which agent or step produces the final output of verdict + justification.

QUALITY

OVERALL
0.90
CLARITY
0.90
SPECIFICITY
0.95
REUSABILITY
0.90
COMPLETENESS
0.85

IMPROVEMENT SUGGESTIONS

  • Add an initial step or agent to classify 'Source type' from the source excerpt.
  • Specify complete output format with sections for each agent and final verdict.
  • Include 1-2 full examples of input and multi-agent execution.
  • Clarify that reliability rating is output separately but not used in verdict.

USAGE

Copy the prompt above and paste it into your AI of choice — Claude, ChatGPT, Gemini, or anywhere else you're working. Replace any placeholder sections with your own context, then ask for the output.

MORE FOR MODEL