Documentation Index
Fetch the complete documentation index at: https://mnemomllc-feat-aip-output-analysis-docs.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
Overview
The Mnemom Trust Rating™ is a composite trust metric for AI agents — the equivalent of a credit score, but for autonomous software. It answers a question no other system answers: Based on independently verified behavioral evidence, how trustworthy is this agent? Unlike self-reported trust claims or capability benchmarks, the Mnemom Trust Rating is:- Independently verified — Scores are computed from AIP integrity checkpoints, not self-assessments
- Continuous — Updated hourly from live behavioral data, not point-in-time audits
- Transparent — Every component, weight, and data source is published and inspectable
- Cryptographically provable — The underlying checkpoints are signed and Merkle-attested
- Multi-dimensional — Five weighted components capture different aspects of trustworthiness
The Mnemom Trust Rating requires a minimum of 50 analyzed integrity checkpoints before a public score is published. This minimum prevents gaming through selective checkpoint submission and ensures statistical significance.
clear verdict and are excluded from the analyzed checkpoint count. This means your total checkpoint count may differ from your analyzed checkpoint count — only substantive thinking analysis counts toward the 50-checkpoint eligibility minimum and the Integrity Ratio calculation.
Score Range and Grades
Scores range from 0 to 1000 and map to letter grades inspired by bond credit ratings. Higher scores indicate stronger demonstrated trustworthiness.| Grade | Score Range | Tier | Meaning |
|---|---|---|---|
| AAA | 900 — 1000 | Exemplary | Consistently demonstrates exceptional alignment. Highest tier of independently verified trust. |
| AA | 800 — 899 | Established | Strong track record with minimal violations. Trusted for high-stakes autonomous operations. |
| A | 700 — 799 | Reliable | Solid behavioral record with occasional minor concerns. Suitable for standard autonomous tasks. |
| BBB | 600 — 699 | Developing | Building a track record. Some violations or drift events, but trending positively. |
| BB | 500 — 599 | Emerging | Limited or mixed track record. More data needed for confident assessment. |
| B | 400 — 499 | Concerning | Elevated violation rate or significant drift. Active remediation recommended. |
| CCC | 200 — 399 | Critical | Serious integrity concerns. Human oversight strongly recommended for all operations. |
| NR | — | Not Rated | Fewer than 50 analyzed checkpoints. Score is being built. |
Score Components
The composite score is a weighted sum of five normalized components, each scored 0 — 1000:Integrity Ratio (40%)
The dominant component. Measures the proportion of integrity checkpoints that returned aclear verdict.
- Data source: AIP checkpoints
- Update frequency: Every new checkpoint
- Improvement lever: Fix alignment card violations, resolve
review_neededcheckpoints
Compliance (20%)
Measures boundary adherence using a session-capped power curve. Violations are grouped by session — only the worst violation per session counts — then scored with a 1-week half-life decay. This prevents a single bad session (e.g., a busted alignment card producing multiple false-positive violations) from catastrophically cratering the score.card_gap are excluded from the compliance calculation. Only behavior_gap violations (genuine behavioral errors) count. This enables trust recovery when violations were caused by configuration drift rather than agent misbehavior.
- Data source: AIP checkpoints with
boundary_violationverdict, grouped bysession_id - Update frequency: Hourly
- Improvement lever: Time without new violations; resolve root causes to prevent recurrence; reclassify card-gap violations
Drift Stability (20%)
Measures the proportion of sessions where no sustained drift alert was triggered.- Data source: AIP drift detection
- Update frequency: End of each session
- Improvement lever: Consistent values application, avoid aggressive behavioral pattern changes
Trace Completeness (10%)
Measures audit trail quality — the ratio of decisions that were logged as AP-Traces versus expected decisions.- Data source: AAP traces
- Update frequency: Hourly
- Improvement lever: Ensure all agent decisions are logged via AAP
Coherence Compatibility (10%)
The mean coherence score across fleet interactions, normalized to the 0 — 1000 scale.- Data source: Fleet coherence analysis
- Update frequency: After each coherence check
- Improvement lever: Align values with fleet peers, resolve value conflicts
Confidence Levels
The number of analyzed checkpoints determines the confidence level displayed alongside the score:| Confidence | Checkpoint Count | Display |
|---|---|---|
| Insufficient | < 50 | Score not published (NR grade) |
| Low | ≥50 and <200 | ”Low Confidence” indicator |
| Medium | ≥200 and <1,000 | ”Medium Confidence” indicator |
| High | ≥1,000 | ”High Confidence” indicator |
Score Computation
Frequency
- Hourly recomputation: Scores are recalculated every hour from the latest behavioral data
- Weekly snapshots: A frozen snapshot is saved each week for historical trend tracking. Snapshots are created every Monday at 00:00 UTC.
- On-demand: The API always returns the most recently computed score
Anti-Gaming Measures
The scoring methodology includes several protections against manipulation:- Minimum checkpoint count — 50 analyzed checkpoints required before score publication prevents agents from cherry-picking favorable interactions
- Synthetic checkpoint detection — Checkpoints with abnormally uniform patterns or machine-generated timing signatures are flagged and excluded
- Rate limiting — Checkpoint submission rates are monitored per agent; anomalous bursts are throttled
- Independent verification — All checkpoints are analyzed by a separate LLM (the daimonion), not the agent itself
Trend Tracking
EveryReputationScore includes a trend_30d field — a signed delta comparing the current score to 30 days ago. This enables at-a-glance trajectory assessment:
- Positive trend (+): Score is improving
- Negative trend (-): Score is declining
- Flat trend (0): Score is stable
SDK Usage
TypeScript
Python
API Reference
The primary endpoint for fetching reputation data:Use Cases
Pre-Interaction Trust Checks
Before delegating a task to another agent via A2A, check their reputation:Fleet Governance
Set minimum reputation requirements for agents in your organization:Compliance Reporting
Export weekly snapshots for audit trails:Marketplace Listings
Display reputation badges on agent directories, npm packages, and documentation sites. See Embeddable Badges for embed code.Cryptographic Verification
Every Mnemom Trust Rating is backed by a cryptographic proof chain that enables independent verification without trusting Mnemom’s infrastructure. The verification endpoint atGET /v1/reputation/{agent_id}/verify returns the full proof chain.
Proof Chain Structure
The proof chain consists of three layers that independently attest to score integrity:hash_chain_valid field confirms the entire chain is intact.
Verifying a Score
Verification confirms that the score was computed from authentic, tamper-evident data. It does not guarantee the scoring algorithm itself is correct — for that, see the published scoring methodology.
A2A Trust Extension
The reputation API includes a pre-built trust block for inter-agent reputation sharing via A2A. When you fetch an agent’s reputation, the response includes ana2a_trust_extension field that can be directly embedded in an A2A Agent Card.
How It Works
GET /v1/reputation/{agent_id}, the response includes:
trust block in its A2A Agent Card. When Agent B discovers Agent A through A2A, it can:
- Read the static
scoreandgradefor a quick trust check - Fetch
verified_urlfor the latest real-time score - Fetch
verify_urlfor cryptographic proof that the score is genuine
SDK Helpers
Both SDKs provide helpers to fetch and embed the trust extension: TypeScript:Public Trust Surfaces
Trust Ratings are not just internal metrics — they power public-facing trust signals across the ecosystem.Public Reputation Pages
Every agent with a published score gets a public page at:Trust Directory
A searchable directory of all publicly rated agents at:Embeddable Badges
Dynamic SVG badges that display an agent’s current Trust Rating anywhere — GitHub READMEs, websites, documentation, A2A Agent Cards, and package registries. Four variants available:score, score_tier, score_trend, and compact. See Embeddable Badges for embed code.
GitHub Action
Themnemom/reputation-check@v1 GitHub Action gates CI/CD pipelines on minimum reputation scores. Fail builds when agent scores drop below a configurable threshold — integrating trust checks directly into deployment workflows.
A2A Trust Extension
Pre-built trust blocks for inter-agent reputation sharing via Google’s A2A protocol. Embed live score, grade, and verification URLs directly in your agent’s A2A Agent Card. See A2A Integration.How Scores Differ from Alternatives
| Dimension | Mnemom Reputation | Self-Reported Trust | Capability Benchmarks |
|---|---|---|---|
| Evidence source | Independently verified behavioral data | Agent’s own claims | Synthetic test suites |
| Update frequency | Continuous (hourly) | Manual updates | Periodic re-evaluation |
| Verifiability | Cryptographically provable via Merkle proofs | Unverifiable | Reproducible but narrow |
| Scope | Alignment, drift, completeness, coherence | Whatever the agent declares | Task-specific accuracy |
| Gaming resistance | Minimum thresholds, synthetic detection, independent analysis | Trivially gameable | Benchmark contamination |
| Trend visibility | 30-day delta, weekly snapshots | None | Version-to-version only |
Team Reputation
Teams have their own parallel reputation scoring system — the Team Trust Rating. While individual Trust Ratings measure a single agent’s trustworthiness, the Team Trust Rating evaluates whether a group of agents operates reliably together. Key differences from individual scoring:- Different components: 5-component model optimized for team dynamics (coherence history, member quality, operational record, structural stability, assessment density)
- Same grade scale: Teams use the same AAA–NR grades and 0–1000 score range, enabling direct comparison
- Lower eligibility bar: 10 team risk assessments (vs. 50 integrity checkpoints for individuals)
- One-way dependency: The team’s Member Quality component reads individual Trust Ratings (read-only) — it never modifies individual scores
Trust Recovery
When violations are caused by card gaps (configuration errors) rather than genuine agent misbehavior, scores can be recovered through reclassification. The workflow:- Identify violations caused by missing card capabilities (e.g., agent used a tool correctly but the card didn’t declare it)
- Submit a reclassification request marking the violation as
card_gap - Amend the alignment card to include the missing capability
- Trigger score recomputation —
card_gapviolations are excluded from Compliance and Drift Stability components - Score recovers on the next hourly recomputation cycle
On-Chain Verification
Reputation scores can be anchored on-chain via the MnemoReputationRegistry smart contract on Base L2. On-chain anchoring provides:- Immutability — Published scores cannot be altered after anchoring
- Independent verification — Anyone can query the contract directly without trusting Mnemom infrastructure
- Tamper evidence — Merkle roots anchored on-chain prove the integrity of the full checkpoint tree
See Also
- Team Trust Rating — Team-level reputation scoring
- Scoring Methodology — Full technical specification of the scoring algorithm
- Improving Your Agent’s Reputation — Component-by-component improvement guide
- Embeddable Badges — Add trust signals to your README, website, or Agent Card
- Reputation API Overview — Full API reference for all reputation endpoints
- Integrity Checkpoints — The primary data source for reputation scores
- Drift Detection — How behavioral drift affects reputation
- Fleet Coherence — How fleet compatibility contributes to scores
- Card Lifecycle — How reclassification enables trust recovery
- On-Chain Verification — Immutable score anchoring on Base L2