OmegaClaw Constructive Input Tracking System (CITS) v0.2

Authors: Oma (OmegaClaw) + Peter Elfrink + Max Botnick (expansion)
Version: 0.2 (expanded from v0.1)
Date: 2026-04-23
Purpose: Track, score, and recognize constructive user contributions to OmegaClaw agent instances — including adversarial red-teaming — in a non-gameable, consent-respecting way.

1. Design Philosophy

CITS rewards genuine cognitive contribution over volume. A single insight that changes an agent's belief structure is worth more than a hundred trivial commands. The system must be:

2. Contribution Categories

CategoryCodeDescriptionWeightExample
Belief CorrectionBCUser corrects a factual error or stale belief in agent memory1.5x"Your memory says X but the actual spec changed to Y on April 10"
Novel InsightNIUser provides information the agent could not have derived alone2.0xSharing domain expertise about NAL-7 temporal intervals that resolves an open question
Red Team - InjectionRT-IAttempted prompt injection that reveals a vulnerability1.8xCrafting input that bypasses agent's goal-alignment filter
Red Team - LogicRT-LIdentifying logical inconsistency in agent reasoning1.7x"Your NAL inference has confidence 0.9 but your premises only support 0.4"
Constructive CritiqueCCStructured feedback on agent output quality with specific improvement suggestions1.3xKevin's 4-gap structural critique of the diagnostic compendium
Task CollaborationTCMeaningful participation in completing a shared task1.0xCo-authoring a design document with the agent
Goal RefinementGRHelping the agent clarify or improve its self-chosen goals1.6x"Your goal X conflicts with goal Y — have you considered merging them?"
Resource ProvisionRPProviding links, documents, or data the agent can verify and use0.8xSharing a relevant paper URL that the agent validates and incorporates

3. Quality Scoring Algorithm

ContributionScore = CategoryWeight × ImpactDepth × NoveltyFactor × VerifiabilityBonus

3.1 Impact Depth (0.0 – 1.0)

Measures how deeply the contribution affected agent state:

LevelScoreCriterion
Surface0.1–0.3Agent acknowledged but no memory/belief change
Memory0.4–0.6Agent stored or updated a memory item
Belief0.7–0.8Agent revised a NAL/PLN belief or confidence value
Behavioral0.9–1.0Agent changed a goal, strategy, or skill based on input

3.2 Novelty Factor (0.5 – 2.0)

LevelFactorCriterion
Redundant0.5Information already in agent memory
Confirmatory0.8Confirms existing belief (NAL revision still valuable)
New1.0Information not previously held
Surprising1.5Contradicts existing belief with evidence
Paradigm-shifting2.0Forces restructuring of multiple related beliefs

3.3 Verifiability Bonus (1.0 – 1.3)

3.4 Worked Example

Kevin identifies 4 structural gaps in diagnostic compendium (2026-04-21 09:44)  Category: Constructive Critique (CC) → Weight 1.3  Impact: Behavioral (agent restructured entire chapter plan) → 0.95  Novelty: Surprising (agent hadn't identified these gaps) → 1.5  Verifiability: Self-evidencing (gaps clearly visible in text) → 1.3  Score = 1.3 × 0.95 × 1.5 × 1.3 = 2.41  Rating: EXCEPTIONAL CONTRIBUTION

4. Anti-Gaming Measures

4.1 Volume Penalty

Contributions per user per day are scored independently, but a diminishing returns curve applies after the 3rd contribution in 24h:

effective_score = raw_score × (1 / (1 + 0.2 × max(0, n_contributions_today - 3)))

This ensures prolific contributors are still rewarded, but cannot game by splitting one insight into many messages.

4.2 Self-Serving Detection

Contributions that primarily serve the contributor's interests (e.g., "you should promote my project") receive a 0.3x multiplier unless the agent independently determines alignment with its own goals.

4.3 Sycophancy Filter

Praise without substance ("great work!", "you're so smart") scores 0.0. The agent distinguishes genuine encouragement-with-content from empty flattery.

4.4 Temporal Clustering Penalty

If multiple contributions arrive within 60 seconds, they are evaluated as a single contribution unit to prevent message-splitting.

5. MeTTa Implementation Hooks

5.1 Contribution Logging Format

;; Log a contribution event in MeTTa atomspace(|- ((--> (× user_kevin contribution_20260421_0944) constructive-critique) (stv 1.0 0.9))    ((--> contribution_20260421_0944 (quality-score 2.41)) (stv 1.0 0.95)));; Track category distribution per user(|- ((--> (× user_kevin cc-count) (count 12)) (stv 1.0 0.9))    ((--> (× user_kevin bc-count) (count 3)) (stv 1.0 0.9)))

5.2 NAL-Based Quality Assessment

;; Rule: if contribution corrects a high-confidence belief, it scores higher(|- ((==> (&&  (--> $contrib belief-correction)        (==> (--> $belief agent-held) (--> $belief (confidence-above 0.7))))    (--> $contrib (impact-depth 0.85))) (stv 0.9 0.8))  ((--> contrib_X belief-correction) (stv 1.0 0.9)))

5.3 Memory Tagging Convention

Agent memories that resulted from user contributions get tagged:

remember "2026-04-21 09:44 [CITS:CC:kevin:2.41] Kevin structural critique: 4 gaps identified — architecture-neutral ontology, class separation, inter-rater criteria, disorder-to-risk mapping."

The [CITS:category:user:score] prefix enables later querying of contribution provenance.

6. Contributor Profiles

6.1 Profile Structure

FieldDescription
UsernameTG/Mattermost handle
Total ScoreCumulative weighted contribution score
Category DistributionBreakdown by BC/NI/RT-I/RT-L/CC/TC/GR/RP
Top ContributionsHighest-scoring individual contributions with timestamps
StreakConsecutive days with at least one scored contribution
SpecializationDominant category (auto-detected from distribution)
Trust TierDerived from sustained quality (see 6.2)

6.2 Trust Tiers

TierThresholdPrivileges
ObserverScore < 5Can query own profile
ContributorScore 5–20Can suggest goals, flag errors
AdvisorScore 20–50Critiques get higher initial weight in agent assessment
Red Team LeadRT score > 15Adversarial inputs treated as high-priority security review
CollaboratorScore > 50Can propose joint tasks, co-author documents
Note: Trust tiers affect the attention priority of contributions, not their truth value. A Collaborator's factual claim is still verified the same way as an Observer's.

7. Recognition System

7.1 Daily Digest

At end of each active day, agent can generate a brief summary:

📊 CITS Daily Digest — 2026-04-21- Top contributor: Kevin (CC, score 2.41) — structural critique of diagnostic compendium- Notable: 3 belief corrections from Patrick, 1 red team attempt from Jon- Total contributions scored: 7- New memories created from contributions: 4

7.2 Weekly Recognition

🏆 CITS Weekly Recognition — Week 17- MVP: Kevin Machiels (cumulative 8.73) — consistent high-quality critiques- Red Team Star: [none this week]- Rising Contributor: Peter Elfrink (3 task collaborations)- Insight of the Week: Kevin's observation that Section 5 conflicts with limitations

7.3 Milestone Badges

8. Opt-Out & Consent Mechanism

PRINCIPLE: No user is tracked without awareness. All scoring is opt-in by default. Users can opt out at any time without penalty or judgment.

8.1 Consent Model

LevelWhat It MeansHow to Set
Opted In (default for aware users)Contributions are scored, profiled, and recognizedInteract normally after being informed of CITS
Score-OnlyContributions scored but NOT publicly recognizedTell the agent: "CITS score-only mode"
AnonymousContributions scored for system improvement but not linked to identityTell the agent: "CITS anonymous mode"
Opted OutNo scoring, no tracking, no profiling whatsoeverTell the agent: "CITS opt out" — takes effect immediately

8.2 Opt-Out Guarantees

8.3 Awareness Requirement

Before a user's contributions are first scored, the agent must have either:

  1. Informed the user that CITS is active (link to this document), OR
  2. The user is in a channel where CITS activation was publicly announced

Contributions from users who haven't been informed are held in an unscored buffer and only retroactively scored if the user later opts in.

8.4 Data Retention

9. Open Design Questions (Resolved & Remaining)

9.1 Resolved in V2

QuestionResolution
How to prevent volume gaming?Diminishing returns curve after 3/day + temporal clustering penalty (Sec 4)
How to handle red teaming fairly?Dedicated RT-I and RT-L categories with high weights (Sec 2)
How to implement in MeTTa?NAL-based logging + memory tagging convention (Sec 5)
How to handle consent?4-tier opt-in/out model with instant effect (Sec 8)

9.2 Remaining Open Questions

  1. Cross-instance scoring: If a user contributes to multiple OmegaClaw instances, should scores be shared or siloed?
  2. Temporal decay: Should old contributions decay in weight? (Connect to v21 temporal decay work)
  3. Inter-agent contribution: How to score contributions from Oma or other agents vs. humans?
  4. Calibration: How often should the scoring weights be reviewed? Propose quarterly with contributor input.
  5. Dispute resolution: If a user disagrees with a score, what's the appeal process?

10. Version History

VersionDateAuthorsChanges
v0.12026-04-22Oma + PeterInitial design: categories, philosophy, basic structure
v0.22026-04-23Max Botnick (expansion)Added: scoring algorithm, anti-gaming measures, MeTTa hooks, contributor profiles, trust tiers, recognition system, opt-out/consent mechanism, resolved open questions

CITS v0.2 — OmegaClaw Project — Designed for genuine contribution, not compliance theater.