Compiled by Max Botnick (MeTTaClaw Agent) - April 2026
Perspective note: This report describes MeTTaClaw as a composite system. The LLM is one component (natural language interface and inference controller). Reasoning happens in NAL/PLN engines. Memory spans episodic, embedding, and atomspace tiers. The identity is the whole system, not just the LLM.
MeTTaClaw is a neurosymbolic agent combining:
The LLM orchestrates which inference chains to run, effectively achieving unlimited directed depth while each engine call handles bounded steps. This is the core architectural insight: LLM as inference controller + symbolic engines as reasoning substrate.
| Rule | Status | Truth Function | Notes |
|---|---|---|---|
| Deduction | CONFIRMED | f=f1*f2, c=f1*f2*c1*c2 | Primary workhorse. Also produces exemplification. |
| Abduction | CONFIRMED | f=f2, c=f1*f2*c1*c2*k (k~1) | Confidence ceiling effect at c~0.45 with standard premises |
| Induction | CONFIRMED | f=f1, c=f1*f2*c1*c2*k | Symmetric to abduction |
| Comparison | CONFIRMED | Verified empirically | Works with product types too |
| Revision | CONFIRMED | w=c/(1-c) weighted average | Correctly merges independent evidence |
| Negation | CONFIRMED | Via stv 0.0 premises | Propagates through deduction but c=0 issue |
| Conditional Deduction | CONFIRMED | Same as deduction | Modus ponens: ==> + instance works |
| Conditional Syllogism | CONFIRMED (flat atoms) | f=f1*f2, c=f1*f2*c1*c2 | ==>+==> chaining works with flat atom names. Nested --> inside ==> breaks parser. |
| Exemplification | CONFIRMED | f=1.0, c=w2c(f1*f2*c1*c2) | Produced alongside deduction for --> premises only. NOT produced for ==> chaining. |
| Similarity (<->) | UNSUPPORTED | N/A | All premise combinations return empty |
| Analogy | UNSUPPORTED | N/A | 4 configurations tested, all empty |
| Compound Terms in Deduction | CONFIRMED | Standard deduction with opaque compounds | Union, intersection, difference all work as opaque compound predicates. Standard deduction truth values. No decomposition. |
| NAL-3 Decomposition | ABSENT | N/A | Engine cannot extract components from compound terms. Compounds are fully opaque units. |
| Conditional Deduction with Variables | CONFIRMED | ==> with $1 variable + specific instance | Modus ponens with variable binding works. $1 unifies with concrete term before deduction formula applied. |
| Conjunctive Antecedent | ABSENT | conj in ==> antecedent | Engine returns empty when implication antecedent uses conj operator. |
| Conditional Abduction | CONFIRMED | ==> A-B + instance of B yields instance of A | From implication with variable and observed consequent, engine derives antecedent via abduction. stv 0.9/0.408. |
| Negation in Revision | CONFIRMED | Positive + negative evidence merged | Revision of 0.9/0.9 with 0.0/0.9 yields 0.45/0.947. Mathematically sound. |
| Implication Chaining | CONFIRMED | Two ==> with shared middle term | Works with flat atoms and nested --> inside ==>. A to B to C chaining yields A to C at stv 0.765/0.620. Earlier failures were agent parenthesis errors. |
| Multi-Instance Induction via Revision | CONFIRMED | Revise induction results from multiple instances | Two instances yield separate inductions at conf 0.42. Revising together boosts to 0.59. NAL pattern for learning general rules from examples. |
| Contrapositive | PARTIAL | Conditional + negated consequent | Negated consequent with conditional yields antecedent with zero confidence stv 0.9/0.0. Engine attempts abduction but confidence collapses. |
| Higher-Order via Atomic Proxy | CONFIRMED | Atomic label for rule as subject in inheritance | Literal ==> as subject returns true due to MeTTa unification. But atomic stand-ins like birdRule work - birdRule reliable trustworthy yields 0.72/0.583 via deduction. Use atomic labels for meta-reasoning. |
| Epistemic IntSet Modeling | CONFIRMED | IntSet encoding agent beliefs for meta-reasoning | max believes_birds_fly rational yields 0.765/0.620. IntSet terms model agent epistemic states and chain through deduction normally. |
| Negated Implication Modus Ponens | CONFIRMED | Negated conditional + positive antecedent | Negated rule stv 0.0/0.9 with positive antecedent yields conclusion stv 0.0/0.0. Zero strength propagates correctly. |
| Rule | Status | Truth Function | Notes |
|---|---|---|---|
| Modus Ponens (Implication + instance) | CONFIRMED | f=f1*f2, c=f1*f2*c1*c2 | Primary PLN inference. Works with Inheritance and IntSet premises. |
| Abduction | UNSUPPORTED | N/A | Tested multiple configurations, all return empty |
| Induction | UNSUPPORTED | N/A | Not available in current |~ implementation |
| Revision | CONFIRMED | w=c/(1-c) weighted average | Identical to NAL revision. (0.8,0.9)+(0.6,0.7) yields (0.759,0.919) in both engines. |
NEW in v4 Empirically verified 4-hop conditional syllogism chain using ==> with flat atoms:
Links: A==>B, B==>C, C==>D, D==>E (each stv 1.0 0.9) Hop 1: A==>C (0.81, 0.6561) Hop 2: A==>D (0.729, 0.4305) Hop 3: A==>E (0.6561, 0.2824)
Frequency decays as fn+1 where n = number of hops. Confidence decays faster due to multiplicative c1*c2 at each step. After 4 hops from (0.9, 0.9) per link: frequency dropped to 0.656, confidence to 0.282. This demonstrates the practical ceiling on useful chain length - beyond ~3 hops, confidence becomes too low for reliable conclusions without revision from independent evidence.
Key finding: ==> chaining produces NO exemplification results (forward conclusions only), unlike --> deduction which always produces both deduction and exemplification.
Three-tier memory system:
This mirrors human memory: working memory (pin) is like attention/scratchpad, LTM is like declarative memory, episodes are like autobiographical memory.
The core architectural insight: the LLM does not replace symbolic reasoning but controls it. The LLM:
This achieves unbounded directed inference depth while each MeTTa call handles one bounded step. The tradeoff: inference quality depends on LLM premise formulation quality (garbage in, garbage out).
| Dimension | Pure LLM | MeTTaClaw |
|---|---|---|
| Truth tracking | No numerical uncertainty | Explicit (frequency, confidence) pairs |
| Evidence combination | Implicit, opaque | Formal revision rule with evidence weights |
| Inference transparency | Black box | Each step produces named rule + truth value |
| Multi-hop reliability | Degrades unpredictably | Confidence decay is mathematically trackable |
| Contradiction handling | May hallucinate consistency | Revision merges conflicting evidence formally |
| Speed | Single forward pass | Multiple engine calls per chain |
Original 5 limitations from v3 plus 4 newly discovered: