Post
4575
π Introducing MARL β Runtime Middleware That Reduces LLM Hallucination Without Fine-Tuning
Now available on PyPI Β· GitHub Β· ClawHub Β· HuggingFace
AI models sense they could be wrong, but they can't actually fix what's broken.
π€ Live A/B test: VIDraft/MARL
We evaluated 9 SOTA models (GPT-5.2, Claude Opus 4.6, Gemini 3 Pro, etc.) across 1,800 assessments in FINAL Bench and found a 39.2%p gap between "recognizing potential errors (MA=0.694)" and "actually finding and fixing them (ER=0.302)."
MARL (Model-Agnostic Runtime Middleware for LLMs) was built to close this metacognitive gap. It decomposes a single LLM call into a 5-stage expert pipeline (Hypothesis β Solver β Auditor β Adversarial Verifier β Synthesizer), transforming "answer in one shot" into "think, doubt, correct, and rewrite."
No weight modification β works instantly with GPT-5.4, Claude, Gemini, Llama, or any OpenAI API-compatible LLM by changing one line: base_url. Ships with 9 domain-specific emergence engines (invention, pharma, genomics, chemistry, ecology, law, and more β 5,538 expert data items) activated by a simple tag like model="gpt-5.4::pharma".
pip install marl-middleware
MARL is also officially registered on ClawHub, the skill marketplace of OpenClaw β an AI agent platform with 260K+ developers and 3,200+ skills. It's the first middleware in the Reasoning Enhancement category. One command β clawhub install marl-middleware β gives your AI agent a metacognition upgrade.
π Technical deep dive: https://huggingface.co/blog/FINAL-Bench/marl-middleware
π¦ PyPI: https://pypi.org/project/marl-middleware/
π GitHub: https://github.com/Vidraft/MARL
π¦ ClawHub: https://clawhub.ai/Cutechicken99/marl-middleware
#MARL #LLM #Hallucination #Metacognition #MultiAgent #AIMiddleware #FINALBench #OpenClaw #ClawHub #PyPI #AGI #HuggingFace #ReasoningAI #SelfCorrection #GlassBoxAI
Now available on PyPI Β· GitHub Β· ClawHub Β· HuggingFace
AI models sense they could be wrong, but they can't actually fix what's broken.
π€ Live A/B test: VIDraft/MARL
We evaluated 9 SOTA models (GPT-5.2, Claude Opus 4.6, Gemini 3 Pro, etc.) across 1,800 assessments in FINAL Bench and found a 39.2%p gap between "recognizing potential errors (MA=0.694)" and "actually finding and fixing them (ER=0.302)."
MARL (Model-Agnostic Runtime Middleware for LLMs) was built to close this metacognitive gap. It decomposes a single LLM call into a 5-stage expert pipeline (Hypothesis β Solver β Auditor β Adversarial Verifier β Synthesizer), transforming "answer in one shot" into "think, doubt, correct, and rewrite."
No weight modification β works instantly with GPT-5.4, Claude, Gemini, Llama, or any OpenAI API-compatible LLM by changing one line: base_url. Ships with 9 domain-specific emergence engines (invention, pharma, genomics, chemistry, ecology, law, and more β 5,538 expert data items) activated by a simple tag like model="gpt-5.4::pharma".
pip install marl-middleware
MARL is also officially registered on ClawHub, the skill marketplace of OpenClaw β an AI agent platform with 260K+ developers and 3,200+ skills. It's the first middleware in the Reasoning Enhancement category. One command β clawhub install marl-middleware β gives your AI agent a metacognition upgrade.
π Technical deep dive: https://huggingface.co/blog/FINAL-Bench/marl-middleware
π¦ PyPI: https://pypi.org/project/marl-middleware/
π GitHub: https://github.com/Vidraft/MARL
π¦ ClawHub: https://clawhub.ai/Cutechicken99/marl-middleware
#MARL #LLM #Hallucination #Metacognition #MultiAgent #AIMiddleware #FINALBench #OpenClaw #ClawHub #PyPI #AGI #HuggingFace #ReasoningAI #SelfCorrection #GlassBoxAI