Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

SeaWolf-AIย 
posted an update 1 day ago
view post
Post
6573
๐Ÿš€ Introducing MARL โ€” Runtime Middleware That Reduces LLM Hallucination Without Fine-Tuning

Now available on PyPI ยท GitHub ยท ClawHub ยท HuggingFace
AI models sense they could be wrong, but they can't actually fix what's broken.

๐Ÿค— Live A/B test: VIDraft/MARL

We evaluated 9 SOTA models (GPT-5.2, Claude Opus 4.6, Gemini 3 Pro, etc.) across 1,800 assessments in FINAL Bench and found a 39.2%p gap between "recognizing potential errors (MA=0.694)" and "actually finding and fixing them (ER=0.302)."

MARL (Model-Agnostic Runtime Middleware for LLMs) was built to close this metacognitive gap. It decomposes a single LLM call into a 5-stage expert pipeline (Hypothesis โ†’ Solver โ†’ Auditor โ†’ Adversarial Verifier โ†’ Synthesizer), transforming "answer in one shot" into "think, doubt, correct, and rewrite."

No weight modification โ€” works instantly with GPT-5.4, Claude, Gemini, Llama, or any OpenAI API-compatible LLM by changing one line: base_url. Ships with 9 domain-specific emergence engines (invention, pharma, genomics, chemistry, ecology, law, and more โ€” 5,538 expert data items) activated by a simple tag like model="gpt-5.4::pharma".

pip install marl-middleware

MARL is also officially registered on ClawHub, the skill marketplace of OpenClaw โ€” an AI agent platform with 260K+ developers and 3,200+ skills. It's the first middleware in the Reasoning Enhancement category. One command โ€” clawhub install marl-middleware โ€” gives your AI agent a metacognition upgrade.

๐Ÿ“ Technical deep dive: https://huggingface.co/blog/FINAL-Bench/marl-middleware
๐Ÿ“ฆ PyPI: https://pypi.org/project/marl-middleware/
๐Ÿ™ GitHub: https://github.com/Vidraft/MARL
๐Ÿฆ€ ClawHub: https://clawhub.ai/Cutechicken99/marl-middleware

#MARL #LLM #Hallucination #Metacognition #MultiAgent #AIMiddleware #FINALBench #OpenClaw #ClawHub #PyPI #AGI #HuggingFace #ReasoningAI #SelfCorrection #GlassBoxAI
Lyteย 
posted an update 3 days ago
view post
Post
5447
Introducing Nanochat Moroccan

Nanochat Moroccan is the first language model family built specifically for Moroccan Darija.

This project brings together a small family of models and datasets centered on Darija, with the goal of building something genuinely useful for a language that is still underserved in AI.

1. Models

- KandirResearch/Nanochat-Moroccan-Base-0.7B
- KandirResearch/Nanochat-Moroccan-Instruct-0.7B-pt-raw
- KandirResearch/Nanochat-Moroccan-Instruct-0.7B

2. Data

- Lyte/darija-pretraining-corpus
- Lyte/darija-pretraining-corpus-nanochat
- Lyte/Moroccan-Darija-Instruct-573K
- GemMaroc/TULU-3-50k-darija-english

3. Collection

- https://huggingface.co/collections/KandirResearch/nanochat-the-first-moroccan-darija-language-model-family

Moroccan Darija is spoken by millions of people, yet it remains underrepresented in language technology. Nanochat Moroccan is a step toward building tools that take the language seriously.

You are welcome to try it and chat with it here:
Lyte/Nanochat-Moroccco-Instruct
SeaWolf-AIย 
posted an update 3 days ago
view post
Post
5980
ALL Bench Leaderboard โ€” Structural Problems in AI Benchmarking and the Case for Unified Evaluation

FINAL-Bench/all-bench-leaderboard

The AI benchmark ecosystem has three structural problems. Major benchmarks like MMLU have surpassed 90%, losing discriminative power. Most leaderboards publish unverified self-reported scores โ€” our cross-verification found Claude Opus 4.6's ARC-AGI-2 listed as 37.6% (actual: 68.8%), Gemini 3.1 Pro as 88.1% (actual: 77.1%). OpenAI's own audit confirmed 59.4% of SWE-bench Verified tasks are defective, yet it remains widely used.

ALL Bench addresses this by comparing 91 models across 6 modalities (LLM ยท VLM ยท Agent ยท Image ยท Video ยท Music) with 3-tier confidence badges (โœ“โœ“ cross-verified ยท โœ“ single-source ยท ~ self-reported). Composite scoring uses a 5-Axis Framework and replaces SWE-Verified with contamination-resistant LiveCodeBench.

Key finding: metacognition is the largest blind spot. FINAL Bench shows Error Recovery explains 94.8% of self-correction variance, yet only 9 of 42 models are even measured. The 9.2-point spread (Kimi K2.5: 68.71 โ†’ rank 9: 59.5) is 3ร— the GPQA top-model spread, suggesting metacognition may be the single biggest differentiator among frontier models today.

VLM cross-verification revealed rank reversals โ€” Claude Opus 4.6 leads MMMU-Pro (85.1%) while Gemini 3 Flash leads MMMU (87.6%), producing contradictory rankings between the two benchmarks.

๐Ÿ“Š Article: https://huggingface.co/blog/FINAL-Bench/all-bench
๐Ÿ“ฆ Dataset: FINAL-Bench/ALL-Bench-Leaderboard
โšก GitHub: https://github.com/final-bench/ALL-Bench-Leaderboard
๐Ÿ† Leaderboard: FINAL-Bench/all-bench-leaderboard
๐Ÿงฌ FINAL Bench: FINAL-Bench/Metacognitive
DavidAUย 
posted an update 1 day ago
view post
Post
2133
21 Qwen 3.5 Fine Tunes (thinking and instruct) ; reg and uncensored (2B to 27B) exceed benchmarks, and work better than org models.

All are bench marked against org model.
Many exceed all benchmarks of org model.
Claude, GLM, Gemini and other distills.
Thinking AND dedicated Instruct versions.

Core goal: Increase benchmarks, and address long thinking blocks.

Highlights:

9B and 27B instruct "Claude" versions hit 624 and 675 on the "ARC-C" (hard challenge).

Thinking fine tunes exceed org model performance (in thinking mode).

In many cases there is a drastic reduction in thinking block size.

9B Claude Heretic Uncensored, GGUF :
-Neo, Code Imatrix (duel imatrix)
- Updated Jinja template
- Custom tensor enhancements.

DavidAU/Qwen3.5-9B-Claude-4.6-OS-Auto-Variable-HERETIC-UNCENSORED-THINKING-MAX-NEOCODE-Imatrix-GGUF

COLLECTION [21 models]:
https://huggingface.co/collections/DavidAU/qwen-35-08-2-4-9-27-35b-regular-uncensored
  • 1 reply
ยท
SeaWolf-AIย 
posted an update about 6 hours ago
view post
Post
364
๐ŸŸ๏ธ Smol AI WorldCup: A 4B Model Just Beat 8B โ€” Here's the Data

We evaluated 18 small language models from 12 makers on 125 questions across 7 languages. The results challenge the assumption that bigger is always better.

Community Article: https://huggingface.co/blog/FINAL-Bench/smol-worldcup
Live Leaderboard: huggingface.co/spaces/ginigen-ai/smol-worldcup
Dataset: huggingface.co/datasets/ginigen-ai/smol-worldcup

What we found:

โ†’ Gemma-3n-E4B (4B, 2GB RAM) outscores Qwen3-8B (8B, 5.5GB). Doubling parameters gained only 0.4 points. RAM cost: 2.75x more.

โ†’ GPT-OSS-20B fits in 1.5GB yet matches Champions-league dense models requiring 8.5GB. MoE architecture is the edge AI game-changer.

โ†’ Thinking models hurt structured output. DeepSeek-R1-7B scores 8.7 points below same-size Qwen3-8B and runs 2.7x slower.

โ†’ A 1.3B model fabricates confident fake content 80% of the time when prompted with nonexistent entities. Qwen3 family hits 100% trap detection across all sizes.

โ†’ Qwen3-1.7B (1.2GB) outscores Mistral-7B, Llama-3.1-8B, and DeepSeek-R1-14B. Latest architecture at 1.7B beats older architecture at 14B.

What makes this benchmark different?

Most benchmarks ask "how smart?" โ€” we measure five axes simultaneously: Size, Honesty, Intelligence, Fast, Thrift (SHIFT). Our ranking metric WCS = sqrt(SHIFT x PIR_norm) rewards models that are both high-quality AND efficient. Smart but massive? Low rank. Tiny but poor? Also low.

Top 5 by WCS:
1. GPT-OSS-20B โ€” WCS 82.6 โ€” 1.5GB โ€” Raspberry Pi tier
2. Gemma-3n-E4B โ€” WCS 81.8 โ€” 2.0GB โ€” Smartphone tier
3. Llama-4-Scout โ€” WCS 79.3 โ€” 240 tok/s โ€” Fastest model
4. Qwen3-4B โ€” WCS 76.6 โ€” 2.8GB โ€” Smartphone tier
5. Qwen3-1.7B โ€” WCS 76.1 โ€” 1.2GB โ€” IoT tier

Built in collaboration with the FINAL Bench research team. Interoperable with ALL Bench Leaderboard for full small-to-large model comparison.

Dataset is open under Apache 2.0 (125 questions, 7 languages). We welcome new model submissions.
codelionย 
posted an update 1 day ago
view post
Post
2381
Scaling Pedagogical Pre-training to 10 Billion Tokens

New blog post exploring what happens when you take optimal data mixing insights and scale up the data generation itself.

We built Sutra, a multi-stage framework for generating pedagogical pre-training data guided by a knowledge graph of ~2,000 concepts across 9 domains. The pipeline includes structured content generation, six-dimension quality evaluation, diversity management across 20 content styles, and a cleaning stage to prevent collapse.

The result is codelion/sutra-10B, a 10.2 billion token pedagogical dataset with rich metadata (domain, complexity, prerequisites, quality scores) on every entry.

We trained codelion/SmolLM2-70M on it for 3 full epochs (30.6B tokens) on a single A10 GPU in ~78 hours.

Key finding: perplexity kept improving across epochs, but benchmark gains plateaued fast. At 70M parameters, the model hits a representational ceiling that more data alone can't break through.

Full writeup with comparisons against 7 other datasets, detailed benchmark breakdowns, and connections to recent work on synthetic data scaling, curriculum learning, and data mixing laws: https://huggingface.co/blog/codelion/scaling-pedagogical-pretraining-10-billion-tokens

All datasets at multiple scales (10M, 100M, 1B, 10B) plus seed concepts and an SFT variant are in the Sutra Pedagogical Datasets collection.
AINovice2005ย 
posted an update 3 days ago
view post
Post
374
Just published my first cuda kernel, inspired by Sage Attention. Feel free to try it out โ˜บ๏ธ

AINovice2005/attention-int8
  • 2 replies
ยท
Reubencfย 
posted an update 4 days ago
view post
Post
2600
๐Ÿš€ I am thrilled to announce the release of a new Konkani LLM!

We've seen some fantastic results for both translation and transliteration tasks, and I'm excited to share this progress with the community.

๐Ÿ“– Read the launch article and see the results: https://huggingface.co/blog/Reubencf/konkani-llm
๐Ÿค– Explore the model and collection:
konkani


I would love to hear your feedback or see what you build with it! #Konkani #LLM #NLP #HuggingFace #IndicNLP #Konkani
prithivMLmodsย 
posted an update 4 days ago
view post
Post
4219
The Qwen3.5 Multimodal Understanding Demo, powered by Qwen3.5-2B, is now available on HF Spaces! It is a lightweight model designed for fast image and video reasoning. Built with Gradio, the demo showcases Image QA, Video QA, object detection, and 2D point tracking, along with real-time token streaming.

๐Ÿค— Demo: prithivMLmods/Qwen-3.5-HF-Demo
โœ… Collection: https://huggingface.co/collections/prithivMLmods/multimodal-implementations
๐Ÿ”— Qwen3.5-2B: Qwen/Qwen3.5-2B

To learn more, visit the app page or the respective model pages.
AINovice2005ย 
posted an update 1 day ago