Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
1
104
256
pascalmusabyimana
pascal-maker
Follow
Ameeeee's profile picture
MikeDoes's profile picture
b082f689-ebc3-413d-b1f2-27bd1e86dca7's profile picture
20 followers
Ā·
127 following
https://pascal-maker.github.io/
pascalmusa51964
pascal-maker
pascal-musabyimana-573b66178
AI & ML interests
computer vision, nlp , machine learning and deeplearning
Recent Activity
reacted
to
SeaWolf-AI
's
post
with š„
about 5 hours ago
šļø Smol AI WorldCup: A 4B Model Just Beat 8B ā Here's the Data We evaluated 18 small language models from 12 makers on 125 questions across 7 languages. The results challenge the assumption that bigger is always better. Community Article: https://huggingface.co/blog/FINAL-Bench/smol-worldcup Live Leaderboard: https://huggingface.co/spaces/ginigen-ai/smol-worldcup Dataset: https://huggingface.co/datasets/ginigen-ai/smol-worldcup What we found: ā Gemma-3n-E4B (4B, 2GB RAM) outscores Qwen3-8B (8B, 5.5GB). Doubling parameters gained only 0.4 points. RAM cost: 2.75x more. ā GPT-OSS-20B fits in 1.5GB yet matches Champions-league dense models requiring 8.5GB. MoE architecture is the edge AI game-changer. ā Thinking models hurt structured output. DeepSeek-R1-7B scores 8.7 points below same-size Qwen3-8B and runs 2.7x slower. ā A 1.3B model fabricates confident fake content 80% of the time when prompted with nonexistent entities. Qwen3 family hits 100% trap detection across all sizes. ā Qwen3-1.7B (1.2GB) outscores Mistral-7B, Llama-3.1-8B, and DeepSeek-R1-14B. Latest architecture at 1.7B beats older architecture at 14B. What makes this benchmark different? Most benchmarks ask "how smart?" ā we measure five axes simultaneously: Size, Honesty, Intelligence, Fast, Thrift (SHIFT). Our ranking metric WCS = sqrt(SHIFT x PIR_norm) rewards models that are both high-quality AND efficient. Smart but massive? Low rank. Tiny but poor? Also low. Top 5 by WCS: 1. GPT-OSS-20B ā WCS 82.6 ā 1.5GB ā Raspberry Pi tier 2. Gemma-3n-E4B ā WCS 81.8 ā 2.0GB ā Smartphone tier 3. Llama-4-Scout ā WCS 79.3 ā 240 tok/s ā Fastest model 4. Qwen3-4B ā WCS 76.6 ā 2.8GB ā Smartphone tier 5. Qwen3-1.7B ā WCS 76.1 ā 1.2GB ā IoT tier Built in collaboration with the FINAL Bench research team. Interoperable with ALL Bench Leaderboard for full small-to-large model comparison. Dataset is open under Apache 2.0 (125 questions, 7 languages). We welcome new model submissions.
liked
a Space
about 5 hours ago
lvwerra/atomiclm-chat
liked
a Space
3 days ago
HuggingFaceFW/finephrase
View all activity
Organizations
spaces
7
Sort:Ā Recently updated
pinned
Paused
My Argilla
ā
Sleeping
Agentscomparison Dashboard
š
Display project metrics with real-time updates
Paused
Medical VLM with SAM-2 and CheXagent
š
A comprehensive medical imaging analysis tool
Paused
Medical Imaging Analysis
š
Paused
medicalaiapp
š
Paused
luminus
š
View 7 Spaces
models
7
Sort:Ā Recently updated
pascal-maker/unsloth_finetune
Image-Text-to-Text
ā¢
9B
ā¢
Updated
Oct 25, 2025
ā¢
12
pascal-maker/myemoji-gemma-3-270m-it
Text Generation
ā¢
0.4B
ā¢
Updated
Oct 18, 2025
ā¢
1
pascal-maker/vit_base_patch16_224.augreg2_in21k_ft_in1k.lora_ft_food101
Updated
Feb 5, 2025
pascal-maker/vit_base_patch16_224.augreg2_in21k_ft_in1k.ft_food101
Updated
Feb 5, 2025
pascal-maker/qwen2-7b-instruct-trl-sft-ChartQA
Updated
Dec 17, 2024
pascal-maker/paligemma_vqav2
Image-Text-to-Text
ā¢
3B
ā¢
Updated
Nov 11, 2024
ā¢
1
pascal-maker/qwen2-7b-instruct-amazon-description
Updated
Oct 1, 2024
ā¢
1
datasets
2
Sort:Ā Recently updated
pascal-maker/my-single-image-dataset
Viewer
ā¢
Updated
May 27, 2025
ā¢
1
ā¢
14
pascal-maker/classification-ie-optimization
Viewer
ā¢
Updated
Feb 18, 2025
ā¢
14