EXOKERN
AI & ML interests
Force/torque data infrastructure for Physical AI and robotics foundation models. Specializing in contact-rich manipulation datasets, sim-to-real calibration, and tactile sensing for Large Behavior Models (LBMs). Building the data engine behind the next generation of humanoid robots.
Recent Activity
EXOKERN β Pre-trained Manipulation Skills for Physical AI
Force/torque-validated. Multi-seed evaluated. Open & reproducible.
We build pre-trained robotic manipulation skills for contact-rich tasks β where vision alone fails and precise force control matters. Every skill ships with full F/T ablation studies, multi-seed evaluation, and standardized benchmarks.
π― The Problem
Over 95% of robotic manipulation approaches are vision-only. They fail at contact-rich tasks like insertion, assembly, and snap-fit β where sub-Newton force control is the difference between success and broken parts.
Training a robot for a new contact task takes weeks of engineering. Validating that it works safely takes even longer.
π‘ What EXOKERN Provides
Skills β Pre-trained policies ready for deployment
| Skill | Version | Status |
|---|---|---|
| skill-forge-peginsert-v0 | v0 (fixed conditions) | β Live |
| skill-forge-peginsert-v0.1.1 | v0.1.1 (domain randomized) | π Training |
More skills in development: Screw Driving, Snap-Fit Assembly, Gear Meshing, Wire Routing, Pick-and-Place.
Datasets β Contact-rich training data with F/T annotations
| Dataset | Episodes | Domain Randomization | Status |
|---|---|---|---|
| contactbench-forge-peginsert-v0 | 2,221 | No | β Live |
| contactbench-forge-peginsert-v0.1 | 5,000 | Yes | β Live |
| contactbench-forge-peginsert-v0.1.1 | 5,000 | Yes | β Live (production) |
Tools β Evaluate any manipulation policy
pip install exokern-eval
exokern-eval --policy your_checkpoint.pt --env Isaac-Forge-PegInsert-Direct-v0 --episodes 100
- PyPI:
exokern-eval - GitHub: github.com/Exokern/exokern_eval
π Validated Results β Peg Insert v0
Evaluated across 3 random seeds Γ 2 conditions Γ 100 episodes = 600 rollouts total:
- 100% success rate across all seeds and conditions
- 38% average force reduction with F/T-aware policies (3.2N vs 5.2N)
- Consistent across architectures: MLP, Temporal CNN, and Diffusion Policy all show F/T benefit
- Reproducible: All checkpoints, configs, and eval scripts published
π§ What Makes EXOKERN Different
- Force/torque ablation on every skill β quantified proof that F/T data improves performance
- Multi-seed evaluation β not cherry-picked single runs, statistically validated results
- Open benchmarks β use
exokern-evalto test your own policies against our baselines - LeRobot v3.0 compatible β datasets plug directly into standard training pipelines
- Industrially relevant tasks β insertion, assembly, contact-rich manipulation
- Full data provenance β capture methodology, sensor specs, evaluation protocol documented
ποΈ Built With
- NVIDIA Isaac Lab for high-fidelity physics simulation
- Diffusion Policy architecture (71.3M parameters)
- Franka FR3 (7-DOF) robot platform
- LeRobot v3.0 format for interoperability
- Simulated 6-axis F/T sensing (Bota Systems SensONE modeled)
πΊοΈ Roadmap
Building toward a full catalog of contact-rich manipulation skills with multi-robot support:
| Version | What's New | Timeline |
|---|---|---|
| v0 | First Skill, Sim-only, Proof-of-Concept | β Done |
| v0.1 | Domain Randomization, robustere Policy | π In Progress |
| v0 Catalog | More tasks (Screw, Snap-Fit, Gear, Wire, Pick-Place) | MayβSep 2026 |
| v1 | Multi-Robot via Task-Space + Adapter Layer | JunβSep 2026 |
| v2 | Real-World validated, Enterprise-ready | 2027 |
See the full roadmap on GitHub.
π¬ Contact
- π€ HuggingFace: huggingface.co/EXOKERN
- π GitHub: github.com/Exokern
- π§ Contact: info@exokern.com
EXOKERN β Bridging the Haptic Gap in Robotic Manipulation