
Tabletop Dual-Arm
Research-grade manipulation, lab-desk footprint.
- Configuration
- 2 × 6-DoF arms, fixed base
- Payload
- 3 kg per arm
- Reach
- 650 mm
- Control
- Up to 1000 Hz, joint + Cartesian
- Sensing
- Dual wrist cameras + scene camera
Now accepting waitlist
Dual-arm hardware — tabletop and mobile — paired with soda OS, our agentic OS for teleop, data collection, and VLM-orchestrated manipulation. Out of the box.
Research labs and AI developers only for now. Why we’re building this →

Why SOMA
Foundation models can see and plan. What they cannot do, reliably, is act in the world. The bottleneck is no longer intelligence — it is a platform. A hardware-software stack that lets any AI developer collect data, run policies, and ship skills without building a robotics company first.
SOMA is that platform. Dual-arm robots that arrive working, with an agentic OS that treats manipulation the way Claude Code treats software: a VLM orchestrates a library of vision-action skills, calls tools, recovers from failure, and learns from every teleop session.
We build both layers because the interface between them is where every current system breaks.
Hardware
Shipped calibrated, with soda OS pre-installed.

Research-grade manipulation, lab-desk footprint.

Navigate, manipulate, return. One platform.
On the roadmap
A Kickstarter campaign for hobbyists, classrooms, and indie AI developers. Same soda OS, same skill library, one arm instead of two.
soda OS
A VLM sits at the top of the stack and orchestrates a growing library of vision-action skills — pick, place, pour, wipe, open, hand-off. Skills are trained from teleop data you collect, or from ours. Think Claude Code, but the tools are physical.
> load the red mug into the dishwasher
→ plan: locate(red_mug) ▸ grasp ▸ open(dishwasher) ▸ place
→ skill: grasp.mug [vla-policy v0.4]
→ skill: open.dishwasher
✓ done · 14.3sVR or Leader-arm input. Every session becomes training data.
Trajectories, videos, proprioception, and natural-language annotation — captured by default, exportable to LeRobot format.
High-level natural-language goals decomposed into skill calls. Failure recovery, retries, and human-in-the-loop handoff built in.
Ship pretrained primitives. Fine-tune with a few dozen demos. Register your own skill in a single Python file.
Traction
GRASP Summit
Demoed live at Penn GRASP Robotics Summit — 200+ roboticists in the room.
Research demand
Multiple US robotics labs have requested early-access quotes after live demos.
Join the waitlist for early access, research pricing, and soda OS beta.
Questions? hello@somarobotics.ai