000
welcome to synjuku
Physical Intelligence requires Physical Data
We built this missing data layer. The world’s most physics-rich human demonstration dataset that fuses vision with the invisible forces, interactions, and inertial measurements of the real world, enabling robotics models to finally scale from "seeing" to "understanding."
001
the problem
We need the "Data Vault" for Physical Intelligence
Physical intelligence has not yet scaled like LLMs, a major bottleneck is the data that powers the models.
LLMS
Scaled because the internet provided trillions of tokens of text (logic + language).
Scaled because the internet provided billions of hours of video (pixels + observation).
Cannot embody today’s robotics hardware without high fidelity, physics-rich human demonstration data.
002
the gap
Why Current Human Demo Data Quality is Not Enough
Current state-of-the-art datasets (Open X-Embodiment, Egocentric 10k, DROID) rely on two main modes:
egocentric video
Head mounted camera or VR headset. While useful for high-level planning, this modality hits a hard ceiling for manipulation. I.E. Egocentric 10k.
Both of the methods are not hardware future-prone, as the degree-of-freedom increases, gesture mapping becomes increasingly difficult. At the moment, they only provide intermediate steps towards true physical intelligence.
No physical interaction information recorded, poor pose estimation and 3D scene reconstruction through egocentric video feed. All these contribute to poor transferability.
Models trained on data can only be as good as the data is. This sets a pessimistic ceiling of how good the models will be trained on the teleoperation dataset.
We capture the "invisible" data. By fusing high frequency inertial data (IMU), physical forces (wearable force sensors) with 3D Volumetric Point Clouds, we provide the ground truth for kinematics and dynamics of poses.
003
The TECHNOLOGY
Industrial-Grade Infrastructure
We operate a proprietary capture and cleaning pipeline designed for high-precision imitation learning, deployed in real-world factories rather than sterile labs.
the hardware
We partner with industry leading motion-capture companies with 10+ years track records to deliver the best hardware options possible.
Multi-view RGB + RGB-D (Depth/Volumetric).
Finger-mounted IMUs on a wearable glove assembly captures fine-grained vibration and motion dynamics, while refining pose estimation accuracy.
Auto hardware triggers and key point detection, ensuring multimodal synchronization.
the software
We deliver cleaned, accurately labeled, and sliced to trajectory dataset via a proprietary post-processing pipeline.
Pose estimation, point cloud generation, labeling, and slicing.
Automated pre-scoring of consistency, path optimality, and task completeness.
004
The product
most curated, physics-rich, scalable human demonstration dataset
We function as the TSMC of robotics data—a neutral foundry powering the ecosystem.
A new benchmark
A gold-standard, human expert demonstration dataset with a wide range of task diversity. Designed to outperform video-only baselines and serve as the industry reference, scaled with both high volume and great expertise.
005
Value proposition
why us, why now
Our team owns the supply (factory access), the process (cleaning pipeline), and customers (cooperate clients from top AI labs); and we are creating the new SOTA standard (physics-rich schema).
There aren’t many ways a dataset can scale in labs. Unlike lab-based collection, our distributed factory workbench hardware scales across global industrial sites, capturing the diversity of the real world while ensuring an incomplete data volume.
Hardware is approaching human capability; VLA models are ready and data-hungry; Data is the only missing layer. It’s a classic egg or chicken problem. Once the new dataset is there, the research direction aligns.