Records
4
HF AUTHOR CLUSTER
4 robotics-tagged HF records from HuggingFaceVLA, totaling 36,775 cumulative downloads. Some records cite published arxiv research.
DIRECT ANSWER
Author clusters consolidate every record from one publisher into a single buyer-review surface. HuggingFaceVLA ships 4 robotics datasets on Hugging Face. Top license: apache-2.0. Tier breakdown: 4 indexable as Tier A, 0 as Tier B, 0 demoted (those URLs redirect here).
4
36,775
apache-2.0
DATASETS
4 of 4 datasets
24,673 downloads · apache-2.0
This dataset was created using LeRobot. Dataset Structure meta/info.json: { "codebase_version": "v3.0", "robot_type": "panda", "total_episodes": 1693, "total_frames": 273465, "total_tasks": 40, "chunk
4,529 downloads · apache-2.0
Lerobot Community Datasets v3 - A Cross-Embodiment Pretraining Dataset for Vision Language Action Models A large-scale robotics dataset for vision-language-action learning, featuring 791 datasets acro
4,357 downloads · apache-2.0
Community Dataset v1 A large-scale community-contributed robotics dataset for vision-language-action learning, featuring 128 datasets from 55 contributors worldwide. We used this dataset to pretrain S
3,216 downloads · apache-2.0
Community Dataset v2 A large-scale community-contributed robotics dataset for vision-language-action learning, featuring 340 datasets from 117 contributors worldwide. This dataset represents the secon
RESEARCH PATHS
A dataset record is only useful when it connects into the rest of the buyer workflow. The next review step is usually not another summary; it is a fit check, rights triage, source comparison, or custom bounty spec that names the missing proof.
For physical AI teams, the hard question is whether the public source can support a specific model objective under real deployment constraints. That requires adjacent dataset records, tools, comparisons, and sourcing paths, plus external references that a reviewer can open and challenge.
Use the links below to keep the review grounded. Start broad when discovery is incomplete, move into profile and comparison pages when the candidate source is known, and switch to custom collection when the blocker is rights, consent, geography, robot embodiment, or target environment coverage.
INTERNAL LINKS
Use the catalog to compare source-backed dataset profiles by modality, task, rights signal, consent risk, and deployment fit.
Scan the broader robotics dataset surface before narrowing into promoted profiles, comparisons, and custom collection specs.
Track source updates, licensing notes, and buyer-readiness changes that should trigger a renewed review.
Score whether a public source is enough for the model, rights path, modalities, and target environment.
Separate source license language from contributor consent, redistribution, private-space risk, and model-use assumptions.
Turn a public-source gap into a scoped capture request with sample QA, metadata, and delivery requirements.
Compare data providers when the answer is not another public dataset but a better sourcing or capture route.
Use the company index to separate annotation vendors, data engines, marketplaces, and specialist capture teams.
EXTERNAL REFERENCES
Market context for why physical AI systems need custom, enriched, real-world data beyond generic labeling workflows.
Robotics dataset and tooling context for Hugging Face based collection, sharing, conversion, and training workflows.
A cross-embodiment robotics dataset reference for comparing trajectory scale, robot diversity, and VLA training assumptions.
A large in-the-wild robot manipulation dataset reference for real-world trajectory capture and deployment transfer risk.
TRUELABEL ROUTING
If the Hub records don't carry the license, consent, or deployment fit your team needs, commission a custom collection on the same modality with explicit commercial terms.