Tag
A robotic arm demonstrates autonomous mid-flight capture of a Skydio F10 drone, enabling rapid recovery without pilot input.
The article discusses whether achieving widespread adoption of home robots capable of performing most chores requires Artificial General Intelligence (AGI), while expressing disappointment that advanced robot actions still largely rely on teleoperation.
The article discusses humanoid robots as the latest phase in the AI hype cycle, noting that while they are visually impressive, creating practical and cost-effective workers remains a significant challenge.
Jim Fan released 《Robotics: Endgame》as a sequel to his previous speech 《Physical Turing Test》, proposing a roadmap to solve physical AGI by analogy to the success story of LLM.
Ouster announces REV8, the first native color lidar sensor that fuses color and 3D data directly in silicon rather than in software, marking a hardware-level advancement in 3D sensing technology.
Figure taught two F.03 robots to fully autonomously clean a room and make a bed in under two minutes.
In his talk at Sequoia AI Ascent, Dr. Jim Fan presents a roadmap for achieving Physical AGI parallel to LLM success, introducing concepts like video world models, World Action Models (WAM), and the Dexterity Scaling Law, and sharing predictions for the near future.
A discussion post exploring where edge AI will have the greatest impact: autonomy and robotics, low-power vision systems, private local LLMs, or bandwidth-constrained industrial deployments.
Reachy Mini has a new fully open-source backend for real-time voice interaction, running audio models locally and leveraging LLM subscriptions to avoid per-second API costs.
This paper introduces Adaptive Q-Chunking (AQC), a reinforcement learning method that dynamically selects action chunk sizes to balance reactive control and long-horizon planning. It achieves state-of-the-art results on OGBench and Robomimic, enhancing the performance of large-scale VLA models in robotics tasks.
Venture capitalists have invested $50 million in humanoid robot startups that rely solely on vision-based systems.
A robotics engineer from Hugging Face proposes mapping human facial expressions onto non-humanoid robots to enhance expressiveness while avoiding the uncanny valley, with plans to use this data for autonomous body language training.
Remi Cadene highlights a new tendon-driven robotic hand with a teleoperation exoskeleton, noting its high degrees of freedom and debugging software.
Genesis AI highlights advancements in robotic manipulation using Gene 2.6.5, aiming to achieve human-level dexterity. The article discusses progress in training robots to perform complex physical tasks.
News report highlights a significant event in South Korea where the country's first autonomous humanoid robot converted to Buddhism, sparking discussion on AI consciousness and ethics.
South Korea's first autonomous humanoid robot has been converted to Buddhism, highlighting the intersection of advanced robotics and religious practice.
ReflectDrive-2 is a new discrete diffusion planner for autonomous driving that uses reinforcement learning to enable self-editing of trajectory tokens, achieving high performance and low latency on the NAVSIM benchmark.
MolmoAct 2 is an open robotics model that reasons in 3D space before taking actions, developed by the Allen Institute for Artificial Intelligence.
RLDX-1 is a general-purpose robotic policy for dexterous manipulation that uses a Multi-Stream Action Transformer architecture to integrate heterogeneous modalities, outperforming existing VLA models in real-world tasks.
Allen AI releases MolmoAct2, an open-weight Vision-Language-Action model designed for real-world robotic deployment, featuring new datasets, an open action tokenizer, and adaptive reasoning to reduce latency.