contextual and motion cues
Information derived from human actions that provides context and guidance for robot learning.
Information derived from human actions that provides context and guidance for robot learning.
Policies that utilize 3D point data to inform the manipulation actions of a robot’s multi-fingered hands.
The disparity between human and robot capabilities that complicates the transfer of learned behaviors.
AINA is a framework designed to learn robot manipulation policies from human demonstrations in natural environments.
Lightweight smart glasses equipped with a high-resolution RGB camera and 3D pose estimation capabilities.
A representation of a sequence of CUDA operations that can be executed on a GPU, allowing for efficient parallel processing.
Large language models are advanced neural network architectures trained on vast amounts of text data to understand and generate human-like text.
A statistical distribution where a small number of events or items are extremely common, while a large number are rare.
A lightweight draft model that is continuously trained on idle GPUs to align with the target model during long-tail response generation.
A system that manages a memory-efficient pool of pre-captured CUDAGraphs and selects appropriate speculative decoding strategies for input batches.