Software-Hardware Co-Design For Embodied AI Robots

📅 2024-07-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address high latency and energy consumption in LLM-driven robotic manipulation—caused by frame-by-frame action generation—this paper proposes Corki, an algorithm-architecture co-design framework. Methodologically, Corki introduces (i) a novel trajectory-level LLM inference paradigm that predicts entire action trajectories instead of per-frame actions, drastically reducing inference frequency; (ii) hardware-level parallel execution of computation, low-level control, and data communication; and (iii) an end-to-end pipelined workflow enabled by software-hardware co-compiled optimizations. Evaluation demonstrates that Corki reduces LLM inference frequency by up to 8.0×, decreases end-to-end latency by 3.6×, and improves task success rate by 17.3%. The framework is fully open-sourced, enabling complete reproducibility.

Technology Category

Application Category

📝 Abstract
Embodied AI robots have the potential to fundamentally improve the way human beings live and manufacture. Continued progress in the burgeoning field of using large language models to control robots depends critically on an efficient computing substrate. In particular, today's computing systems for embodied AI robots are designed purely based on the interest of algorithm developers, where robot actions are divided into a discrete frame-basis. Such an execution pipeline creates high latency and energy consumption. This paper proposes Corki, an algorithm-architecture co-design framework for real-time embodied AI robot control. Our idea is to decouple LLM inference, robotic control and data communication in the embodied AI robots compute pipeline. Instead of predicting action for one single frame, Corki predicts the trajectory for the near future to reduce the frequency of LLM inference. The algorithm is coupled with a hardware that accelerates transforming trajectory into actual torque signals used to control robots and an execution pipeline that parallels data communication with computation. Corki largely reduces LLM inference frequency by up to 8.0x, resulting in up to 3.6x speed up. The success rate improvement can be up to 17.3%. Code is provided for re-implementation. https://github.com/hyy0613/Corki
Problem

Research questions and friction points this paper is trying to address.

High latency and energy in AI robotic manipulation systems
Disjointed execution pipeline in embodied AI robots
Frequent LLM inference slowing robotic task performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Algorithm-architecture co-design for robotic manipulation
Decouples LLM inference, control, and data communication
Predicts future trajectory to reduce LLM inference frequency
🔎 Similar Papers
No similar papers found.
Y
Yiyang Huang
ICT, Chinese Academy of Sciences, China
Y
Yuhui Hao
ICT, Chinese Academy of Sciences, China
B
Bo Yu
Shenzhen Institute of Artificial Intelligence and Robotics for Society, China
F
Feng Yan
Meituan, China
Y
Yuxin Yang
ICT, Chinese Academy of Sciences, China
F
Feng Min
ICT, Chinese Academy of Sciences, China
Y
Yinhe Han
ICT, Chinese Academy of Sciences, China
L
Lin Ma
Meituan, China
Shaoshan Liu
Shaoshan Liu
PerceptIn
Embodied AIAutonomous Machine ComputingComputer SystemsTechnology Policy
Q
Qiang Liu
Tianjin University, China
Y
Yiming Gan
ICT, Chinese Academy of Sciences, China