Knowledge Diversion for Efficient Morphology Control and Policy Transfer
Fu Feng ⋅ Ruixiao Shi ⋅ Yucheng Xie ⋅ Jianlu Shen ⋅ Jing Wang ⋅ Xin Geng
Abstract
Universal morphology control aims to learn a universal policy that generalizes across heterogeneous robot morphologies, with Transformer-based controllers emerging as a dominant choice. However, such architectures incur substantial computational costs, resulting in high deployment overhead, and existing methods exhibit limited cross-task generalization, necessitating training from scratch for each new task. To this end, we propose DivMorph, a modular training paradigm that leverages knowledge diversion to learn \textit{decomposable controllers}. DivMorph factorizes randomly initialized Transformer weights into \textit{basic knowledge units} via SVD and employs dynamic soft gating, conditioned on task and morphology embeddings, to adaptively modulate these units into universal \textit{learngenes} and morphology- and task-specific \textit{tailors} during training, thereby achieving knowledge disentanglement. By selectively activating relevant components, DivMorph adaptively recomposes the controller, enabling efficient policy deployment and effective policy transfer to novel tasks. Extensive experiments demonstrate that DivMorph achieves state-of-the-art performance, improving sample efficiency for cross-task transfer by 3.3$\times$ and reducing model size for single-agent deployment by 16.7$\times$.
Successful Page Load