
【#Tech24H】On January 29, Unitree Technology open-sourced UnifoLM-VLA-0, a Visual-Language-Action (VLA) large model for general-purpose humanoid robot operations. As part of the UnifoLM series, this model aims to overcome the limitations of traditional Vision-Language Models (VLMs) in physical interactions. Through continued pre-training on robotic operational data, it evolves from general "image-text understanding" to an "embodied brain" equipped with physical common sense. Additionally, the model incorporates full-chain kinetic prediction data, enhancing its task generalization capabilities.

