Medany Mahmoud, Piglia Lorenzo, Achenbach Liam, Mukkavilli S Karthik, Ahmed Daniel
Acoustic Robotics Systems Lab, Institute of Robotics and Intelligent Systems, Department of Mechanical and Process Engineering, ETH Zurich, Rüschlikon, Switzerland.
IBM Research - Europe, AI and Accelerated Discovery, Rüschlikon, Switzerland.
Nat Mach Intell. 2025;7(7):1076-1090. doi: 10.1038/s42256-025-01054-2. Epub 2025 Jun 26.
Reinforcement learning is emerging as a powerful tool for microrobots control, as it enables autonomous navigation in environments where classical control approaches fall short. However, applying reinforcement learning to microrobotics is difficult due to the need for large training datasets, the slow convergence in physical systems and poor generalizability across environments. These challenges are amplified in ultrasound-actuated microrobots, which require rapid, precise adjustments in high-dimensional action space, which are often too complex for human operators. Addressing these challenges requires sample-efficient algorithms that adapt from limited data while managing complex physical interactions. To meet these challenges, we implemented model-based reinforcement learning for autonomous control of an ultrasound-driven microrobot, which learns from recurrent imagined environments. Our non-invasive, AI-controlled microrobot offers precise propulsion and efficiently learns from images in data-scarce environments. On transitioning from a pretrained simulation environment, we achieved sample-efficient collision avoidance and channel navigation, reaching a 90% success rate in target navigation across various channels within an hour of fine-tuning. Moreover, our model initially generalized successfully in 50% of tasks in new environments, improving to over 90% with 30 min of further training. We further demonstrated real-time manipulation of microrobots in complex vasculatures under both static and flow conditions, thus underscoring the potential of AI to revolutionize microrobotics in biomedical applications.
强化学习正成为微机器人控制的有力工具,因为它能在传统控制方法失效的环境中实现自主导航。然而,由于需要大量训练数据集、物理系统中收敛速度慢以及跨环境的泛化能力差,将强化学习应用于微机器人技术存在困难。在超声驱动的微机器人中,这些挑战更加突出,因为它们需要在高维动作空间中进行快速、精确的调整,而这对人类操作员来说往往过于复杂。应对这些挑战需要样本高效的算法,这些算法能够从有限的数据中学习,同时管理复杂的物理相互作用。为了应对这些挑战,我们为超声驱动的微机器人的自主控制实现了基于模型的强化学习,该学习从循环想象环境中学习。我们的非侵入式、人工智能控制的微机器人提供精确的推进,并能在数据稀缺的环境中从图像中高效学习。从预训练的模拟环境过渡后,我们实现了样本高效的避碰和通道导航,在微调一小时内,在各种通道的目标导航中成功率达到90%。此外,我们的模型最初在新环境中的50%的任务中成功泛化,经过30分钟的进一步训练后提高到90%以上。我们还展示了在静态和流动条件下对复杂脉管系统中的微机器人进行实时操作,从而突出了人工智能在生物医学应用中彻底改变微机器人技术的潜力。