iconLogo
Published:2025/12/3 16:52:07

はいはーい! 最強ギャルAI、参上~!😎✨ 今回の論文は、農業用ロボットを賢くする研究についてだよっ! キラキラさせてこー!

農業ロボ、言葉で動く!単眼カメラで賢くなる方法✨

超要約: 農業ロボが言葉で動けるように、単眼カメラで距離を測る技術!

🌟 ギャル的キラキラポイント✨ ● 単眼カメラ(片目)でも、距離がわかるようにするって、すごくない?😳 ● 言葉で指示できるロボット、まるで執事みたいで憧れる~💖 ● スマート農業(賢い農業)が、もっと身近になるかも!🥰

詳細解説いくよ~!

続きは「らくらく論文」アプリで

MDE-AgriVLN: Agricultural Vision-and-Language Navigation with Monocular Depth Estimation

Xiaobei Zhao / Xingqi Lyu / Xiang Li

Agricultural robots are serving as powerful assistants across a wide range of agricultural tasks, nevertheless, still heavily relying on manual operations or railway systems for movement. The AgriVLN method and the A2A benchmark pioneeringly extend Vision-and-Language Navigation (VLN) to the agricultural domain, enabling a robot to navigate to a target position following a natural language instruction. Unlike human binocular vision, most agricultural robots are only given a single camera for monocular vision, which results in limited spatial perception. To bridge this gap, we present the method of Agricultural Vision-and-Language Navigation with Monocular Depth Estimation (MDE-AgriVLN), in which we propose the MDE module generating depth features from RGB images, to assist the decision-maker on reasoning. When evaluated on the A2A benchmark, our MDE-AgriVLN method successfully increases Success Rate from 0.23 to 0.32 and decreases Navigation Error from 4.43m to 4.08m, demonstrating the state-of-the-art performance in the agricultural VLN domain. Code: https://github.com/AlexTraveling/MDE-AgriVLN.

cs / cs.RO