超要約:いちご収穫ロボットの「見えない!ズレる!」を解決して、賢く収穫する技術だよ☆
✨ ギャル的キラキラポイント ✨ ● 難しいこと全部お任せ!マルチタスクAIが優秀すぎ✨ ● ズレも落下も怖くない!自律(じりつ)で問題解決! ● データ活用で未来の農業をハッピーにしちゃお♪
🍓 詳細解説 🍓 背景 人手不足(ひとでぶそく)の農業を助けるべく、ロボット開発が進んでるんだけど、従来のロボは「いちごが見えない」「掴(つか)めない」「落としちゃう」って問題があったんだよね😭
方法 そこで!マルチタスクAI「SRR-Net」を開発!✨いちごを見つける、形を認識する、熟(う)れ具合を判断する…全部同時にできちゃうんだって!さらに、ズレを直す技術や、落ちそうなのを感知する機能も搭載(とうさい)!🤖賢すぎ💖
続きは「らくらく論文」アプリで
Strawberry harvesting robots faced persistent challenges such as low integration of visual perception, fruit-gripper misalignment, empty grasping, and strawberry slippage from the gripper due to insufficient gripping force, all of which compromised harvesting stability and efficiency in orchard environments. To overcome these issues, this paper proposed a visual fault diagnosis and self-recovery framework that integrated multi-task perception with corrective control strategies. At the core of this framework was SRR-Net, an end-to-end multi-task perception model that simultaneously performed strawberry detection, segmentation, and ripeness estimation, thereby unifying visual perception with fault diagnosis. Based on this integrated perception, a relative error compensation method based on the simultaneous target-gripper detection was designed to address positional misalignment, correcting deviations when error exceeded the tolerance threshold. To mitigate empty grasping and fruit-slippage faults, an early abort strategy was implemented. A micro-optical camera embedded in the end-effector provided real-time visual feedback, enabling grasp detection during the deflating stage and strawberry slip prediction during snap-off through MobileNet V3-Small classifier and a time-series LSTM classifier. Experiments demonstrated that SRR-Net maintained high perception accuracy. For detection, it achieved a precision of 0.895 and recall of 0.813 on strawberries, and 0.972/0.958 on hands. In segmentation, it yielded a precision of 0.887 and recall of 0.747 for strawberries, and 0.974/0.947 for hands. For ripeness estimation, SRR-Net attained a mean absolute error of 0.035, while simultaneously supporting multi-task perception and sustaining a competitive inference speed of 163.35 FPS.