ジャガイモ収穫現場のRGB-Dカメラ映像から“完全な三次元”を再構成:「3DPotatoTwin」データを公開~収穫機の不完全3Dを補うAI基盤で、収量評価とスマート農業に貢献~

2025-10-28 東京大学

東京大学大学院農学生命科学研究科の研究チームは、ジャガイモ収穫現場のRGB-Dカメラ映像から完全な3D形状を再構成できるAI基盤「3DPotatoTwin」データセットを公開した。北海道更別村で3品種・300個体以上を撮影し、デプスカメラによる不完全3Dと高精度SfMモデルをペア化。これにより、AIが欠損データを補完し、収量評価・栽培技術の効果検証・自動収穫ロボットの精密マッピングなどに活用できる。低コストセンサーでも高精度な三次元解析が可能となり、スマート農業推進に貢献する。

ジャガイモ収穫現場のRGB-Dカメラ映像から“完全な三次元”を再構成:「3DPotatoTwin」データを公開~収穫機の不完全3Dを補うAI基盤で、収量評価とスマート農業に貢献~
「3DPotatoTwin」データの概要

<関連情報>

3DPotatoTwin: 3Dマルチ感覚融合のためのジャガイモ塊茎ペアデータセット 3DPotatoTwin: a Paired Potato Tuber Dataset for 3D Multi-Sensory Fusion

Haozhou Wang, Pieter M. Blok, James Burridge, Ting Jiang, Minato Miyauchi, Kyosuke Miyamoto, Kunihiro Tanaka, Wei Guo
Plant Phenomics  Available online: 6 October 2025
DOI:https://doi.org/10.1016/j.plaphe.2025.100123

Abstract

Accurate 3D phenotyping of agricultural produce remains challenging due to the inherent trade-offs between reconstruction quality and acquisition throughput in existing sensing technologies. While RGB-D cameras enable high-throughput scanning in operational settings like harvesting conveyors, they produce incomplete, low-quality 3D models. Conversely, close-range Structure-from-Motion (SfM) produces high-fidelity reconstructions but lacks high-throughput field applicability. This study bridges this gap through 3DPotatoTwin, a paired dataset containing 339 tuber samples across three cultivars collected in Hokkaido, Japan. Our dataset uniquely combines: (1) conveyor-acquired RGB-D point clouds, (2) ground measurement, (3) SfM reconstructions under indoor controlled environment, and (4) aligned model pairs with transformation matrices. The multi-sensory alignment employs an semi-supervised pin-guided pipeline incorporating single-pin extraction and referencing, cross-strip matching, and binary-color-enhanced ICP, achieving 0.59 ± 0.11 mm registration accuracy. Beyond serving as a benchmark for 3D phenotyping algorithms, the dataset enables training of 3D completion networks to reconstruct high-quality 3D models from partial RGB-D point clouds. Meanwhile, the proposed semi-automated annotation pipeline has the potential to accelerate 3D dataset generation for similar studies. The presented methodology demonstrates broader applicability for multi-sensor data fusion across crop phenotyping applications. The dataset and pipeline source code are publicly available at HuggingFace and GitHub, respectively.

1200農業一般
ad
ad
Follow
ad
タイトルとURLをコピーしました