超要約: 点群データに画像とテキストを足して、欠損した3Dモデルを最強に作り直す研究だよ!
✨ ギャル的キラキラポイント ✨
● 点群(3Dデータ)に画像とテキストを組み合わせる発想が天才的!まるで推しの"顔面偏差値"と"性格の良さ"を足して最強にするみたい💖 ● Modality Dropout(一部情報を無視)って、ちょっとした"ツンデレ"みたいな?入力の揺らぎに強くなるってことね! ● 大規模データセット「MGPC-1M」で、色んな3Dモデルに対応できるのがすごい!まるで推しの"イケメン度"の幅が広がる感じ😍
詳細解説
続きは「らくらく論文」アプリで
Point cloud completion aims to recover complete 3D geometry from partial observations caused by limited viewpoints and occlusions. Existing learning-based works, including 3D Convolutional Neural Network (CNN)-based, point-based, and Transformer-based methods, have achieved strong performance on synthetic benchmarks. However, due to the limitations of modality, scalability, and generative capacity, their generalization to novel objects and real-world scenarios remains challenging. In this paper, we propose MGPC, a generalizable multimodal point cloud completion framework that integrates point clouds, RGB images, and text within a unified architecture. MGPC introduces an innovative modality dropout strategy, a Transformer-based fusion module, and a novel progressive generator to improve robustness, scalability, and geometric modeling capability. We further develop an automatic data generation pipeline and construct MGPC-1M, a large-scale benchmark with over 1,000 categories and one million training pairs. Extensive experiments on MGPC-1M and in-the-wild data demonstrate that the proposed method consistently outperforms prior baselines and exhibits strong generalization under real-world conditions.