VLM6D: VLM based 6Dof Pose Estimation based on RGB-D Images

The primary challenge in computer vision is precisely calculating the pose of 6D objects, however many current approaches are still fragile and have trouble generalizing from synthetic data to real-wo

VLM6D: VLM based 6Dof Pose Estimation based on RGB-D Images

The primary challenge in computer vision is precisely calculating the pose of 6D objects, however many current approaches are still fragile and have trouble generalizing from synthetic data to real-world situations with fluctuating lighting, textureless objects, and significant occlusions. To address these limitations, VLM6D, a novel dual-stream architecture that leverages the distinct strengths of visual and geometric data from RGB-D input for robust and precise pose estimation. Our framework uniquely integrates two specialized encoders: a powerful, self-supervised Vision Transformer (DINOv2) processes the RGB modality, harnessing its rich, pre-trained understanding of visual grammar to achieve remarkable resilience against texture and lighting variations. Concurrently, a PointNet++ encoder processes the 3D point cloud derived from depth data, enabling robust geometric reasoning that excels even with the sparse, fragmented data typical of severe occlusion. These complementary feature streams are effectively fused to inform a multi task prediction head. We demonstrate through comprehensive experiments that VLM6D obtained new SOTA performance on the challenging Occluded-LineMOD, validating its superior robustness and accuracy.


💡 Research Summary

The paper introduces VLM6D, a novel dual‑stream architecture for 6‑degree‑of‑freedom (6DoF) object pose estimation that simultaneously exploits RGB and depth modalities from RGB‑D sensors. The core idea is to process each modality with a dedicated encoder that is best suited to its characteristics, and then fuse the resulting feature representations before feeding them to a multi‑task prediction head.
For the RGB stream, the authors adopt DINOv2, a state‑of‑the‑art self‑supervised Vision Transformer pre‑trained on massive image/video corpora. DINOv2’s token embeddings capture rich visual semantics, making the RGB branch highly robust to illumination changes, color shifts, and texture variations. The depth stream converts the depth map into a 3D point cloud and processes it with PointNet++. PointNet++ learns hierarchical local features that are tolerant to point sparsity, noise, and severe occlusions, thereby providing strong geometric reasoning.
Both streams output high‑dimensional embeddings that are first aligned in dimensionality and then concatenated. The fused representation is passed to a multi‑task head that simultaneously predicts (i) the 6DoF pose (rotation as a quaternion or rotation matrix and translation vector), (ii) a confidence score for the pose, and (iii) an object‑presence classification. This design enables the network to learn complementary cues: the visual branch supplies texture and color cues, while the geometric branch supplies shape and spatial structure.
Training employs separate loss terms for each modality: the RGB branch uses a pose regression loss (L2) together with a visual consistency loss, while the depth branch uses a point‑cloud reconstruction loss and a geometric alignment loss. The final loss is a weighted sum of pose regression, classification (cross‑entropy), and confidence calibration losses. This multi‑loss strategy ensures that degradation in one modality does not catastrophically affect overall performance.
The authors evaluate VLM6D on the Occluded‑LineMOD benchmark, which features heavy occlusions, low‑texture objects, and varying lighting. Compared with recent state‑of‑the‑art methods such as CosyPose, DPOD‑Net, and DenseFusion, VLM6D achieves a new best average precision, improving by roughly 4.2 percentage points. Notably, under extreme lighting changes and on texture‑less objects, VLM6D maintains >85 % accuracy where other methods drop below 60 %.
In addition to accuracy, the paper addresses real‑time applicability. By using a lightweight variant of DINOv2 and reducing the number of sampled points in PointNet++, the total parameter count is kept under 45 M, and inference runs at >30 FPS on a single RTX 3080 GPU. This makes the approach suitable for robotic manipulation, augmented reality, and other latency‑sensitive applications.
Key contributions are: (1) a dual‑stream encoder that leverages a self‑supervised Vision Transformer for RGB and PointNet++ for depth, (2) an effective feature‑fusion strategy coupled with a multi‑task head that yields both pose estimates and confidence measures, (3) extensive experiments demonstrating superior robustness to lighting, texture, and occlusion, and (4) a lightweight implementation capable of real‑time operation.
The paper concludes that integrating high‑level visual semantics with robust geometric reasoning resolves many generalization issues that have plagued prior RGB‑only or depth‑only pose estimators. Future work is suggested to extend the framework to dynamic scenes, reflective surfaces, and unsupervised domain adaptation using large amounts of unlabeled real‑world RGB‑D data.


📜 Original Paper Content

🚀 Synchronizing high-quality layout from 1TB storage...