Skip to content

Deep learning pipeline for 6D pose estimation (translation and rotation) of rigid objects from RGB-D images, based on a modified PointNet++ architecture. It leverages both geometric and image features for high-accuracy results tailored for robotic manipulation tasks.

Notifications You must be signed in to change notification settings

BonoGiorgio02/6DPose_Estimation

Repository files navigation

MIT license PyTorch

6DPose_Estimation

This project addresses the task of 6D object pose estimation on the LINEMOD preprocessed dataset.

Moreover, the proposed architecture is trained to detect the pose of cones in a simulator of the student autonomous driving team of Politecnico of Turin Squadra Corse driverless.

How to run the code

To run the code, execute the notebook.

If you want to run with Colab:

  • download repo and store on Drive
  • extract it
  • open the notebook
  • connect to Drive (set MOUNT_DRIVE to True)
  • follow the notebook step by step

If you want to run locally:

  • you can clone the repo, but you may need to rename all the 6DPose_Estimation-main into 6DPose_Estimation

How to run inference

  • execute all the cells of Set up the project, Download dataset, Modify Dataset, Data Exploration, Define CustomDataset, Data Preprocessing for Object Detection Model, Visualize data
  • execute the inference cell (Inference Baseline or Inference Extension), it uses the test set (it may take a while to create training, validation, and test sets)

The repository is structured such that:

  • checkpoints contains saved models
  • data contains custom dataloader and dataset classes
  • datasets contains datasets
  • images contains images
  • models contains model architectures, metrics
  • utils contains multiple functions (init, data exploration, plot)
  • notebook and associated training logic
  • requirements contains necessary packages

Results obtained in LINEMOD dataset

Linemod_1 Linemod_2

Figure 1: Object 05.    Figure 2: Object 06.

Linemod_3 Linemod_4

Figure 3: Object 08.    Figure 4: Object 09.

Linemod_5 Linemod_6

Figure 5: Object 13.    Figure 6: Object 14.

Evaluation results

Overall:

Extension ADD Score Accuracy
RGB-D 0.0138 80.03%
YOLO + RGB-D 0.0144 77.03%

Results by object:

Object Ours
(baseline, RGB)
Ours
(baseline pipeline, YOLO + RGB)
Ours
(extension, RGB-D)
Ours
(extension pipeline, YOLO + RGB-D)
ape (01) 0.0 0.0 51.1 31.7
bench vi. (02) 8.8 1.1 89.6 84.1
camera (04) 0.0 1.1 70.0 65.0
can (05) 2.2 1.7 81.0 80.5
cat (06) 1.1 0.0 79.7 82.0
driller (08) 3.9 0.5 90.5 86.5
duck (09) 0.0 0.0 52.1 47.9
eggbox (10) 13.3 0.5 100.0 100.0
glue (11) 18.0 6.0 100.0 100.0
hole p. (12) 1.1 0.5 59.7 61.3
iron (13) 2.9 0.5 95.4 88.4
lamp (14) 8.2 1.6 91.9 92.9
phone (15) 2.7 2.1 81.5 83.2
MEAN 4.8 1.9 80.0 77.0

Table: Comparison of 6D pose estimation methods on the LineMOD dataset. Results are reported as accuracy (%) under the ADD(-S) metric.

Results obtained in autonomous driving dataset

Left camera frame Right camera frame

Figure 1: Left camera frame.    Figure 2: Right camera frame.

Left camera frame processed by YOLO Right camera frame processed by YOLO

Figure 3: Left camera frame processed by YOLO.    Figure 4: Right camera frame processed by YOLO.

Left camera frame processed by YOLO Right camera frame processed by YOLO

Figure 5: Cropped image of left cone.    Figure 6: Cropped image of right cone.

LiDAR pointcloud of the cone projected in left image LiDAR pointcloud of the cone projected in right image

Figure 7: LiDAR pointcloud of the left cone.    Figure 8: LiDAR pointcloud of the right cone.

LiDAR pointcloud of the cone projected in left image LiDAR pointcloud of the cone projected in right image

Figure 9: 6D pose estimation of the left cone.    Figure 10: 6D pose estimation of the right cone.

About

Deep learning pipeline for 6D pose estimation (translation and rotation) of rigid objects from RGB-D images, based on a modified PointNet++ architecture. It leverages both geometric and image features for high-accuracy results tailored for robotic manipulation tasks.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •