 Autonomous Systems LabOpen OpportunitiesSemantic segmentation augments visual information from cameras or geometric information from LiDARs by classifying what objects are present in a scene. Fusing this semantic information with visual or geometric sensor data can improve the odometry estimate of a robot moving through the scene. Uni-modal semantic odometry approaches using camera images or LiDAR point clouds have been shown to outperform traditional single-sensor approaches. However, multi-sensor odometry approaches typically provide more robust estimation in degenerate environments. - Computer Vision, Image Processing, Intelligent Robotics, Signal Processing
- Master Thesis, Semester Project
| Lidar-Visual-Inertial odometry approaches [1-3] aim to overcome the limitations of the individual sensing modalities by estimating a pose from heterogenous measurements. Lidar-inertial odometry often diverges in environments with degenerate geometric structures and visual-inertial odometry can diverge in environments with uniform texture. Many existing lidar-visual-inertial odometry approaches use independent lidar-inertial and visual-inertial pipelines [2-3] to compute odometry estimates that are combined in a joint optimisation to obtain a single pose estimate. These approaches are able to obtain a robust pose estimate in degenerate environments but often underperform lidar-inertial or visual-inertial methods in non-degenerate scenarios due to the complexity of maintaining and combining odometry estimates from multiple representations. - Computer Vision, Intelligent Robotics, Signal Processing
- Master Thesis, Semester Project
| Existing lidar-inertial odometry approaches (e.g., FAST-LIO2 [1]) are capable of providing sufficiently accurate pose estimation in structured environments to capture high quality 3D maps of static structures in real-time. However, the presence of dynamic objects in an environment can reduce the accuracy of the odometry estimate and produce noisy artifacts in the captured 3D map. Existing approaches to handling dynamic objects [2-4] focus on detecting and filtering them from the captured 3D map but typically operate independently from the odometry pipeline, which means that the dynamic filtering does not improve the pose estimation accuracy. - Computer Vision, Engineering and Technology, Intelligent Robotics, Signal Processing
- Master Thesis, Semester Project
| This project focuses on the generation of detailed 3D models from a user-specified set of 3D cuboids. - Computer Vision
- ETH Zurich (ETHZ), Master Thesis, Semester Project
|
|