SiROP
Login   
Language
  • English
    • English
    • German
Home
Menu
  • Login
  • Register
  • Search Opportunity
  • Search Organization
  • Create project alert
Information
  • About SiROP
  • Team
  • Network
  • Partners
  • Imprint
  • Terms & conditions
Register now After registration you will be able to apply for this opportunity online.
This opportunity is not published. No applications will be accepted.

Closing Sim-to-Real Visual Domain Gap for Transparent and Reflective Objects

Robots need to manipulate a wide range of unknown objects, from transparent to shiny surfaces. The goal of this project is to investigate learning techniques to bridge the visual domain gap between high-fidelity rendered scenes and real-world images for scene understanding.

Keywords: computer vision, sim-to-real, machine learning, perception, robotics

  • While humans are adept at differentiating an object and its texture properties, robots are easily fooled by common household objects such as glass tables, shiny toasters, or plastic bottles. The limitations of depth sensing on current hardware limit real-world data collection for these objects. This affects the generalization of the learned models to different lighting conditions and scene variations [1]. Prior works [1, 2] rely on stereo images based sparse cost volume computation or real-world. Recent advances in physically-based rendering show promise in collecting high-fidelity renderings of diverse synthetic scenes and creating large datasets for learning [2]. The goal of this project is to leverage these advances and investigate how to close the visual domain gap for deploying learned models for the 3D shape estimation of transparent or shiny objects. References: - Sajjan, Shreeyak, et al. "Clear grasp: 3d shape estimation of transparent objects for manipulation." ICRA (2020). - Kollar, Thomas, et al. "Simnet: Enabling robust unknown object manipulation from pure synthetic data via stereo." CoRL (2022).

    While humans are adept at differentiating an object and its texture properties, robots are easily fooled by common household objects such as glass tables, shiny toasters, or plastic bottles. The limitations of depth sensing on current hardware limit real-world data collection for these objects. This affects the generalization of the learned models to different lighting conditions and scene variations [1]. Prior works [1, 2] rely on stereo images based sparse cost volume computation or real-world. Recent advances in physically-based rendering show promise in collecting high-fidelity renderings of diverse synthetic scenes and creating large datasets for learning [2]. The goal of this project is to leverage these advances and investigate how to close the visual domain gap for deploying learned models for the 3D shape estimation of transparent or shiny objects.

    References:
    - Sajjan, Shreeyak, et al. "Clear grasp: 3d shape estimation of transparent objects for manipulation." ICRA (2020).
    - Kollar, Thomas, et al. "Simnet: Enabling robust unknown object manipulation from pure synthetic data via stereo." CoRL (2022).

  • - Setup a learning environment for synthetic data generation using NVIDIA Omniverse / Isaac Sim - Literature review on online learning techniques, such as automatic domain randomization, and recent advances in implicit representations, such as neural fields - Designing a method to perform 3D shape estimation using only simulated data - Show generalization of the learned architecture on real-world data

    - Setup a learning environment for synthetic data generation using NVIDIA Omniverse / Isaac Sim
    - Literature review on online learning techniques, such as automatic domain randomization, and recent advances in implicit representations, such as neural fields
    - Designing a method to perform 3D shape estimation using only simulated data
    - Show generalization of the learned architecture on real-world data

  • - Highly motivated and autonomous student - Prior knowledge and experience in 3D Vision - Programming experience in Python and PyTorch

    - Highly motivated and autonomous student
    - Prior knowledge and experience in 3D Vision
    - Programming experience in Python and PyTorch

  • While submitting your application, please include your CV including relevant experience and the latest transcript. - Mayank Mittal (mittalma@ethz.ch) - Vaishakh Patil (patilv@ethz.ch) - René Zurbrügg (zrene@ethz.ch)

    While submitting your application, please include your CV including relevant experience and the latest transcript.

    - Mayank Mittal (mittalma@ethz.ch)
    - Vaishakh Patil (patilv@ethz.ch)
    - René Zurbrügg (zrene@ethz.ch)

  • Not specified

  • Not specified

Calendar

Earliest start2023-09-01
Latest endNo date

Location

Robotic Systems Lab (ETHZ)

Other involved organizations
ETH Competence Center - ETH AI Center (ETHZ)

Labels

Master Thesis

Topics

  • Information, Computing and Communication Sciences
  • Behavioural and Cognitive Sciences
SiROP PARTNER INSTITUTIONS