Vision Based Localization Specialist at Auterion
München, Bayern, Germany -
Full Time


Start Date

Immediate

Expiry Date

06 Nov, 24

Salary

0.0

Posted On

08 Aug, 24

Experience

0 year(s) or above

Remote Job

No

Telecommute

No

Sponsor Visa

No

Skills

Camera Calibration, Robotics, Opencv, Matlab, Localization, Communication Skills, Python, Computer Science, C++, Engineers, Navigation Systems, Programming Languages

Industry

Information Technology/IT

Description

We are seeking a highly skilled and experienced Specialist for vision based navigation systems with a proven track record of implementing and deploying vision based navigation systems on robotics systems such as UAS in real-life scenarios. The ideal candidate will have hands-on experience in open-source visual inertial-based localization, visual SLAM (Simultaneous Localization and Mapping), and expertise in the underlying optimization frameworks. You will play a pivotal role in developing and implementing state-of-the-art visual inertial navigation algorithms and techniques, enabling drones to navigate autonomously and accurately in a variety of real-world applications. This is a hands-on position, where we expect you to deploy vision based navigation algorithms on UAS, that are actively deployed with customers, in short iterations in order to create a robust and reliable system.
If you love autonomous robotics, solving challenging real world problems and delivering high-quality experiences, we want to talk with you!

QUALIFICATIONS AND SKILLS:

  • Master or Ph.D. degree in Computer Science, Electrical Engineering, Robotics, or a related field.
  • Strong hands-on experience in developing and implementing visual inertial navigation systems for drones, with a focus on open-source visual inertial-based localization and visual SLAM frameworks.
  • Deep understanding of computer vision techniques, including feature detection and tracking/matching, bundle adjustment, optical flow, depth estimation and camera calibration.
  • Excellent knowledge of sensor fusion techniques, including integrating visual and inertial data for accurate pose estimation and localization.
  • Expertise in optimization frameworks for visual inertial navigation systems.
  • Proficiency in programming languages such as C++, Python, or MATLAB, and thorough experience with ROS2 and OpenCV.
  • Track record of having implemented and deployed visual inertial navigation systems on robotic platforms in real applications.
  • Strong analytical and problem-solving skills, with the ability to design and evaluate complex algorithms in real-world scenarios.
  • Effective communication skills and the ability to work collaboratively in multidisciplinary teams, interacting with researchers, engineers, and external stakeholders.
Responsibilities
  • Develop and optimize visual inertial navigation algorithms for drones, focusing on open-source visual inertial-based localization and visual SLAM techniques.
  • Implement and customize visual inertial navigation frameworks, ensuring seamless integration with existing drone systems and hardware.
  • Conduct extensive testing and validation of navigation algorithms in real-world scenarios, fine-tuning parameters and optimizing performance for robust and accurate navigation.
  • Evaluate and select appropriate sensors, cameras, IMUs, and other hardware components for visual inertial navigation systems.
  • Develop and maintain calibration and synchronization procedures for visual inertial sensor setups, ensuring accurate data fusion and pose estimation.
  • Develop repeatable testing methods to evaluate the performance and accuracy of the system for each iteration.
  • Collaborate with computer vision and robotics experts to leverage the latest advancements in the field and integrate them into visual inertial navigation systems.
  • Document and present research findings, technical specifications, and implementation details to internal and external stakeholders.
  • Provide technical guidance and support to cross-functional teams during the implementation and deployment of visual inertial navigation systems in real applications.
Loading...