Research Scientist Intern, Robotics Dexterous Manipulation (PhD) at Meta
Redmond, Washington, United States -
Full Time


Start Date

Immediate

Expiry Date

25 Apr, 26

Salary

12134.0

Posted On

25 Jan, 26

Experience

0 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Robotics, Machine Learning, C/C++, Python, Deep Learning, Control Theory, Optimization Algorithms, Computer Vision, Dexterous Manipulation, Multimodal Learning, Reinforcement Learning, Imitation Learning, Robotics Frameworks, Embedded Systems, Tactile Sensing, Collision Avoidance

Industry

Software Development

Description
Meta Reality Labs Research is seeking a Robotics Research Scientist to drive foundational and applied research aimed at advancing state-of-the-art robotic architectures, dynamic in-hand manipulation, generalizable dexterous and functional manipulation, as well as tactile and multimodal sensing and perception. In this position, you will collaborate closely with a multidisciplinary team of researchers and engineers to push the boundaries of robotics science and technology. Our internships are twelve (12) to twenty-four (24) weeks long and we have various start dates throughout the year. Responsibilities Explore forward-looking ways and novel robotic architectures to advance dynamic in-hand manipulation, generalizable dexterous and functional manipulation (rigid to deformable objects in-hand to against the environment) Research novel sensors, robotic systems architectures, planning and control algorithms that involve tactile and multimodal vision and perception, dexterous manipulation and collision avoidance Develop machine learning models and end-to-end applications for robotics, dexterous manipulation, and contextual understanding Invent novel paradigms for robotics sensing and intelligence, leveraging a variety of sensing and machine perception modalities (images, video, audio, tactile, etc.) Develop efficient models deployable to robotic systems, integrate research into full system-level robotic prototypes Minimum Qualifications Currently has, or is in the process of obtaining a PhD degree in Computer Science, Electrical Engineering, Machine Learning, Robotics or related field Research experience in at least one of the following research areas: robot learning, dexterous manipulation, control theory, optimization algorithms, representation learning, self-supervised learning, multimodal learning, vision-language-action (VLA) models, reinforcement learning, imitation learning, robotics policy development, computer vision, egocentric perception, embodied AI and/or LLMs Experience in C/C++ and Python and deep learning frameworks (e.g., PyTorch, TensorFlow) Must obtain work authorization in country of employment at the time of hire, and maintain ongoing work authorization during employment Preferred Qualifications Proven track record of achieving significant results as demonstrated by grants, fellowships, patents, as well as first-authored publications at leading workshops or conferences such as Robotics (RSS, ICRA, IROS, CoRL, T-RO, IJRR), Machine Learning (NeurIPS, ICML, ICLR, AAAI, JMLR), and Computer Vision (CVPR, ICCV, ECCV, TPAMI) Experience with research and development of humanoid robots and dexterous robot hands Experience with embedded systems and communication protocols Experience with robotics frameworks such as Robot Operation System (ROS), along with experience working with robotic simulations and real-world robots Experience bringing-up and debugging prototype/scientific software-hardware systems (e.g., robotics platforms, multi-camera sensing/tracking solutions, wearable sensing systems) Experience working and communicating cross-functionally in a team environment Intent to return to degree program after the completion of the internship/co-op $7,313/month to $12,134/month + benefits
Responsibilities
The intern will explore novel robotic architectures and develop machine learning models for dexterous manipulation. They will collaborate with a multidisciplinary team to advance robotics science and technology.
Loading...