Member of Technical Staff, MM Inference at Microsoft
Redmond, Washington, United States -
Full Time


Start Date

Immediate

Expiry Date

24 Feb, 26

Salary

0.0

Posted On

26 Nov, 25

Experience

10 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

C, C++, C#, Java, JavaScript, Python, Generative AI, Distributed Computing, Technical Project Leadership, Architectural Decisions, Model Deployment, Inference Engine Development, Performance Tuning, Benchmarking, Data Gathering, Product Development

Industry

Software Development

Description
Develop and maintain the inference engine for multimodal generative models. Develop and maintain the model deployment pipeline for various product lines. Benchmark, profile and tune the model inference performance with model and hardware specific techniques. Gather data and insights to develop the multimodal inference roadmap. Find a path to get things done despite roadblocks to get your work into the hands of users quickly and iteratively. Enjoy working in a fast-paced, design-driven, product development cycle. Embody our Culture and Values. Bachelor's Degree in Computer Science or related technical field AND 6+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python Master's Degree in Computer Science or related technical field AND 8+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR Bachelor's Degree in Computer Science or related technical field AND 12+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR equivalent experience. Experience in leading technical projects and supporting architectural decisions. Experience with generative AI. Experience with distributed computing.
Responsibilities
Develop and maintain the inference engine for multimodal generative models and the model deployment pipeline for various product lines. Benchmark, profile, and tune model inference performance while gathering insights to develop the multimodal inference roadmap.
Loading...