Member of Technical Staff - Platform Engineer (LLM Infrastructure & Backend at INFLECTION AI (UK) Limited
Palo Alto, California, United States -
Full Time


Start Date

Immediate

Expiry Date

14 Mar, 26

Salary

350000.0

Posted On

14 Dec, 25

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

Yes

Skills

Backend Engineering, Python, TypeScript, Node.js, Production ML Systems, APIs, FastAPI, Postgres, Redis, Kubernetes, React, Model Integration, Inference Orchestration, Data Flow, Clean Code, Collaboration, Infrastructure

Industry

technology;Information and Internet

Description
At Inflection AI, our public benefit mission is to harness the power of AI to improve human well-being and productivity. The next era of AI will be defined by agents we trust to act on our behalf. We’re pioneering this future with human-centered AI models that unite emotional intelligence (EQ) and raw intelligence (IQ)—transforming interactions from transactional to relational, to create enduring value for individuals and enterprises alike. Our work comes to life in two ways today: Pi, your personal AI, designed to be a kind and supportive companion that elevates everyday life with practical assistance and perspectives. Platform — large-language models (LLMs) and APIs that enable builders, agents, and enterprises to bring Pi-class emotional intelligence into experiences where empathy and human understanding matter most. We are building toward a future of AI agents that earn trust, deepen understanding, and create aligned, long-term value for all. About the Role We are seeking a Platform Engineer to join our team building backend infrastructure for new ML-powered enterprise products. This role is a unique opportunity to work at the intersection of backend engineering and machine learning systems, focusing on inference orchestration, model integration, and real-time deployment. The ideal candidate will have experience with backend development, production ML systems, and tools that scale enterprise-level applications. This is a good role for you if you: Backend engineering experience with Python, TypeScript, or Node.js. Hands-on experience working with production PyTorch models, model checkpoints, and inference logic. Strong knowledge of building APIs and services that are scalable, stable, and secure. Passion for bridging backend engineering and ML systems, especially at the infrastructure layer. Familiarity with tools such as FastAPI, Postgres, Redis, Kubernetes, and React. Desire to be hands-on and contribute to shaping the foundation of a new enterprise ML product. Have a bachelor’s degree or equivalent in a related field to the offered position requirements. Responsibilities include: Build and maintain backend services to support LLM integration, inference orchestration, and data flow. Write clean, reliable Python code for experimentation, model integration, and production systems. Collaborate closely with ML researchers to rapidly iterate on product ideas and deploy features. Design and implement infrastructure to handle scalable inference workloads and enterprise-level use cases. Own system components and ensure reliability, observability, and maintainability from day one. Have a bachelor’s degree or equivalent in a related field to the offered position requirements. Compensation & Benefits Salary Range: $175,000 – $350,000 USD per year (based on experience and location) Equity: Competitive stock options Benefits: Diverse medical, dental and vision options 401k matching program Unlimited paid time off Parental leave and flexibility for all parents and caregivers Support of country-specific visa needs for international employees living in the Bay Area
Responsibilities
Build and maintain backend services to support LLM integration and inference orchestration. Collaborate closely with ML researchers to rapidly iterate on product ideas and deploy features.
Loading...