Principal Solution Architect - AI Infrastructure at Toyota
Plano, TX 75024, USA -
Full Time


Start Date

Immediate

Expiry Date

07 Nov, 25

Salary

0.0

Posted On

10 Aug, 25

Experience

10 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Good communication skills

Industry

Information Technology/IT

Description

WHO WE ARE

Collaborative. Respectful. A place to dream and do. These are just a few words that describe what life is like at Toyota. As one of the world’s most admired brands, Toyota is growing and leading the future of mobility through innovative, high-quality solutions designed to enhance lives and delight those we serve. We’re looking for talented team members who want to Dream. Do. Grow. with us.
An important part of the Toyota family is Toyota Financial Services (TFS), the finance and insurance brand for Toyota and Lexus in North America. While TFS is a separate business entity, it is an essential part of this world-changing company- delivering on Toyota’s vision to move people beyond what’s possible. At TFS, you will help create best-in-class customer experience in an innovative, collaborative environment.
To save time applying, Toyota does not offer sponsorship of job applicants for employment-based visas or any other work authorization for this position at this time.

WHO WE’RE LOOKING FOR

Toyota Financial Services is seeking a Principal Solution Architect - AI Infrastructure to design the systems and platforms that enable enterprise-scale MLOps and LLMOps. This is a senior-level individual contributor role within the Architecture organization, focused on building the foundational infrastructure for deploying, managing, and scaling AI and GenAI workloads in production.
You’ll work across cloud infrastructure, platform engineering, data, and cybersecurity teams to architect robust, secure, and performant environments that power model training, inference, orchestration, and retrieval-augmented generation (RAG) systems.
This role is ideal for architects who deeply understand infrastructure, cloud-native systems, and the unique demands of production-scale AI workloads.

Responsibilities
  • Architect and evolve cloud-native infrastructure to support AI/ML and LLM workloads in production
  • Build platform capabilities for MLOps and LLMOps—including model training, versioning, deployment, monitoring, and rollback
  • Enable GPU-accelerated compute environments optimized for model performance, scalability, and cost-efficiency
  • Integrate and standardize infrastructure for vector databases, model registries, and orchestration frameworks
  • Establish reusable patterns for model serving, inference scaling, prompt management, and latency optimization
  • Design secure, multi-tenant environments that enforce access controls, auditability, and usage limits for AI models
  • Partner with engineering, platform, and data teams to ensure seamless data flow, observability, and operational resiliency
  • Contribute architecture documentation, governance models, and standards to support AI infrastructure delivery across teams
  • Stay informed on emerging technologies in GenAI, distributed systems, and infrastructure trends
Loading...