Technology Engineer - AI/ML Platform at GSSTech Group
Bengaluru, karnataka, India -
Full Time


Start Date

Immediate

Expiry Date

28 Jun, 26

Salary

0.0

Posted On

30 Mar, 26

Experience

5 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Kubernetes, OpenShift, MLOps, GenAI, LLM, Triton Inference Server, vLLM, Prometheus, Grafana, Jenkins, Groovy, Ansible, Terraform, Python, AWS, SageMaker

Industry

IT Services and IT Consulting

Description
We are seeking a highly skilled Technology Engineer specializing in AI/ML Platforms, MLOps, and GenAI infrastructure to design, build, and scale next-generation AI systems. The ideal candidate will have strong experience in containerized environments, model serving, and cloud-based AI architecture, with a focus on performance, scalability, and resilience. Key Responsibilities Design, build, and maintain containerized applications using OpenShift, OpenShift AI, Kubernetes, and Helm Charts Deploy and optimize AI inference engines such as Triton Inference Server and vLLM for high-performance model serving Lead end-to-end model lifecycle management, including deployment, monitoring, scaling, and retraining workflows Implement monitoring, logging, and alerting systems using Prometheus and Grafana Collaborate on GenAI and LLM-based projects, including Agentic AI solutions Build and automate CI/CD pipelines using Jenkins, Groovy, Ansible, and Terraform Develop automation scripts and internal tools using Python Architect and manage AI/ML solutions on AWS, leveraging services like SageMaker and Bedrock (preferred) Build and enhance AI platforms across on-premise and cloud environments Ensure systems are highly scalable, fault-tolerant, and performance-optimized Contribute to architecture design, platform roadmap, and strategic technical decisions Required Skills & Qualifications Bachelor’s or Master’s degree in Computer Science, Engineering, or related field Strong hands-on experience with: Kubernetes / OpenShift ecosystem MLOps and AI/ML deployment pipelines Inference optimization (TensorRT / ONNX / Triton / vLLM) Experience with CI/CD tools (Jenkins, Groovy, Ansible, Terraform) Proficiency in Python scripting and automation Experience with monitoring tools like Prometheus and Grafana Solid understanding of distributed systems, microservices, and cloud-native architecture Hands-on experience with AWS Cloud services (SageMaker, Bedrock preferred) Preferred Qualifications Experience working on GenAI / LLM / Agentic AI use cases Knowledge of GPU acceleration and performance tuning Exposure to hybrid cloud (on-prem + cloud) AI platforms Familiarity with enterprise-scale AI platform engineering What We Offer Opportunity to work on cutting-edge AI/GenAI platforms Exposure to large-scale enterprise AI deployments Collaborative and innovation-driven engineering environment Competitive compensation and growth opportunities
Responsibilities
The role involves designing, building, and maintaining containerized AI/ML platforms using technologies like OpenShift and Kubernetes, focusing on deploying and optimizing high-performance inference engines such as Triton and vLLM. Key tasks include managing the end-to-end model lifecycle, implementing robust monitoring, and developing CI/CD pipelines for scalable AI solutions.
Loading...