Senior Technology Engineer - AI/ML Platforms at GSSTech Group
Bengaluru, karnataka, India -
Full Time


Start Date

Immediate

Expiry Date

28 Jun, 26

Salary

0.0

Posted On

30 Mar, 26

Experience

5 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

AI/ML Platforms, MLOps, Kubernetes, OpenShift, Inference Server, LLM, GenAI, Triton, vLLM, Prometheus, Grafana, Jenkins, Ansible, Terraform, Python, AWS

Industry

IT Services and IT Consulting

Description
We are seeking a highly skilled Senior Technology Engineer with strong expertise in AI/ML platforms, MLOps, and scalable inference systems. The ideal candidate will play a critical role in designing, deploying, and optimizing AI-driven solutions across on-premises and cloud environments, with a focus on GenAI and LLM-based applications. Key Responsibilities Design, build, and maintain containerized applications using OpenShift, OpenShift AI, Kubernetes, and Helm Charts. Integrate and optimize AI inference engines such as Triton Inference Server and vLLM for high-performance model serving. Lead end-to-end model lifecycle management, including deployment, monitoring, scaling, and maintenance in production environments. Implement robust monitoring and alerting frameworks using Prometheus and Grafana. Collaborate on GenAI and Large Language Model (LLM) initiatives, including Agentic AI systems. Develop and manage CI/CD pipelines using Jenkins, Ansible, Groovy, and Terraform. Build automation tools and scripts using Python to improve system efficiency and reliability. Architect and manage AI/ML solutions on AWS Cloud, leveraging services such as Amazon SageMaker and AWS Bedrock (preferred). Design and enhance AI platforms across hybrid environments (on-premise and cloud). Ensure systems are scalable, resilient, and high-performing. Contribute to architecture design decisions and define the future roadmap of AI platform capabilities. Required Skills & Experience Strong experience in MLOps, AI/ML platform engineering, or related roles. Expertise in container orchestration using Kubernetes/OpenShift. Hands-on experience with AI inference optimization (TensorRT, ONNX Runtime, Triton, vLLM). Proficiency in cloud platforms (AWS preferred). Experience with CI/CD and Infrastructure as Code tools (Jenkins, Terraform, Ansible). Strong programming skills in Python. Knowledge of monitoring tools like Prometheus and Grafana. Experience working with LLMs, GenAI, or AI-based products is highly desirable. Preferred Qualifications Experience with GPU acceleration and performance tuning. Exposure to Agentic AI frameworks and LLM orchestration tools. Familiarity with hybrid cloud architectures. Soft Skills Strong problem-solving and analytical thinking Ability to work in fast-paced, collaborative environments Excellent communication and stakeholder management skills
Responsibilities
The role involves designing, building, and maintaining containerized AI/ML platforms using technologies like OpenShift and Kubernetes, focusing on optimizing high-performance inference systems for GenAI and LLM applications. Key duties include leading end-to-end model lifecycle management, implementing robust monitoring, and developing CI/CD pipelines.
Loading...