Assistant Manager - AI Research at WNS Global Services
Gurgaon, haryana, India -
Full Time


Start Date

Immediate

Expiry Date

18 May, 26

Salary

0.0

Posted On

17 Feb, 26

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

LLMOps, LLM Handling, Deployment, Monitoring, Kubernetes, Docker, Cloud, AWS, Azure, GCP, Inference Infrastructure, CI/CD Pipelines, RAG Systems, GenAI Security, MLOps, Evaluation Frameworks

Industry

Business Consulting and Services

Description
Company Description WNS (Holdings) Limited (NYSE: WNS), is a leading Business Process Management (BPM) company. We combine our deep industry knowledge with technology and analytics expertise to co-create innovative, digital-led transformational solutions with clients across 10 industries. We enable businesses in Travel, Insurance, Banking and Financial Services, Manufacturing, Retail and Consumer Packaged Goods, Shipping and Logistics, Healthcare, and Utilities to re-imagine their digital future and transform their outcomes with operational excellence.We deliver an entire spectrum of BPM services in finance and accounting, procurement, customer interaction services and human resources leveraging collaborative models that are tailored to address the unique business challenges of each client. We co-create and execute the future vision of 400+ clients with the help of our 44,000+ employees. Job Description Role Overview: We are hiring an LLMOps Engineer to join our AI Research team, a highly technical group working on cutting-edge advancements in the AI industry. The team focuses on building scalable, production-grade LLM systems, fine-tuning strategies, evaluation frameworks, and next-generation deployment architectures. This role requires hands-on experience operating LLMs beyond simple API integration. The ideal candidate understands the architectural, operational, and evaluation complexities that differentiate LLMOps from traditional MLOps. Responsibilities:- Manage the end-to-end lifecycle of LLMs: registry, packaging, versioning, deployment, monitoring, and rollback. Deploy and operate self-hosted / open-source LLMs (not limited to OpenAI API usage). Design and manage scalable inference infrastructure, including GPU-aware deployments. Implement CI/CD pipelines for LLM deployment and continuous evaluation. Monitor system performance including latency, throughput, token usage, cost, drift (model and data), and hallucinations. Ensure secure, compliant, and resilient cloud-based model deployments. Collaborate with research and engineering for deployments. Skills:- Strong hands-on experience with LLM handling, hosting, and operationalization. Clear understanding of how LLMOps differs from traditional MLOps (prompt management, non-deterministic outputs, semantic evaluation, token economics, guardrails etc.). Experience with Kubernetes, Docker, and containerized deployments. Cloud expertise (AWS / Azure / GCP) including compute, storage, IAM, networking, and monitoring. Experience building scalable inference and model-serving architectures. Familiarity with tools such as MLflow, Kubeflow etc. (good to have). Understanding of vector databases, RAG systems, and evaluation frameworks (preferred). Knowledge of GenAI security considerations (prompt injection, data leakage prevention). Qualifications Bachelor’s degree in Computer Science, Engineering, or related field.DevOps certification (e.g., AWS DevOps Engineer, Azure DevOps, or equivalent). 3–5 years of experience in MLOps, LLMOps, ML Engineering, or related roles. Bachelor’s or master’s degree in computer science, Artificial Intelligence, Data Science, or a related technical field. Demonstrated experience deploying ML/LLM systems in production environments.
Responsibilities
The role involves managing the complete lifecycle of Large Language Models (LLMs), including registry, versioning, deployment, monitoring, and rollback, with a focus on deploying and operating self-hosted or open-source LLMs.
Loading...