Senior DevOps Engineer - AI Services

at  ANACONDA INC

Remote, Oregon, USA -

Start DateExpiry DateSalaryPosted OnExperienceSkillsTelecommuteSponsor Visa
Immediate06 Feb, 2025Not Specified07 Nov, 20244 year(s) or aboveGood communication skillsNoNo
Add to Wishlist Apply All Jobs
Required Visa Status:
CitizenGC
US CitizenStudent Visa
H1BCPT
OPTH4 Spouse of H1B
GC Green Card
Employment Type:
Full TimePart Time
PermanentIndependent - 1099
Contract – W2C2H Independent
C2H W2Contract – Corp 2 Corp
Contract to Hire – Corp 2 Corp

Description:

SUMMARY:

Anaconda is seeking a talented Senior Devops Engineer to join our rapidly growing company. This is an excellent opportunity for you to leverage your experience and skills and apply it to the world of data science, artificial intelligence, and machine learning.

What You’ll Do:

  • Design and implement scalable AWS infrastructure, with particular focus on Lambda functions, RDS, and message bus architectures
  • Build and maintain robust MLOps pipelines for deploying and monitoring LLM models in production environments
  • Develop and optimize real-time communication systems using WebSockets and WebRTC for ML inference services
  • Create and maintain Python packages with C extensions, focusing on performance optimization and reliability
  • Design and implement comprehensive monitoring and telemetry systems across our infrastructure
  • Manage and optimize Kubernetes clusters for ML workloads, ensuring efficient resource utilization and high availability
  • Architect and maintain efficient CI/CD pipelines for both infrastructure and application deployments
  • Collaborate with AI and research teams to understand and implement infrastructure requirements for new ML models and features
  • Optimize system performance and cost efficiency across our AWS infrastructure
  • Lead technical discussions and provide expertise in infrastructure and deployment strategies
  • Implement and maintain security best practices across our infrastructure
  • Participate in on-call rotations and lead incident response efforts when necessary

How To Apply:

Incase you would like to apply to this job directly from the source, please click here

Responsibilities:

  • Design and implement scalable AWS infrastructure, with particular focus on Lambda functions, RDS, and message bus architectures
  • Build and maintain robust MLOps pipelines for deploying and monitoring LLM models in production environments
  • Develop and optimize real-time communication systems using WebSockets and WebRTC for ML inference services
  • Create and maintain Python packages with C extensions, focusing on performance optimization and reliability
  • Design and implement comprehensive monitoring and telemetry systems across our infrastructure
  • Manage and optimize Kubernetes clusters for ML workloads, ensuring efficient resource utilization and high availability
  • Architect and maintain efficient CI/CD pipelines for both infrastructure and application deployments
  • Collaborate with AI and research teams to understand and implement infrastructure requirements for new ML models and features
  • Optimize system performance and cost efficiency across our AWS infrastructure
  • Lead technical discussions and provide expertise in infrastructure and deployment strategies
  • Implement and maintain security best practices across our infrastructure
  • Participate in on-call rotations and lead incident response efforts when necessar


REQUIREMENT SUMMARY

Min:4.0Max:7.0 year(s)

Information Technology/IT

IT Software - System Programming

Software Engineering

Graduate

Computer science engineering or related field

Proficient

1

Remote, USA