companyaccellor logo

Senior MLOps Engineer

accellorHyderabad, Telangana, India
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Experience Level

Senior

Qualifications

Requirements:Bachelor’s degree in Computer Science, Engineering, or a related discipline.5+ years of experience in DevOps, Platform Engineering, or MLOps roles with 1-2+ years dedicated to ML/AI infrastructure. Expert programming skills in Python; familiarity with Bash, Go, or Java is advantageous. Hands-on experience with ML pipeline orchestration tools such as Kubeflow, MLflow, Airflow, or Vertex AI Pipelines. Proficient in containerization (Docker) and orchestration (Kubernetes, Helm). Experience with cloud-native ML services on AWS (SageMaker), Azure (Azure ML), or GCP (Vertex AI). Familiarity with model serving frameworks like TorchServe, Triton Inference Server, vLLM, or TGI. Knowledge of Infrastructure as Code tools (Terraform, Pulumi, or CloudFormation). Experience with monitoring and observability tools (Prometheus, Grafana, Datadog, or similar). Solid understanding of software engineering principles, version control (Git), and CI/CD methodologies.

About the job

Join our dynamic team at accellor as a Senior MLOps Engineer, where you will play a crucial role in architecting, building, and sustaining the infrastructure and pipelines essential for operationalizing AI and Machine Learning systems at scale. This position serves as a vital link between model development and production deployment, ensuring that ML and GenAI workloads are dependable, observable, cost-effective, and continuously optimized across enterprise environments.

Key Responsibilities

  • Design and execute comprehensive ML pipelines involving data ingestion, feature engineering, model training, evaluation, and deployment.
  • Develop and oversee CI/CD pipelines for ML models, incorporating automated testing, validation, and rollback strategies.
  • Architect and sustain model serving infrastructure for real-time and batch inference workloads, including deployments of LLM and agentic AI.
  • Implement systems for model monitoring, drift detection, and alerting to maintain production model reliability and health.
  • Oversee experiment tracking, model versioning, and artifact registries to facilitate reproducibility and governance.
  • Optimize computing costs and inference latency for both GPU and CPU workloads on cloud platforms (AWS, Azure, or GCP).
  • Utilize Docker and Kubernetes for the containerization and orchestration of ML workloads.
  • Automate data pipeline workflows and manage feature stores for training and inference processes.
  • Collaborate with AI Engineers, Data Scientists, and Platform teams to streamline the transition from prototype to production.
  • Establish and uphold MLOps best practices, standards, and documentation across the engineering team.

About accellor

At accellor, we are committed to leveraging cutting-edge technologies to drive innovation and transformation in the AI and machine learning domains. Our team thrives on collaboration, creativity, and delivering impactful solutions that empower businesses to achieve their strategic goals.

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.