Overview

We are looking for a detail-oriented and motivated Senior Systems Engineer with a strong focus on Data DevOps/MLOps to join our team.

The ideal candidate should possess a deep understanding of data engineering, automation of data pipelines, and integration of machine learning models into operational environments. This role is for a collaborative professional adept at building, deploying, and managing scalable data and ML pipelines aligned with strategic objectives.

Responsibilities

  • Design CI/CD pipelines for data integration and machine learning model deployment
  • Deploy and maintain infrastructure for data processing and model training using cloud services
  • Automate processes like data validation, transformation, and workflow orchestration
  • Coordinate with data scientists, software engineers, and product teams to integrate ML models into production environments
  • Enhance performance and reliability by optimizing model serving and monitoring processes
  • Ensure data versioning, lineage tracking, and reproducibility across ML experiments
  • Identify improvements for deployment processes, scalability, and infrastructure resilience
  • Implement security measures to safeguard data integrity and maintain compliance
  • Resolve issues in the data and ML pipeline lifecycle

Requirements

  • Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field
  • 5 or more years of experience in Data DevOps, MLOps, or related professions
  • Proficiency in cloud platforms such as Azure, AWS, or GCP
  • Background in Infrastructure as Code (IaC) tools like Terraform, CloudFormation, or Ansible
  • Expertise in containerization and orchestration tools such as Docker and Kubernetes
  • Skills in using data processing frameworks like Apache Spark or Databricks
  • Proficiency in Python, with familiarity with data manipulation and ML libraries such as Pandas, TensorFlow, or PyTorch
  • Familiarity with CI/CD tools like Jenkins, GitLab CI/CD, or GitHub Actions
  • Knowledge of version control systems, such as Git, and MLOps platforms like MLflow or Kubeflow
  • Understanding of monitoring, logging, and alerting systems like Prometheus or Grafana
  • Strong problem-solving abilities with the capability to work both independently and collaboratively
  • Effective communication and documentation skills

Nice to have

  • Familiarity with DataOps practices and tools like Airflow or dbt
  • Understanding of data governance frameworks and tools like Collibra
  • Knowledge of Big Data technologies such as Hadoop or Hive
  • Credentials in cloud platforms or data engineering activities

India

Opportunity to work on technical challenges that may impact across geographies

Vast opportunities for self-development: online university, knowledge sharing opportunities globally, learning opportunities through external certifications

Opportunity to share your ideas on international platforms

Sponsored Tech Talks & Hackathons

Unlimited access to LinkedIn learning solutions

Possibility to relocate to any EPAM office for short and long-term projects

Focused individual development

Benefit package:

  • Health benefits
  • Retirement benefits
  • Paid time off
  • Flexible benefits

Forums to explore beyond work passion (CSR, photography, painting, sports, etc.)

[epamgdo] India (About EPAM)

EPAM is a leading global provider of digital platform engineering and development services. We are committed to having a positive impact on our customers, our employees, and our communities. We embrace a dynamic and inclusive culture. Here you will collaborate with multi-national teams, contribute to a myriad of innovative projects that deliver the most creative and cutting-edge solutions, and have an opportunity to continuously learn and grow. No matter where you are located, you will join a dedicated, creative, and diverse community that will help you discover your fullest potential.