We are looking for a highly skilled Senior ML / Data Pipeline Engineer who can translate complex machine learning and multimodal concepts into scalable, production-ready pipelines and workflows.

Requirements

  • 5+ years of experience in data engineering, ML pipelines, or distributed systems.
  • Strong experience building scalable data pipelines for large datasets (video/audio preferred).
  • Hands-on experience with cloud platforms (AWS, Azure, or GCP).
  • Experience working with GPU-based environments and distributed computing.
  • Strong programming skills in Python, Scala, or similar languages.
  • Experience with data processing frameworks (Spark, Ray, Kafka, Airflow, or similar).
  • Understanding of ML workflows, training pipelines, and inference systems.
  • Experience designing fault-tolerant, high-availability systems.
  • Strong knowledge of data storage systems (data lakes, object storage, distributed file systems).
  • Ability to handle high-throughput, large-scale data ingestion and processing.