Overview

We are looking for a Lead Data Software Engineer to drive scalable data solutions and guide end-to-end technical delivery with clients. You will partner directly with stakeholders, shape robust batch and streaming pipelines, and set a high bar for architecture and quality across the program—apply now.

Responsibilities

  • Develop and maintain the codebase for ETL and ELT pipelines, large batch/micro batch processing, and streaming systems
  • Build the infrastructure needed for efficient extraction, transformation, and loading from diverse data sources using ADF, Spark, Kafka, or similar technologies
  • Identify, design, and implement internal process improvements, including automating manual steps, optimizing data delivery, and reworking infrastructure for better scalability
  • Design and implement modern data services solutions using Spring Boot, ReactJS, NoSQL, or other UI and API related technologies
  • Ensure governance of delivery management and production processes according to the selected delivery model
  • Serve as the single accountable owner for top management and stakeholders on delivery topics, including escalations, upsells, and ramp-downs
  • Provide technical leadership across delivery by planning sound, future-proof architecture and ensuring implementation meets technical quality standards
  • Write stories and corresponding acceptance criteria to support agile/scrum execution
  • Coordinate work across multiple disciplines and stakeholders
  • Ensure projects are delivered in line with defined processes and methodologies, with a strong focus on agile approaches
  • Establish a continuous delivery risk management strategy that enables proactive decisions and actions throughout the delivery life cycle
  • Measure and improve delivery productivity, act as a consultant to Data Engineers, and perform production support and deployment activities

Requirements

  • 5+ years of experience in data software engineering
  • High proficiency with SQL, Spark, and Scala
  • Deep expertise in Databricks
  • Solid understanding of ETL/ELT pipelines, batch and streaming data processing
  • Working knowledge of delivery management, agile methodologies, and technical leadership
  • Ability to communicate directly with clients in English at B2 (Upper-Intermediate) level or higher

Nice to have

  • Knowledge of Spark Streaming
  • Familiarity with Kafka
  • Background in Azure

[GTS] Benefits (generic, except India)

  • International projects with top brands
  • Work with global teams of highly skilled, diverse peers
  • Healthcare benefits
  • Employee financial programs
  • Paid time off and sick leave
  • Upskilling, reskilling and certification courses
  • Unlimited access to the LinkedIn Learning library and 22,000+ courses
  • Global career opportunities
  • Volunteer and community involvement opportunities
  • EPAM Employee Groups
  • Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn