Overview

We are looking for a highly skilled and experienced Data Software Engineer to join our team and contribute to developing data-centric applications.

You will collaborate with state-of-the-art Big Data tools, cloud technologies, and cross-functional teams to design innovative solutions for complex business challenges.

Responsibilities

  • Build and enhance data software applications utilized by Data Integration Engineers
  • Create and deploy sophisticated analytical solutions with Spark, PySpark, NoSQL, and other Big Data tools
  • Integrate advanced features using cloud platforms like AWS to optimize data workflows
  • Partner with product and engineering teams to gather insights and support decision-making
  • Coordinate with architects, technical leads, and other teams to ensure unified solutions
  • Assess business needs and technical environments to deliver well-suited technical implementations
  • Conduct code reviews to promote best practices and ensure high-quality code
  • Validate and test solutions against functional, technical, and performance standards
  • Document projects comprehensively to facilitate future reference and development
  • Engage with clients to understand their needs and provide expert technical guidance

Requirements

  • Degree in Computer Science, Software Engineering, or a related field at the bachelor’s or master’s level
  • Background in Data Software Engineering, with 2+ years of experience in Big Data technologies
  • Knowledge of data engineering concepts, including data management, storage, visualization, operations, and security
  • Strong understanding of data ingestion pipelines, Data Warehousing, and Data Lakes
  • Background in programming languages such as Python, Java, Scala, or Kotlin
  • Expertise in SQL and NoSQL databases
  • Proficiency in Big Data tools, particularly Spark and PySpark
  • Capability to design and deploy solutions on AWS cloud platforms, including Glue and RedShift
  • Understanding of CI/CD pipelines and workflows for integration and deployment
  • Familiarity with containerization and tools like Docker, Kubernetes, and Yarn
  • Skills in utilizing Databricks for advanced data analytics and engineering
  • Advanced proficiency in English, both written and spoken, at a B2 level or higher

Nice to have

  • Familiarity with additional Big Data tools, including Hadoop, Hive, and Flink
  • Understanding of SDLC methodologies, focusing on Agile practices
  • Competency in implementing and managing the software development lifecycle effectively

[GTS] Benefits (generic, except India)

  • International projects with top brands
  • Work with global teams of highly skilled, diverse peers
  • Healthcare benefits
  • Employee financial programs
  • Paid time off and sick leave
  • Upskilling, reskilling and certification courses
  • Unlimited access to the LinkedIn Learning library and 22,000+ courses
  • Global career opportunities
  • Volunteer and community involvement opportunities
  • EPAM Employee Groups
  • Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn