Overview
We are looking for a highly skilled and experienced Data Software Engineer to join our team and contribute to developing data-centric applications.
You will collaborate with state-of-the-art Big Data tools, cloud technologies, and cross-functional teams to design innovative solutions for complex business challenges.
Responsibilities
- Build and enhance data software applications utilized by Data Integration Engineers
- Create and deploy sophisticated analytical solutions with Spark, PySpark, NoSQL, and other Big Data tools
- Integrate advanced features using cloud platforms like AWS to optimize data workflows
- Partner with product and engineering teams to gather insights and support decision-making
- Coordinate with architects, technical leads, and other teams to ensure unified solutions
- Assess business needs and technical environments to deliver well-suited technical implementations
- Conduct code reviews to promote best practices and ensure high-quality code
- Validate and test solutions against functional, technical, and performance standards
- Document projects comprehensively to facilitate future reference and development
- Engage with clients to understand their needs and provide expert technical guidance
Requirements
- Degree in Computer Science, Software Engineering, or a related field at the bachelor’s or master’s level
- Background in Data Software Engineering, with 2+ years of experience in Big Data technologies
- Knowledge of data engineering concepts, including data management, storage, visualization, operations, and security
- Strong understanding of data ingestion pipelines, Data Warehousing, and Data Lakes
- Background in programming languages such as Python, Java, Scala, or Kotlin
- Expertise in SQL and NoSQL databases
- Proficiency in Big Data tools, particularly Spark and PySpark
- Capability to design and deploy solutions on AWS cloud platforms, including Glue and RedShift
- Understanding of CI/CD pipelines and workflows for integration and deployment
- Familiarity with containerization and tools like Docker, Kubernetes, and Yarn
- Skills in utilizing Databricks for advanced data analytics and engineering
- Advanced proficiency in English, both written and spoken, at a B2 level or higher
Nice to have
- Familiarity with additional Big Data tools, including Hadoop, Hive, and Flink
- Understanding of SDLC methodologies, focusing on Agile practices
- Competency in implementing and managing the software development lifecycle effectively
[GTS] Benefits (generic, except India)
- International projects with top brands
- Work with global teams of highly skilled, diverse peers
- Healthcare benefits
- Employee financial programs
- Paid time off and sick leave
- Upskilling, reskilling and certification courses
- Unlimited access to the LinkedIn Learning library and 22,000+ courses
- Global career opportunities
- Volunteer and community involvement opportunities
- EPAM Employee Groups
- Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn