Overview

Our client is developing an AI-powered analytics layer on top of a network security data platform.

As a Senior Data Engineer on this engagement, you will design and deliver the core data infrastructure to derive security policies from operational technology network logs.

You will work with diverse, complex data exports from industrial security platforms - such as network monitoring systems - and transform them into a well-structured DynamoDB database. Once the data foundation is in place, you will connect it to AWS QuickSight to deliver interactive BI dashboards that give stakeholders immediate visibility into network behavior, security events, and policy gaps.

Responsibilities

  • Design and implement DynamoDB databases to serve as the main analytical store for network and security log data
  • Define table structures, partition keys, sort keys, and Global Secondary Indexes for high-volume, time-series event data
  • Optimize database performance and cost efficiency by leveraging DynamoDB features and establishing data archival strategies
  • Build robust ETL pipelines to ingest data from Excel, CSV, and API feeds from network-monitoring tools
  • Cleanse, validate, normalize, and enrich raw security-event data, handling inconsistent schemas and missing values
  • Automate orchestration, scheduling, and error handling to ensure reliable pipeline execution and timely data delivery
  • Connect DynamoDB to AWS QuickSight to create interactive dashboards and reports for security events and network traffic patterns
  • Design and maintain QuickSight datasets, calculated fields, and visual analyses, iterating on dashboard design based on feedback
  • Collaborate with cybersecurity experts, AI/Analytics Engineers, AI Architects, and stakeholders to translate business requirements into data architecture
  • Produce clear technical documentation including schema definitions, data dictionaries, pipeline specifications, and operational runbooks

Requirements

  • Minimum 3 years of experience in data engineering roles
  • Advanced knowledge of DynamoDB including table design, partition and sort key strategies, GSIs, capacity planning, Streams, and NoSQL data modeling for high-throughput workloads
  • Proven experience building production-grade ETL/ELT pipelines in Python using libraries such as boto3 and pandas, capable of handling messy, multi-format source data and loading into DynamoDB
  • Proficiency with AWS services including S3, Lambda, IAM, and CloudWatch, and their integration with DynamoDB and QuickSight
  • Experience with AWS QuickSight for building datasets, analyses, and interactive dashboards, connecting to AWS data sources, and designing visualizations for non-technical audiences
  • Expertise in NoSQL data modeling, designing denormalized, query-driven data models for key-value and document-oriented workloads
  • Strong background in data quality engineering, implementing validation frameworks, data contracts, and automated testing for pipelines
  • Proficiency in Python for writing clean, maintainable ETL code, including familiarity with virtual environments, testing, and version control using Git
  • Ability to document technical decisions clearly and collaborate effectively with non-technical stakeholders
  • Excellent English communication skills at B2+ level or higher, both written and spoken

Nice to have

  • Experience with AWS Glue and Athena for serverless ETL and ad-hoc querying of S3-based data lakes
  • Familiarity with graph databases such as Neo4j or Amazon Neptune and modeling data as nodes and relationships for representing network topologies and policy structures
  • Experience with CI/CD tools like GitHub Actions or GitLab CI for automated pipeline testing and deployment
  • Background in network-security logs, cybersecurity analytics, or security domain environments
  • Understanding of process mining, including event logs and structures that feed process-mining algorithms
  • Experience with AI-assisted development tools or workflows

[GTS] Benefits (generic, except India)

  • International projects with top brands
  • Work with global teams of highly skilled, diverse peers
  • Healthcare benefits
  • Employee financial programs
  • Paid time off and sick leave
  • Upskilling, reskilling and certification courses
  • Unlimited access to the LinkedIn Learning library and 22,000+ courses
  • Global career opportunities
  • Volunteer and community involvement opportunities
  • EPAM Employee Groups
  • Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn