Data Engineer Ii, Transportation Execution, Speed Team

Amazon Amazon · Big Tech · Bellevue, WA · Business Intelligence

Data Engineer II role focused on building and operating scalable data pipelines and infrastructure for Amazon's transportation logistics. Responsibilities include ETL/ELT development, Redshift cluster management, technical design, and collaboration with data scientists and business teams. The role emphasizes data reliability, efficiency, and optimization for reporting, analysis, and machine learning workloads.

What you'd actually do

  1. Own end-to-end design, development, and operation of ETL/ELT pipelines that extract, transform, and load data from diverse sources using SQL, Python, and AWS big data technologies
  2. Manage and optimize multiple production Redshift clusters, including performance tuning, capacity planning, and cost optimization to support transportation org reporting needs
  3. Lead technical design discussions with Product teams, Data Scientists, Software Developers, and Business Intelligence Engineers to define data infrastructure requirements and deliver scalable solutions
  4. Define and enforce data engineering best practices for your domain, including code quality standards, testing frameworks, documentation, and deployment processes
  5. Conduct thorough code reviews and mentor junior data engineers on technical problem-solving, coding standards, and AWS best practices

Skills

Required

  • 3+ years of data engineering experience
  • Experience with distributed systems as it pertains to data storage and computing
  • Bachelor's degree in Computer Science, Engineering, Mathematics, or a related field
  • 2+ years of experience writing production data pipelines using SQL and Python
  • Experience designing and implementing ETL/ELT solutions with large-scale data processing

Nice to have

  • Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions
  • Experience with AWS services including S3, Redshift, Sagemaker, EMR, Kinesis, Lambda, and EC2
  • Experience with non-relational databases / data stores (object storage, document or key-value stores, graph databases, column-family databases)
  • Experience in mentoring, leading, or managing more junior engineers
  • Master's degree in Engineering, Computer Science, or a related field
  • Experience with data orchestration frameworks such as Apache Airflow, AWS Step Functions, or Glue Workflows
  • Experience with infrastructure-as-code tools (CloudFormation, Terraform, or CDK)