Data Engineer III

JPMorgan Chase JPMorgan Chase · Banking · LONDON, United Kingdom · Corporate Sector

Data Engineer III at JPMorgan Chase in London, UK, focused on building and maintaining cloud-native data platforms and pipelines for a digital investing experience. The role involves working with lakehouse, warehousing, and streaming technologies to support analytics and regulatory reporting.

What you'd actually do

  1. Build and maintain scalable, reusable data processing and data quality frameworks using Python, PySpark, and dbt
  2. Build and operate batch and streaming data pipelines with strong scalability, performance, and fault tolerance
  3. Develop and manage workflow orchestration using tools such as Apache Airflow to support reliable, observable, and well-scheduled data movement and transformations
  4. Implement and optimize data models and warehouse structures to support analytics and business intelligence workloads
  5. Write clean, testable Python/PySpark code using object-oriented principles and unit testing

Skills

Required

  • Degree in Computer Science or a STEM-related field (or equivalent)
  • Experience working in an agile and dynamic environment
  • Experience across the software development lifecycle (requirements, design, architecture, development, testing, deployment, release, and support)
  • Hands-on experience with major cloud technologies (e.g., AWS, Google Cloud, or Azure)
  • Experience writing Python using object-oriented programming and unit/integration testing practices
  • Experience with SQL and familiarity with SQL-based workflow management tools such as dbt
  • Experience with orchestration tools such as Airflow (or similar)
  • Understanding of messaging/streaming systems such as Kafka or Pub/Sub (or similar)
  • Familiarity with infrastructure-as-code (e.g., Terraform) for cloud-based data infrastructure

Nice to have

  • Data modeling skills
  • Experience with data streaming and scalable processing frameworks (e.g., Spark, Flink, Beam, or similar)
  • Experience automating deployment, releases, and testing in continuous integration and continuous delivery pipelines
  • Experience with lakehouse patterns and table formats (e.g., Apache Iceberg)
  • Experience with federated query engines such as Trino
  • Experience designing automated tests (unit, component, integration, and end-to-end), including use of mocking frameworks
  • Experience with containers and container-based deployment environments (e.g., Docker, Kubernetes, or similar)

What the JD emphasized

  • 5 years of recent, hands-on professional experience actively coding as a data engineer