(usa) Senior, Data Engineer

Walmart Walmart · Retail · Bentonville, AR +1

Senior Data Engineer responsible for developing and maintaining end-to-end data solutions, including data ingestion pipelines, transformation workflows, and consumption-layer data products. Integrates diverse technologies and works with big data frameworks, orchestration tools, storage systems, and streaming platforms. Collaborates with Data Scientists, Product Managers, and Platform Engineers. Leverages AI coding agents and developer assistants.

What you'd actually do

  1. Develop and maintain scalable data pipelines and backend systems.
  2. Integrate and optimize data from multiple sources, ensuring quality and reliability.
  3. Implement and enhance data processing frameworks and tools.
  4. Design, manage, and govern data schemas throughout their lifecycle.
  5. Collaborate with cross-functional teams on data-driven initiatives.

Skills

Required

  • Bachelor's degree in Computer Science, Business Administration, Information Systems, or a related field.
  • 5+ years of experience in software engineering.
  • Proficiency in scripting and programming languages including Python, Java, Scala, and SQL.
  • Strong background in Java, multithreading, data structures, algorithms, and design patterns, with experience building high-performance, scalable applications.
  • Proficiency in Java, Spring, and Spring Boot, with hands-on experience building large-scale backend microservices.
  • Strong experience with SQL and NoSQL databases (e.g., Postgres, Cassandra, Cosmos) and distributed SQL engines such as Presto.
  • Required experience with Presto and Trino for distributed SQL query processing across large-scale data sources.
  • Hands-on experience with big data technologies (Hadoop, Apache Spark, Kafka) and workflow orchestration tools (Apache Airflow).
  • Expertise in designing, deploying, and managing scalable and secure solutions on cloud platforms, including Azure and GCP, with a focus on automation, monitoring, and optimization.
  • Strong software engineering practices, including version control and CI/CD (GitHub).
  • Experience with unit, functional, and integration testing using frameworks such as JUnit, Mockito, and Jest, along with code quality tools like Sonar.

Nice to have

  • BS or MS in Computer Science or a related field with 3+ years of experience building large-scale distributed systems.
  • Experience with RESTful APIs, OAuth, JWT, and API gateways.
  • Familiarity with in-memory processing and data formats such as Avro, Parquet, and JSON.
  • Experience using Kibana and Elasticsearch for monitoring, analyzing, and visualizing large-scale logs and datasets.
  • Demonstrated ability to leverage AI coding agents and developer assistants (e.g., GitHub Copilot, Cursor, or similar tools) to accelerate development velocity, improve code quality, and enable faster iteration cycles.

What the JD emphasized

  • Required experience with Presto and Trino for distributed SQL query processing across large-scale data sources.