Senior Software Engineer - Contributory Network

ZoomInfo ZoomInfo · Enterprise · Waltham, MA · 936 Engineering - Data Engineering

Senior Software Engineer to join the Contributory Network team, focusing on building and operating large-scale data pipelines for first-party data ingestion, transformation, and enrichment. The role involves architecting resilient ETL/ELT pipelines, building streaming and batch processing systems, establishing data quality frameworks, and defining observability for pipeline health. This is a critical, greenfield initiative powering intelligence products.

What you'd actually do

  1. Own and drive the design and implementation of large-scale data pipelines that ingest, validate, transform, and enrich first-party contributed data from CRM systems, email providers, and recording platforms
  2. Architect resilient ETL/ELT pipelines handling massive volumes of contact data, opportunity metadata, engagement signals, and activity patterns
  3. Take initiative on complex technical challenges - identify problems proactively, propose solutions, and execute with urgency
  4. Build streaming and batch processing systems for real-time and scheduled data flows using Kafka, Pub/Sub, Apache Beam, or similar
  5. Establish data quality frameworks, ensuring accuracy, consistency, and completeness across contributed data

Skills

Required

  • 5+ years of professional software engineering experience with a strong focus on data engineering
  • Proven track record of building and operating production data pipelines at scale
  • Deep experience with Python and/or Java
  • Hands-on expertise with data processing technologies: Apache Beam, Apache Airflow, Spark, Google Dataflow, or DataProc
  • Strong experience with streaming systems (Apache Kafka, Google Pub/Sub, or similar)
  • Experience with cloud platforms, preferably GCP (BigQuery, GKE, Dataflow)
  • Solid understanding of data modeling, schema evolution, and data quality management
  • Experience designing and operating large-scale ETL/ELT pipelines processing terabytes of data
  • Demonstrated ability to drive complex technical initiatives end to end - from scoping through delivery
  • Track record of operating with high autonomy and a bias toward action
  • Ability to push through ambiguity, make pragmatic decisions under uncertainty, and unblock progress
  • Experience influencing technical direction within a team or across teams
  • Strong code review and technical mentorship skills
  • Proven ability to balance quality with velocity - you ship, iterate, and improve
  • Bachelor's degree in Computer Science, Software Engineering, or a related field
  • Exceptional interpersonal skills with a proven ability to build productive cross-departmental relationships
  • Strong communicator who can explain complex systems to diverse audiences
  • Entrepreneurial mindset - comfortable pioneering new capabilities and wearing multiple hats

Nice to have

  • Experience with Kubernetes (GKE/EKS) for running distributed workloads
  • Familiarity with multi-cloud environments (GCP + AWS)
  • Experience with Snowflake, BigQuery, Starburst/Trino, or similar query engines
  • Experience with Terraform or infrastructure-as-code
  • Knowledge of data integration patterns with CRM systems, email providers, or recording platforms
  • Exposure to AI/LLM-based data processing approaches
  • Experience in a B2B data company or data-as-a-product environment
  • Experience with Apache Spark at expert level

What the JD emphasized

  • driver
  • takes ownership
  • pushes through ambiguity
  • unblocks themselves and others
  • relentlessly moves work forward
  • carve the path
  • make decisions
  • deliver
  • thrive when given a hard problem and the autonomy to solve it
  • drive complex technical initiatives end to end
  • operating with high autonomy
  • bias toward action
  • push through ambiguity
  • make pragmatic decisions under uncertainty
  • unblock progress
  • goes through walls
  • sees a blocker and finds a way
  • default mode is action
  • analysis paralysis