Senior Software Engineer (data Platform)

Databricks Databricks · Data AI · Bangalore, India · Engineering - Pipeline

Databricks is seeking a Senior Software Engineer for their Data Platform team to build the Data Intelligence Platform. This role involves designing and operating systems for logging, orchestration, data transformation, metric storage, governance, and data consumption. The engineer will work with large-scale data pipelines, messaging systems, and workflow orchestration tools, contributing to the company's internal data infrastructure and guiding product direction.

What you'd actually do

  1. Design and run the Databricks metrics store that enables all business units and engineering teams to bring their detailed metrics into a common platform for sharing and aggregation, with high quality, introspection ability and query performance.
  2. Design and run the cross-company Data Intelligence Platform, which contains every business and product metric used to run Databricks. You’ll play a key role in developing the right balance of data protections and ease of shareability for the Data Intelligence Platform as we transition to a public company.
  3. Develop tooling and infrastructure to efficiently manage and run Databricks on Databricks at scale, across multiple clouds, geographies and deployment types. This includes CI/CD processes, test frameworks for pipelines and data quality, and infrastructure-as-code tooling.
  4. Design the base ETL framework used by all pipelines developed at the company.
  5. Partner with our engineering teams to provide leadership in developing the long-term vision and requirements for the Databricks product.

Skills

Required

  • Technical leadership on large projects
  • ETL frameworks
  • metrics stores
  • infrastructure management
  • data security
  • building, shipping and operating reliable multi-geo data pipelines at scale
  • working with and operating workflow or orchestration frameworks (e.g., Airflow, DBT)
  • large-scale messaging systems (e.g., Kafka, RabbitMQ)
  • cross-functional and communication skills
  • consensus builder

Nice to have

  • Passion for data infrastructure
  • enabling others by making their data easier to access

What the JD emphasized

  • 6+ years of industry experience
  • 4+ years of experience providing technical leadership on large projects similar to the ones described above - ETL frameworks, metrics stores, infrastructure management, data security.
  • Experience building, shipping and operating reliable multi-geo data pipelines at scale.
  • Experience working with and operating workflow or orchestration frameworks, including open source tools like Airflow and DBT or commercial enterprise tools.
  • Experience with large-scale messaging systems like Kafka or RabbitMQ or commercial systems.