Senior Software Engineer

ClickHouse ClickHouse · Data AI · Engineering

Senior Software Engineer role focused on building and maintaining JVM-based data framework integrations and connectors for ClickHouse, a high-performance OLAP database. The role involves owning the full lifecycle of integrations, from core drivers to SDKs, impacting how companies process massive datasets. Requires strong experience with big data processing frameworks like Spark, Flink, Kafka Connect, or Beam, and deep JVM knowledge.

What you'd actually do

  1. own the full lifecycle of data framework integrations - from the core database driver that [handles billions of records per second](https://www.linkedin.com/feed/update/urn:li:activity:7265414437384187904/), to SDKs and connectors that make ClickHouse feel native in JVM-based applications.
  2. craft tools that enable Data Engineers to harness ClickHouse's incredible speed and scale.
  3. collaborate closely with the open-source community, internal teams, and enterprise users to ensure our JVM integrations set the standard for performance, reliability, and developer experience.
  4. serve as a core contributor, owning and maintaining critical parts of ClickHouse's Data engineering ecosystem.

Skills

Required

  • 6+ years of software development experience focusing on building and delivering high-quality, data-intensive solutions.
  • Proven experience with the internals of at least one of the following technologies: Apache Spark, Apache Flink, Kafka Connect, or Apache Beam.
  • Experience developing or extending connectors, sinks, or sources for at least one big data processing framework such as Apache Spark, Flink, Beam, or Kafka Connect.
  • Strong understanding of database fundamentals: SQL, data modeling, query optimization, and familiarity with OLAP/analytical databases.
  • A track record of building scalable data integration systems (beyond simple ETL jobs)
  • Strong proficiency in Java and the JVM ecosystem, including deep knowledge of memory management, garbage collection tuning, and performance profiling.
  • Solid experience with concurrent programming in Java, including threads, executors, and reactive or asynchronous patterns.
  • Outstanding written and verbal communication skills to collaborate effectively within the team and across engineering functions.
  • Understanding of JDBC, network protocols (TCP/IP, HTTP), and techniques for optimizing data throughput over the wire.

Nice to have

  • Prior contributions to open-source projects
  • Familiarity with ClickHouse or similar high-performance data platforms.
  • Working knowledge of Python, especially in data engineering contexts (e.g., Pandas, PySpark, Airflow), and ability to contribute to Python tooling when needed.

What the JD emphasized

  • Proven experience with the internals of at least one of the following technologies: Apache Spark, Apache Flink, Kafka Connect, or Apache Beam.
  • A track record of building scalable data integration systems (beyond simple ETL jobs)