Software Engr II

Honeywell Honeywell · Industrial · India

Seeking a Data Engineer with 2-4 years of experience to design, build, and maintain scalable data pipelines using Databricks and cloud platforms. Responsibilities include managing ETL/ELT workflows, optimizing data models, collaborating with data scientists, and ensuring data quality. The role also involves team supervision and mentoring.

What you'd actually do

  1. Design, develop, and maintain data pipelines using Databricks (PySpark / Spark SQL)
  2. Implement and optimize ETL/ELT workflows using Databricks jobs, notebooks, and workflows
  3. Build and manage Delta Lake tables, ensuring data reliability, performance, and ACID compliance
  4. Develop and optimize data models for analytics, BI, and downstream consumption
  5. Work with batch and streaming data processing using Spark Structured Streaming (where applicable)

Skills

Required

  • Databricks platform
  • Python (PySpark)
  • Spark SQL
  • Delta Lake
  • SQL
  • cloud data storage (AWS S3 / Azure Data Lake / GCP Cloud Storage)
  • ETL orchestration tools (Databricks Workflows, Airflow, Azure Data Factory, etc.)
  • Git

Nice to have

  • streaming technologies (Kafka, Event Hubs, Kinesis)
  • dbt
  • Unity Catalog
  • Databricks governance features
  • cloud security
  • IAM
  • cost optimization
  • BI tools (Power BI, Tableau, Looker)
  • data science or ML workflows on Databricks
  • Agile/Scrum