Convergehealth - Data Operations Engineer, Expert Services-innovation_delivery_transformation

Data Operations Engineer for ConvergeHEALTH's Expert Services team, responsible for building and operating cloud-native data pipelines that integrate client healthcare data into Deloitte's Data Studio platform. This role involves data engineering, cloud operations, and applying AI tooling for analytics and anomaly detection, with a focus on creating reliable, decision-ready analytics from complex healthcare data.

What you'd actually do

  1. Design, build, and optimize cloud-native ETL/ELT pipelines that ingest client source data and conform it to the Data Studio platform's foundational data model — making real-world healthcare data ready to power production analytics.
  2. Profile, validate, and QA large, complex healthcare datasets for accuracy, completeness, and conformance to platform standards; combine traditional debugging with LLM-enabled data exploration and ML-based anomaly detection to find and resolve issues faster than manual approaches allow, partnering with client and Deloitte teams as needed when integration issues require it.
  3. Develop the analytics layer of the Data Studio platform — including BI dashboards, self-service reporting, and ML Lab workflows — putting validated, production-ready data in the hands of consulting teams and clients.
  4. Implement and maintain workflow automation, monitoring, and alerting using event-driven architectures and orchestration tools, with the goal of building systems that run reliably without constant intervention.
  5. Act as a hands-on technical voice into the Data Studio platform's evolution — translating real-world delivery learnings into concrete product, data model, and platform enhancement opportunities, and partnering with product and engineering teams to validate and pressure-test new capabilities before they ship.

Skills

Required

  • Expert SQL proficiency, including complex query authoring, data profiling, performance tuning, and query optimization across large-scale, messy datasets
  • Strong Python proficiency for data wrangling, scripting, automation, and integrating ML/AI capabilities into data pipelines
  • Hands-on experience designing and operating cloud-native data pipelines, with judgment around when to use which tool and how to debug distributed systems when things break; practical familiarity with AWS data services (e.g., Redshift, Glue, S3, Step Functions, Lambda) and exposure to AWS AI/ML services (e.g., Bedrock, SageMaker) a plus
  • Sound data modeling judgment, including conforming heterogeneous source data to standardized analytics models without losing fidelity
  • Demonstrated experience working with large, complex datasets across structured, semi-structured, and unstructured formats
  • Forward-thinking engineering mindset, including fluency with modern code collaboration workflows (Git, pull requests, code review), practical use of AI-assisted development tools (e.g., Claude Code, GitHub Copilot), and curiosity about emerging AI/ML techniques such as agentic patterns, RAG, and vector databases
  • Working familiarity with modern BI tools (e.g., Tableau, Power BI, Superset) and workflow orchestration platforms (e.g., Airflow, Step Functions)
  • Strong ownership mindset and comfort with ambiguity — able to self-manage priorities, juggle concurrent workstreams, and adapt as priorities shift
  • Clear communicator who works well across distributed engineering, product, and occasional client or consulting stakeh

Nice to have

  • exposure to AWS AI/ML services (e.g., Bedrock, SageMaker) a plus

What the JD emphasized

  • hands-on technical role
  • designing and operating the cloud-native data pipelines
  • applying emerging AI tooling
  • Expert SQL proficiency
  • Strong Python proficiency
  • Hands-on experience designing and operating cloud-native data pipelines
  • practical use of AI-assisted development tools
  • curiosity about emerging AI/ML techniques such as agentic patterns, RAG, and vector databases

Other signals

  • applied AI tooling
  • LLM-enabled data exploration
  • ML-based anomaly detection
  • AI-assisted development tools
  • agentic patterns
  • RAG
  • vector databases