Staff Data Engineer

Visa Visa · Fintech · Bengaluru, India, IN

Staff Data Engineer on the Predictive Fraud Intelligence (PFI) team, focusing on MLOps and Data Engineering. The role involves building and maintaining data pipelines, developing platforms for ML experimentation and deployment, creating automated monitoring systems, and ensuring compliance for AI/ML models in a financial industry context. The position requires expertise in big data processing, distributed systems, cloud platforms, and MLOps tooling to support fraud-detection models.

What you'd actually do

  1. Building and maintaining reliable data pipelines that deliver high‑quality data across the product lifecycle – Product development to client support.
  2. Developing platforms that support rapid model experimentation, training, evaluation, versioning, and deployment.
  3. Creating automated monitoring systems for data drift, model performance, and operational health to ensure models stay accurate as fraud patterns evolve.
  4. Partnering closely with AI/ML researchers and product teams to reduce time from model concept to production.
  5. Ensuring compliance, security, and traceability across the full ML lifecycle to meet financial‑industry standards.

Skills

Required

  • 5+ years of relevant work experience with a Bachelor’s Degree or at least 2 years of work experience with an Advanced degree (e.g. Masters, MBA, JD, MD) or 0 years of work experience with a PhD, OR 8+ years of relevant work experience.
  • 8+ yrs. work experience with a bachelor’s degree or 6+ years of work experience with a Master's or Advanced Degree in an analytical field such as computer science, statistics, finance, economics, or relevant area.
  • handling big data on-premises as well as on cloud
  • Deep understanding of Hadoop ecosystem and associated technologies
  • good knowledge of cloud analytical solutions available
  • Strong expertise in designing and operating large‑scale data pipelines (batch and streaming) that process terabytes to petabytes of data.
  • Deep proficiency with distributed data‑processing frameworks such as Spark, Flink, Beam, or similar.
  • Solid command of data storage technologies (Delta Lake, Iceberg, Hive, BigQuery, Redshift, or equivalent).
  • Working experience with cloud‑based data‑processing systems (AWS EMR, Dataproc, Glue, Dataflow, Snowflake, BigQuery, Redshift, Databricks or equivalent).
  • Strong programming skills in Python, Scala, or Java, with a focus on building reliabe production systems.
  • Hands‑on experience with orchestration and workflow tools (Airflow, Dagster, equivalent).
  • Proficiency in containerization and orchestration (Docker, Kubernetes).
  • Experience implementing CI/CD pipelines for data and ML workloads.
  • Understanding of data‑quality frameworks, lineage, observability, and monitoring (Great Expectations, Deequ, Monte Carlo, Databand, or similar).
  • Practical knowledge of cloud platforms (AWS, GCP, or Azure) and cloud‑native data systems.

Nice to have

  • Leverage AI and automation tool

What the JD emphasized

  • petabyte-scale datasets
  • petabytes of data
  • compliance, security, and traceability across the full ML lifecycle to meet financial-industry standards
  • financial-industry standards

Other signals

  • MLOps and Data Engineering team designs and operates the platforms, pipelines, and tooling that enable the core product teams to build, deploy, and iterate on models quickly.
  • This group provides the scalable data foundations, model‑orchestration frameworks, and automated workflows required to keep fraud‑detection models continuously updated against emerging fraud schemes and new attack vectors.
  • This is a great opportunity to be part of a Data Engineering and MLOps team that is set out to scale and structure large scale data engineering and ML/AI that drives significant revenue for Visa.