Specialist Data Analysis

AT&T AT&T · Telecom · Naucalpan De Juarez, Mexico

This role focuses on modernizing data pipelines and ETL processes for finance commissions at AT&T, using Databricks, PySpark, and integrating various data sources. It emphasizes data accuracy and reliability for sales compensation calculations.

What you'd actually do

  1. Migrate and refactor legacy ETL processes from Java, Python, and SQL into Databricks using PySpark
  2. Design and maintain scalable data pipelines and workflows using Databricks and Apache Iceberg
  3. Integrate data from MongoDB, SQL Server, and Oracle through Trino connectors
  4. Build and maintain data models optimized for analytics, reporting, and ML applications
  5. Implement data quality checks, validation, and automated anomaly detection to protect compensation accuracy

Skills

Required

  • PySpark
  • Databricks
  • SQL
  • data pipeline orchestration
  • MongoDB
  • SQL Server
  • Oracle
  • Trino
  • Apache Iceberg
  • Delta Lake
  • cloud data platforms (Azure, AWS, or GCP)
  • migrating legacy ETL workloads
  • data modeling
  • partitioning
  • performance optimization
  • CI/CD for data workflows
  • Git
  • Advanced conversational English

Nice to have

  • ML applications

What the JD emphasized

  • accuracy, reliability, and performance of what you build matters