Azure Databricks Lead

Detroit

Published 16 hours ago

Job Overview

We are looking for a highly senior, deeply hands on Databricks Lead to lead a large‑scale Oracle‑to‑Databricks migration, covering schema migration, code conversion, and ODI job modernization. The ideal candidate has extensive experience building enterprise-grade data platforms on Databricks, has executed at least one greenfield Databricks implementation, and is exceptionally strong in PySpark, Spark SQL, framework development, and Databricks Workflows.

 

Key Responsibilities

  • Architect, design, and implement cloud-native data platforms using Databricks (ingestion → transformation → consumption).
  • Lead the full Oracle → Databricks migration including schema translation, ETL/ELT logic modernization, and ODI job replacement.
  • Develop reusable PySpark frameworks, data processing patterns, and orchestration using Databricks Workflows.
  • Build scalable, secure, and cost‑optimized Databricks infrastructure and data pipelines.
  • Collaborate with business and technical stakeholders to drive data modernization strategy.
  • Establish development best practices, coding standards, CI/CD, and DevOps/DataOps patterns.
  • Provide technical mentorship and create training plans for engineering teams.
  • Contribute to building MLOps and advanced operations frameworks.

 

Required Qualifications

  • 14+ years in Data Engineering/Architecture with at least 4+ years hands-on Databricks experience delivering end‑to‑end cloud data solutions.
  • Strong experience migrating from Oracle/on‑prem systems to Databricks, including SQL, PL/SQL, ETL logic, and ODI pipelines.
  • Deep hands-on expertise in: 
    • PySpark, Spark SQL, Delta Lake, Unity Catalog
    • Building reusable data frameworks
    • Designing high‑performance batch and streaming pipelines
  • Proven experience with greenfield Databricks implementations.
  • Strong understanding of cloud-native architectures on AWS and modern data platform concepts.
  • Solid knowledge of data warehousing, columnar databases, and performance optimization.
  • Good understanding of Agile/Scrum development processes.
  • Bonus: Experience designing Data Products, Data Mesh architectures, Data Vault or enterprise data governance models.
  • Good Understanding of Golde Gate.

Contract

Mid-Senior Level

Detroit