Key Responsibilities:
- Design and implement end-to-end Databricks Lakehouse Architecture using Bronze / Silver / Gold medallion framework.
- Develop, manage, and optimize DBT Cloud projects including models, macros, tests, seeds, snapshots, and deployments.
- Build scalable ELT pipelines using Databricks, Delta Lake, Spark SQL, Python, and SQL.
- Configure and manage DBT Cloud jobs, scheduling, environments, and automated deployments.
- Implement enterprise data governance using Unity Catalog, role-based access control, lineage, and data quality controls.
- Design batch and streaming ingestion pipelines using Autoloader, Structured Streaming, ADF, Event Hubs, or similar tools.
- Optimize Databricks clusters, SQL Warehouses, workload performance, and cloud cost management.
- Integrate Databricks with BI tools such as Power BI, Tableau, or Fabric.
- Build CI/CD pipelines using GitHub Actions, Azure DevOps, Terraform, or similar DevOps tools.
- Collaborate with business stakeholders, data engineers, analysts, and leadership teams.
- Provide architecture guidance, best practices, code reviews, and mentor junior developers.
- Support migration from legacy ETL tools / data warehouses to modern Lakehouse platforms.