Job Description

What You’ll Do

  • Design, develop, and implement scalable data processing pipelines using Azure Databricks and Spark

  • Write efficient and optimized Python (PySpark) and SQL code for large-scale data processing

  • Architect and maintain data lakehouse solutions, leveraging Microsoft Fabric for analytics and reporting

  • Implement data governance, cataloging, lineage, and compliance using Microsoft Purview

  • Build and optimize ETL workflows for ingestion, transformation, and integration across Azure services

  • Ensure data quality, security, privacy, and integrity, aligned with healthcare data standards

  • Collaborate with cross-functional teams and stakeholders operating in the US time zone

  • Maintain version control using Git and follow CI/CD best practices for data engineering

  • Produce clear and comprehensive technical documentation (requi...

Ready to Apply?

Take the next step in your AI career. Submit your application to Confidential today.

Submit Application