Job Description
What You’ll Do
Design, develop, and implement scalable data processing pipelines using Azure Databricks and Spark
Write efficient and optimized Python (PySpark) and SQL code for large-scale data processing
Architect and maintain data lakehouse solutions, leveraging Microsoft Fabric for analytics and reporting
Implement data governance, cataloging, lineage, and compliance using Microsoft Purview
Build and optimize ETL workflows for ingestion, transformation, and integration across Azure services
Ensure data quality, security, privacy, and integrity, aligned with healthcare data standards
Collaborate with cross-functional teams and stakeholders operating in the US time zone
Maintain version control using Git and follow CI/CD best practices for data engineering
Produce clear and comprehensive technical documentation (requi...
Ready to Apply?
Take the next step in your AI career. Submit your application to Confidential today.
Submit Application