Job Description
Key Responsibilities:
● Develop Apache Airflow DAGs and PySpark ETL pipelines for high volume data processing.
● Write optimized SQL queries for data transformation and aggregation.
● Build data products serving Business Process, Executive KPIs, and Product Analytics.
● Implement data quality and monitoring solutions.
● Optimize pipeline performance and troubleshoot production issues.
● Collaborate with cross-functional teams.
● Production Pipeline Monitoring (KLO).
Required skills:
● 10+ years of data engineering experience. 7+ years dedicated to the Big data stack.
● Expert in Python and PySpark (DataFrame API, Spark SQL).
● Advanced SQL skills (window functions, complex queries).
● Production experience with...
Ready to Apply?
Take the next step in your AI career. Submit your application to Nucleusteq today.
Submit Application