Job Description

Key Responsibilities

−Build and maintain data pipelines on Databricks using Spark and Delta Lake.

−Integrate semiconductor manufacturing data sources (MES, equipment logs, yield data) into the Lakehouse environment.

−Develop efficient ETL/ELT processes to ensure data quality, consistency, and scalability.

−Collaborate with manufacturing engineers and data scientists to deliver clean, reliable datasets for analytics and modeling.

−Document workflows and processes to ensure reproducibility and transparency.

Required Skills

−Hands-on experience with Databricks (Spark, Delta Lake, Notebooks).

−Strong proficiency in Python (PySpark) and SQL for data processing and transformation.

−Experience with large-scale data processing (batch and streaming).

−Familiarity with cloud platforms (AWS) and their core data services.

−Requires a minimum of 5 years of related experience with a Bachelor’s degree;
or 3 years and a Master...

Ready to Apply?

Take the next step in your AI career. Submit your application to NXP Semiconductors today.

Submit Application