Job Description
Design & Development
- Create and maintain ETL/ELT pipelines for large datasets in formats like CSV, JSON, Parquet, and Avro using Hadoop MapReduce and Spark frameworks. +2
Data Modeling
- Analyze existing schemas and recommend improvements to E-R and dimensional models. Develop logical and physical data models. +2
Performance Optimization
- Configure metadata, fine-tune pipeline performance, and optimize job execution. +10
Collaboration
- Work with cross-functional teams (DevOps, BI, stakeholders) to ensure smooth data workflows and scalable solutions. +10
Governance & Quality
- Apply best practices for data modeling, governance, and quality controls
Skills Required
Csv, Json, Devops, Bi, Etl
Ready to Apply?
Take the next step in your AI career. Submit your application to Han Digital Solution today.
Submit Application