Job Description
- Primary Skills: ETL
- SQL, ETL Concepts and testing techniques, Experience in cloud like AWS or Azure
- You will work with large-scale datasets and leverage the Apache Spark framework to implement data processing workflows.
- You will work closely with data engineers, analysts, and other teams to develop robust data pipelines and contribute to our big data ecosystem.
- This role requires experience in distributed computing and hands-on experience in Spark and PySpark for processing and analyzing large datasets.
- Strong experience with Apache Spark and PySpark for large-scale data processing.
- Proficiency in Python, including working with libraries like Pandas, NumPy, and PySparks DataFrame API.
- Experience with SQL for querying large datasets and performing data transformations.
- Familiarity with cloud platforms such as AWS (EMR),
- Experience in designing and implementing data pipelines and ETL processe...
Ready to Apply?
Take the next step in your AI career. Submit your application to Virtusa today.
Submit Application