Job Description

  • Primary Skills: ETL
  • SQL, ETL Concepts and testing techniques, Experience in cloud like AWS or Azure
  • You will work with large-scale datasets and leverage the Apache Spark framework to implement data processing workflows.
  • You will work closely with data engineers, analysts, and other teams to develop robust data pipelines and contribute to our big data ecosystem.
  • This role requires experience in distributed computing and hands-on experience in Spark and PySpark for processing and analyzing large datasets.
  • Strong experience with Apache Spark and PySpark for large-scale data processing.
  • Proficiency in Python, including working with libraries like Pandas, NumPy, and PySparks DataFrame API.
  • Experience with SQL for querying large datasets and performing data transformations.
  • Familiarity with cloud platforms such as AWS (EMR),
  • Experience in designing and implementing data pipelines and ETL processe...

Ready to Apply?

Take the next step in your AI career. Submit your application to Virtusa today.

Submit Application