Job Description

  • Proficiency in PySpark for distributed data processing and transformation.
  • Solid experience with AWS Glue for ETL jobs and managing data workflows.
  • Hands-on experience with AWS Data Pipeline (DPL) for workflow orchestration.
  • Strong experience with AWS services such as S3, Lambda, Redshift, RDS, and EC2.

Technical Skills:

  • Proficiency in Python and PySpark for data processing and transformation tasks.
  • Deep understanding of ETL concepts and best practices.
  • Familiarity with AWS Glue (ETL jobs, Data Catalog, and Crawlers).
  • Experience building and maintaining data pipelines with AWS Data Pipeline or similar orchestration tools.
  • Familiarity with AWS S3 for data storage and management, including file formats (CSV, Parquet, Avro).
  • Strong knowledge of SQL for querying and manipulating relational and semi-structured data.
  • Experience with Data Warehousing and Big Data technologie...

Ready to Apply?

Take the next step in your AI career. Submit your application to Virtusa today.

Submit Application