Job Description

We are seeking a highly skilled Data Engineer with expertise in ETL, PySpark, AWS, and big data technologies. The ideal candidate will have in-depth knowledge of Apache Spark, Python, and Java programming (Java 8 and above, including Lambda, Streams, Exception Handling, Collections, etc.). This role involves designing and developing scalable data processing pipelines for batch and real-time analytics.

Key Responsibilities

  • Develop data processing pipelines using PySpark.
  • Create Spark jobs for data transformation and aggregation.
  • Optimize query performance using file formats like ORC, Parquet, and AVRO.
  • Design scalable pipelines for both batch and real-time analytics.
  • Perform data enrichment and integrate with

Ready to Apply?

Take the next step in your AI career. Submit your application to Evnek today.

Submit Application