Job Description

We are looking for an experienced Java + Apache Spark Developer with 6+ years of hands-on experience in building scalable, high-performance data processing applications. The ideal candidate should have strong expertise in Java, distributed data processing, and big data ecosystems.


Key Responsibilities

  • Design, develop, and maintain scalable data processing applications using Java and Apache Spark
  • Develop batch and real-time data pipelines
  • Optimize Spark jobs for performance, scalability, and reliability
  • Work with large datasets using distributed computing frameworks
  • Integrate Spark applications with Hadoop ecosystem tools (HDFS, Hive, etc.)
  • Collaborate with data engineers, analysts, and cross-functional teams
  • Troubleshoot production issues and ensure high availability
  • Follow best practices in coding, testing, and deployment


Required Skills & Qualifica...

Ready to Apply?

Take the next step in your AI career. Submit your application to Deloitte today.

Submit Application