Job Description
Responsibilities
Build a universal data platform, meet real-time/offline computing and storage requirements;Define the data pipeline scheme according to the demand scenarios, and deliver it by multiple methods based on various infrastructure;Enhance the data platform, improve the stability and flexibility of data asset, and optimize the resource efficiency.Requirements
Bachelor degree or above, major in computer, big data, mathematics, and more than 5 years of data development experience;Familiar with Hadoop, Spark, Flink, Airflow and other popular data platform components, understand the working principle of them, and have optimization experience about these components;Have a certain understanding of distributed system principles, calculations, and storage;Good communication and logical thinking skills, good self-drive, continuous learning and updating knowledge system;We may use art...
Ready to Apply?
Take the next step in your AI career. Submit your application to Confidential today.
Submit Application