Job Description

We are looking for DevOps Engineers with strong experience in data architecture, containerization, and cloud-native deployments. The ideal candidates will have hands-on expertise in Kafka-based data pipelines, Kubernetes orchestration, CI/CD automation, and data integration within regulated environments (preferably banking/financial services). This role involves building scalable, secure, and high-performance data platforms across multiple environments.
Key Responsibilities:
Implement and configure Helm packages across multiple environments.
Design and manage data ingestion pipelines using Apache Kafka, ensuring high availability and low-latency data delivery.
Translate functional and technical requirements into scalable technical solutions and workflows.
Manage data modelling, schema registry configurations, and serialization strategies.
Monitor, troubleshoot, and resolve issues in data pipelines, schema alignment, and data transformation processes.
Ensure data in...

Ready to Apply?

Take the next step in your AI career. Submit your application to Systems Limited today.

Submit Application