Job Description

Noesis is looking for candidates with the following profile:

Main Tasks and Responsibilities:


  • Develop and maintain data processing workflows using SQL, Python, and PySpark;

  • Design, implement, and optimise data models within Data Warehouse and Data Lake architectures;

  • Work within distributed data ecosystems, supporting platforms such as Hadoop, Hive, Spark, or Databricks:

  • Contribute to the automation of data pipelines and engineering processes:

  • Participate in code reviews and ensure alignment with established development standards.

Requirements:


  • Degree in Computer Science or similar;

  • Minimum of 1 year as a Data Engineer;

  • Foundational knowledge of SQL and Python or similar data processing technologies;

  • Basic understanding of data modelling concepts and data ...

Ready to Apply?

Take the next step in your AI career. Submit your application to Noesis today.

Submit Application