Job Description

Specific Duties
-Reviewing, designing, developing ETL jobs to ingest data into Data Lake, load data to data marts;

-extract data to integrate with various business applications.

-Parse unstructured data, semi structured data such XML etc.

-Design and develop efficient Mapping and workflows to load data to Data Marts

-Map XML DTD schema in Python (customized table definitions)

-Write efficient queries and reports in Hive or Impala to extract data on ad hoc basis for data analysis.

-Identify the performance bottlenecks in ETL Jobs and tune their performance by enhancing or redesigning them.

-Responsible for performance tuning of ETL mappings and queries.

-import tables and all necessary lookup tables to facilitate the ETL process required to process daily XML files in addition to processing the very large (multi-terabytes) historical XML data files

Ready to Apply?

Take the next step in your AI career. Submit your application to ipvisibility today.

Submit Application