ETL Engineer, Data Analyst, Big data Engineer ,Data Modeller

Job Description

Description

Permanent Positions



Calling all Data Analysts, Data Modellers, ETL Engineers and Big Data Engineers



Your new organisation: We are currently partnering with a leading consultancy to source for numerous data roles that will be working onsite at a "big 4" bank on some exciting projects



Your new role:



Data Analyst / Business Analyst


The data analyst is skilled team member with knowledge and skills to turn raw data into information and insight, which can be used to make business decisions.



What you'll need to succeed:


  • Interpret data, analyse results using statistical techniques and provide ongoing reports

  • Identify, analyse, and interpret trends or patterns in complex data sets

  • Working experience with data warehouse and data lake environments.

  • Strong Understanding of overall data domain and related concepts of Data Lineage, Metadata Management , Data Warehouse storage services such as Teradata, Oracle, etc

  • Data storage services such as HBase, Hive, HDFS etc

  • Expertise in data modelling principles/methods

  • Strong skills in Advanced SQL queries and Unix Shell Scripting

  • Expertise in data modelling principles/methods

  • Minimum 5 Years of working experience in Data Analytics, Data warehousing and Data Lake related IT development projects



ETL Engineer



ETL Engineers are responsible for developing, testing, implementing, and maintaining ETL solutions and data pipelines for data warehouse environment.

What you'll need to succeed:
  • Hands on experience working on ETL tools such as Informatica Power Centre, DataStage, data Warehouse storage services such as Teradata, Oracle, etc

  • Good understanding of basic scheduling mechanism such as Control M, data domain and related concepts of Data Modelling, Data Lineage, Metadata Management.

  • Strong skills in Advanced SQL queries and Unix Shell Scripting



Data Modeller



You will be responsible for designing, implementing, and documenting data architecture and data modelling solutions for data warehouse and data lake environments.

What you'll need to succeed:
  • Expertise in data modelling principles/methods including conceptual, logical & physical Data Models, experience in translating/mapping relational data models into XML and Schemas.

  • Strong Understanding of overall data domain and related concepts of Data Lineage, Metadata Management,

  • Strong skills in Advanced SQL queries and Unix Shell Scripting

  • Data Warehouse storage services such as Teradata, Oracle, etc,data storage services such as HBase, Hive, HDFS etc

  • Data analysis and modelling tools (ERWin, ER/Studio)

  • Minimum 8 Years of working experience in Data Analytics, Data warehousing and Data Lake related IT development projects

Big Data Engineer



Big Data Engineer are responsible for developing, testing, implementing, and maintaining big data solutions and data pipelines for data lake environment with the ,ability to design solutions independently based on high-level architecture

What you'll need to succeed:
  • Big Data/Data Lake technologies such as Hortonworks, Cloudera etc

  • Hands on experience working on Spark, Python, Cloud Technologies, preferably Microsoft Azure

  • Data storage services such as HBase, Hive, HDFS etc, Data management services such as Atlas, Cloudera Navigator etc

  • Exposure to both batch and real time processing

  • Good understanding of basic scheduling mechanism such as Control M

  • Understanding of overall data domain and related concepts of Data Modelling, Data Lineage, Metadata Management



What you need to do now:

Please apply or to find out more about this exciting opportunity, please ,contact Menka on 0292492265 or email to menka.tahiliani@hays.com.au [mailto:menka.tahiliani@hays.com.au] for a detailed and confidential discussion . I look forward to hearing from you.



SA Licence number :
LHS 297508

 
 
Apply

There is something wrong with this job ad? Report the error



Related Ads