Job Description
Leading Global organisation firm in Sydney is about to start a Big Data implementation project
The role
In this team you will be part of the the overall data team, and will have involvement in multiple projects to deliver a data service to within the organisation.
Responsibilities
- Analyse content , integrity and coherency of new and ad hoc data sources to be used in data mining
- Collaborate with the data science team
- Create and maintain ETL transforms and flows for data sources being presented to the big data platform
- 2+ years working in data processing development and operations functions
- Experience in working with large amounts of diverse data using S3/HDFS/Hadoop and MapReduce in an AWS framework
- Strong programming skills in one or more of Python, Java, javascript, pig , hive
Desirable
- Experience in implementing google Big Query and or Redshift
- Knowledge of analytics tools such as SAS, SPSS with machine learning being an advantage