Vacancy One:
Ideally we’re looking for someone who has had exposure to the following:
- Data warehouse maintenance (SQL)
- Data Pipeline design and implementation (Python)
- Big data relevant projects (Apache Hadoop, Hive, Spark and Airflow)
- Cloud experience, preferably with AWS
Vacancy Two:
The projects for this role will include modelling so we’re looking for someone who has ideally had exposure to the following:
- Data pipeline design and implementation (Python)
- Prior experience around collaborating with data scientists (feature engineering and model productisation)
- Prior experience with main stream frameworks (such as Apache Spark, Airflow (or equivalent scheduling tool), AWS EMR, AWS S3)