Mission :
Work closely with data analysts in order to build and validate the data model
Develop the processing flows using a generic Spark ETL tool or SQL statements
Document the technical steps taken
Push for new approaches which increase the efficiency of the team
Raise the knowledge level of the team by sharing key details about big data technologies and tools
Compétences demandées :
SQL, distributed processing (Hadoop, Spark)
Cloud technologies (especially GCP - BigQuery, Dataproc) a programming language (Scala or Python) git, CI/CD pipelines