Vacancy caducado!
- Design and build high performing and scalable data pipeline platform using Hadoop, Apache Spark, MongoDB and object storage architecture.
- Design and build the data services on container-based architecture such as Kubernetes and Docker Partner with Enterprise data teams such as Data Management & Insights and Enterprise Data Environment (Data Lake) and identify the best place to source the data
- - Software engineering experience
- ETL (Extract, Transform, Load) Programming experience
- Agile experience
- Hadoop experience
- Java or Python experience
- Design and development experience with columnar databases using Parquet or ORC file formats on Hadoop
- Apache Spark design and development experience using Scala, Java, Python or Data Frames with Resilient - Distributed Datasets (RDDs)
Vacancy caducado!