Vacancy caducado!
- Good knowledge of database structures, theories, principles, and practices.
- Experience in Java and Knowledge in Hadoop (HDFS/SPARK with or prior experience in MapReduce) concepts and ability to write Spark Dataset, Dataframe & SQL jobs. Proven understanding of Hadoop, Spark, Hive and ability to write Shell scripting.
- Familiarity with data loading tools like Sqoop, Flume, and Kafka. Knowledge of workflow/schedulers like Oozie.
- Good aptitude in multi-threading and concurrency concepts.
- Loading data from disparate data source sets.
- Added advantage with certifications like Cloudera Developer (CCA175)/ Hortonworks Developer (Spark & Hive) /Administrator Certifications added advantage.
- Hands-on experience with some NO SQL databases.
- Ability to analyze, identify issues with existing cluster, suggest architectural design changes.
- Ability/Knowledge to implement Data Governance in Hadoop clusters.
Vacancy caducado!