• Expertise in Hadoop ecosystem technologies such as HDFS, Hive, HBase, Spark, Kafka, Flink, and NiFi.
• Experience with Hadoop-based Ecosystems like Cloudera including its different compartments like CDP, CDF, and CML.
• Expertise in ELK Stack (Elastic Search, Kibana, LogStash) and administration of logging systems over them.
• Knowledge of distributed computing, cloud computing architectures, and data processing frameworks, such as Apache Spark.
• Experience with data ingestion and processing tools, such as Kafka and Flume.
• Strong communication and collaboration skills.
• Ability to train and educate junior team members on Big Data technologies and best practices.
• Ability to work independently and as part of a team.
• Expertise in Teradata, PowerBI Server, and Informatica is a plus.