Job Description
1. 2&plus years in Hadoop Platform
2. 3 to 5 years of Big Data experience including use of Flume, Kafka, PySpark, Sqoop
3. Proficient understanding of Underlying infrastructure for Big Data Solutions (Clustered/Distributed Computing, Storage, Data Center Networking)
4. Experience in development of Big Data Applications
5. Experience Hadoop cluster, with all included services Ability to solve any ongoing issues with operating the cluster
6. Proficiency with Hadoop, MapReduce using Python, HDFS
7. Experience with building stream-processing systems, using solutions such as Storm or Spark-Streaming Good knowledge of Big Data querying tools, such as Pig, Hive, and Impala
8. Experience with Spark
9. Experience with integration of data from multiple data sources
10. Experience with NoSQL databases, such as HBase, Cassandra, MongoDB
11. Knowledge of various ETL techniques and frameworks, such as Flume
12. Experience with various messaging systems, such as Kafka.
13. Experience with Big Data Client toolkits, such as Mahout, SparkML, or H2O.