Job Description:
• Experience in design, development, and maintenance and support of Big Data Analytics using Hadoop Ecosystem components like HDFS, Hive, Pig, HBase, Sqoop, Zookeeper, Map Reduce, and oozie.
• Strong working experience with ingestion, storage, querying, processing and analysis of big data.
• Extensive Experience on working with Hadoop Architecture and the components of Hadoop - Map Reduce, HDFS, Job Tracker, Task Tracker, Name Node and Data Node.
• Having exposure to faster processing of large datasets using Apache Spark and distributed messaging and multi-processing of Kafka.
• Hands on experience in writing and tweaking Map Reduce programs according to the requirement.
• Good experience with "Productionalizing" Apache Hadoop, Apache Spark and Apache Kafka applications (such as monitoring, debugging and performance tuning).
• Experience in writing Pig scripts, hive scripts, pig UDF, hive UDF for ETL process workflows.
• Familiar with importing and exporting data using Sqoop into HDFS and have experience in fine tuning the sqoop actions for terabytes of data.
• Experience with SQL, PL/SQL and database concepts and NoSQL databases.
• Good Experience with job workflow scheduling like Oozie, D-series and Autosys
• Good understanding of Experience on creating databases, tables and views in HIVE, IMPALA