Description
As part of Global Data Analytics project team visualizes large data sets from many sources, using a variety of tools. They work closely with business partners, data scientists, Data architects, solution architects and data analysts to better understand and evaluate data requirements and the type of data needed.
Responsibilities

    Candidate should have experience Hadoop MapReduce (MR),
    Big Data eco-system tools like Hadoop, Hive, SQOOP, MySQL
    Candiate should have 4+ years of experience in developing Hive scripts and optimizing the scripts.
    Candidate should have 3+ years of experience developing PySpark modules using RDDs, Dataframes API, Spark SQL and their optimization. Candidate should have 3+ years of experience in developing and scheduling OOZIE workflows with all the supported actions.
    Candidate should have 4+ years of experience in Hadoop commands and shell scripting to develop the adhoc jobs for data importing/exporting.
    Candidate should have 4+ years of experience in Java development.

Skills
Must have

    HiveQL, MapReduce, Sqoop,MySQL,Pysprak,Hive,Oozie Workflow,Java

Nice to have
Languages
English: B2 Upper Intermediate
Seniority
Regular


Source link