Developer will be responsible for analyzing requirements, prototyping data analysis solutions (primarily in HiveQL or Spark and UNIX scripting), designing, developing and unit testing solutions, and facilitating solution deployment and support.
Candidates need to have strong capabilities in HiveQL, UNIX scripting.
Candidates should have experience with the Hadoop ecosystem and working with large data sets.
The system will consist of batch analytic processing on large sets of data.
Experience with Spark is preferred.
Candidate Skill Set
Strong HiveQL and SQL Development skills
Performance Tuning Map Reduce/Hive.
Programming Experience with Hadoop / Map-Reduce with Java
UNIX / Shell scripting
Designing distributed solutions for parallel processing of large data.
Full SDLC Experience (requirements analysis, design, development, unit testing, deployment, support).
Good communication skills
Experience with Spark/Scala programming
Experience with AWS Cloud Technologies
Experience with Agile Development Methodologies.
Experience with Big-Data Technologies in Hadoop Eco System
Test driven development.