— February 2, 2017

Experience : 2 – 4 years
Requirements :
Task Info :

Position Summary:

We are looking for candidates with hands on experience in Big Data technologies to be based out of our Gurgaon office.

Key Responsibilities:

  • Build the Big Data infrastructure to store and process terabytes of data
  • Understand the business need what kind of data, how much data, types of algorithms to be run, load on the system, budget etc.- and recommend optimal solutions
  • Build and implement the solution. This will need you to be hands on to build in quick prototypes / proof of concepts data processing benchmarks
  • Work with the operations team to build systems, process and team required to run and maintain the systems securely, reliably and in a scalable manner
  • Work with the analytics team to understand what data landscaping would be required

Qualifications and Skills:

  • Must have 2-4 years of experience with Big Data technologies such as Hadoop and the related ecosystem
  • Practical experience and in-depth understanding of Map Reduce
  • Hands-on experience with Spark/Hive/Pig/Flume/Sqoop
  • Should have a good programming background with expertise in Java
  • Data infrastructure tools landscape e.g cloud service providers, virtualization software, system monitoring tools and development environments
  • Ability to program and guide junior resources on technical aspects
  • Ability to craft documents that can explain complex ideas in simple terms in order to build consensus or educate
  • Knowledge of R or any other Statistical Programming Language is a plus
  • Degree – Graduates/Postgraduates in CSE or related field

College Preference : no-bar
Min Qualification : ug
Skills : flume, hadoop, hive, mapreduce, pig, r, spark, sqoop
Location : Gurugram

About the Author

Our Top Authors

Download Analytics Vidhya App for the Latest blog/Article

Leave a Reply Your email address will not be published. Required fields are marked *