View Our Website View All Jobs

Lead Engineer, Big Data

Position Expectations:

The Sr. Big data Engineer is a hands-on analytics professional with experience in managing big data engagements involving designing, developing and implementing big data and analytics solutions to solve industry-specific business problems. He/she leads multiple client projects and is responsible for examining and implementing the latest advancements in Big data analytics tools, practices and techniques to deliver greater results.

Principal Responsibilities:

Work across multiple client projects that require developing and implementing big data analytics solutions
Primary client contact for delivery and project management for a large big data project  manage day-to-day communication) or lead 1-2 Sr. Big Data Engineers to manage such projects
Work closely with clients to understand analysis needs and jointly evaluate and institutionalize new big data and analytical methods and techniques
Work with onsite Engagement Manager and offshore delivery team to ensure successful service delivery and client satisfaction
Review project deliverables and guide team members from big data consulting and analytics methodology perspective
Solve problems of large dimensionality in a computationally efficient and statistically effective manner
Evaluate, develop, test and maintain efficient codes in Big data environment
Build pre-processing and reporting workflows including collecting, parsing, managing, analyzing & visualizing large sets of data to turn information into business insights
Test various descriptive and predictive analytics, and text mining solutions / models on Big data platforms, and operationalized the developed  models for ongoing scoring and prediction
Collaborate with client teams to effectively integrate and communicate analysis findings
Advise Fractal analysts on use of appropriate big data and analytics tools, techniques and frameworks
Evaluate emerging technologies that may contribute to Fractal’s analytical platforms

Qualification & Experience

Desired Skills and Experience:

8-10 years of demonstrable experience in designing, developing and testing modular, efficient and scalable code implemented in Big data and analytics environment
Expert level proficiency in at least one of Python (preferred), Scala, Java (preferred) or C++
Minimum 2-3 years of working experience in distributed computing frameworks, particularly Apache Hadoop 2.0+ (YARN; MR & HDFS) and associated technologies -- one or more of Sqoop, Avro, Flume, Oozie, Zookeeper, etc.. 
Hands-on experience with Apache Hive, Apache Spark and its components (Streaming, SQL, MLLib)
Operating knowledge of cloud computing platforms (AWS, especially EMR, EC2, S3, SWF services and the AWS CLI)
Experience working within a Linux computing environment, and use of command line tools including knowledge of shell/Python scripting for automating common tasks
Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of GIT repository
Manage client deliverables, communicate effectively with internal and external teams
Excellent communication skills to be able to effectively translate and articulate the findings to technology, analytics and business stakeholders 

Education

B.E / B.Tech. / B.S. in Computer Science or advanced degree in related technical domain

Read More

Apply for this position

Required*
Apply with Indeed
Attach resume as .pdf, .doc, or .docx (limit 2MB) or Paste resume

Paste your resume here or Attach resume file