skillindiajobs
Hyderabad Jobs
Banglore Jobs
Chennai Jobs
Delhi Jobs
Ahmedabad Jobs
Mumbai Jobs
Pune Jobs
Vijayawada Jobs
Gurgaon Jobs
Noida Jobs
Oil & Gas Jobs
Banking Jobs
Construction Jobs
Top Management Jobs
IT - Software Jobs
Medical Healthcare Jobs
Purchase / Logistics Jobs
Sales
Ajax Jobs
Designing Jobs
ASP .NET Jobs
Java Jobs
MySQL Jobs
Sap hr Jobs
Software Testing Jobs
Html Jobs
IT Jobs
Logistics Jobs
Customer Service Jobs
Airport Jobs
Banking Jobs
Driver Jobs
Part Time Jobs
Civil Engineering Jobs
Accountant Jobs
Safety Officer Jobs
Nursing Jobs
Civil Engineering Jobs
Hospitality Jobs
Part Time Jobs
Security Jobs
Finance Jobs
Marketing Jobs
Shipping Jobs
Real Estate Jobs
Telecom Jobs

Spark ML Engineer StreamFlux

2.00 to 4.00 Years   Mumbai City   26 May, 2021
Job LocationMumbai City
EducationNot Mentioned
SalaryNot Disclosed
IndustryKPO / Analytics
Functional AreaGeneral / Other Software
EmploymentTypeFull-time

Job Description

Fractal Analytics & Streamflux:Fractal is one of the most prominent players in the Artificial Intelligence space. Fractal s mission is to power every human decision in the enterprise and uses the power of AI to help the world s most admired Fortune 500 companies.StreamFlux is a self-serve data science platform that allows the building of a batch and streaming pipelines in a matter of minutes. It performs data processing such as aggregation, transformations, filtering and enrichment of the data, in the cloud or on-premise. Through its advanced analytics and data science capabilities it offers model training, scoring, monitoring, and validating, thus enabling power users of enterprises such as data engineers, data architects, business analysts, and data scientists to build complex data engineering and machine learning pipeline at ease.ResponsibilitiesOur Big Data capability team needs hands-on developers who can produce beautiful & functional code to solve complex analytics problems. If you are an exceptional developer with an aptitude to learn and implement using new technologies, and who loves to push the boundaries to solve complex business problems innovatively, then we would like to talk with you.

  • You would be responsible for evaluating, developing, maintaining and testing big data solutions for advanced analytics projects
  • The role would involve big data pre-processing & reporting workflows including collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into business insights
  • The role would also involve testing various machine learning models on Big Data, and deploying learned models for ongoing scoring and prediction. An appreciation of the mechanics of complex machine learning algorithms would be a strong advantage.
Experience & Tech Stack:
  • 2 to 4 years of demonstrable experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions.
Ideally, this would include work on the following technologies:
  • Expert-level proficiency in at-least one of Java, or Scala
  • Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop 2.0 (YARN; MR & HDFS) and associated technologies -- one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, etc
  • Hands-on experience with Spark Streaming and Spark ML is a must
  • Operating knowledge of cloud computing platforms (AWS, especially EMR, EC2, S3, SWF services and the AWS CLI)
  • Experience working within a Linux computing environment, and use of command line tools including knowledge of shell/Python scripting for automating common tasks
  • Ability to work in a team in an agile setting and clear understanding of how Git works
In addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners.Must Have (hands-on) experience:
  • Java or Scala
  • Linux environment and shell scripting
  • Distributed computing frameworks (Hadoop or Spark)
  • Data Structures
  • Apache Maven / Gradle
  • Kafka
Desirable (would be a plus):
  • Statistical or machine learning experience
  • Distributed and low latency (streaming) application architecture
  • Row store distributed DBMSs such as Cassandra
  • Familiarity with API design
  • Deep Learning experience
Education
  • B.E/B.Tech in Computer Science or related technical degree
,

Keyskills :
cloud computingproblem solvingdata architectsshell scriptingbig datalow latencydata processingdata solutionsdata sciencefortune 500computer science

Spark ML Engineer StreamFlux Related Jobs

© 2020 Skillindia All Rights Reserved