Hadoop/ Spark Developer - Spark SQL Streaming

HUQUO CONSULTING PVT. LTD.
  • Bangalore
  • Confidential
  • 5-8 years
  • Views
  • 27 Nov 2018

  • Software Design & Development

  • Consumer Durables/ Semi Durables
Job Description

Hadoop/Spark Developer - Spark SQL Streaming (5-8 yrs)
 
- Experience with enterprise data management, Business Intelligence, data integration, and SQL database implementations
- Experience with the major big data solutions like - Experience with enterprise data management, Business Intelligence, data integration, and SQL database implementations
- Experience with the major big data solutions like Hadoop, MapReduce, Hive, Spark, Scala, HBase, MongoDB, Cassandra.
- Programming/scripting languages like Java, Linux, PHP, Ruby, Python and/or R. As well as have experience in working with ETL tools such as Informatica, Talend, Pentaho etc.
- He or she should have experience in designing solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high-scale or distributed RDBMS and/or knowledge on NoSQL platforms.
- Experience in data migration from relational databases to Hadoop HDFS
- Propose best practices/standards
- Translate, load and present disparate datasets in multiple formats/sources including JSON, XML etc.

Must Have Skills :
- Hadoop stack including HDFS cluster, MapReduce, Hive, Spark and Impala
- Web Technologies CSS, DHTML, XML, Hight Charts, Linux
- ETL tools such as Informatica, Talend and/or Pentaho.
Query : SQL, No SQL Concepts
Ingest : Kafka, Sqoop, Flume
Orchestration : Zookeeper
Databases : Postgres, Mongo DB, Cassandra, HBase
Languages : Java, Scala
Scripting : JavaScript, DHTML, XML, Shell
Good to have Skills :
Core : AWS, Hadoop, Yarn
Process : Agile-Scrum, Iterative Development, DevOps, CI
Analytics : Descriptive, Predictive (Added advantage)
Tools : Jenkins and TFS
Languages : Python, Java Enterprise
- MapReduce, Hive, Spark, Scala, HBase, MongoDB, Cassandra.
- Programming/scripting languages like Java, Linux, PHP, Ruby, Python and/or R. As well as have experience in working with ETL tools such as Informatica, Talend, Pentaho etc.
- He or she should have experience in designing solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high-scale or distributed RDBMS and/or knowledge on NoSQL platforms.
- Experience in data migration from relational databases to Hadoop HDFS
- Propose best practices/standards
- Translate, load and present disparate datasets in multiple formats/sources including JSON, XML etc.

Must Have Skills :
- Hadoop stack including HDFS cluster, MapReduce, Hive, Spark and Impala
- Web Technologies CSS, DHTML, XML, Hight Charts, Linux
- ETL tools such as Informatica, Talend and/or Pentaho.
Query : SQL, No SQL Concepts
Ingest : Kafka, Sqoop, Flume
Orchestration : Zookeeper
Databases : Postgres, Mongo DB, Cassandra, HBase
Languages : Java, Scala
Scripting : JavaScript, DHTML, XML, Shell
Good to have Skills :
Core : AWS, Hadoop, Yarn
Process : Agile-Scrum, Iterative Development, DevOps, CI
Analytics : Descriptive, Predictive (Added advantage)
Tools : Jenkins and TFS
Languages : Python, Java Enterprise  


Competencies/Skill sets for this job

Javascript Hive Flume NoSQL Sqoop Java Mapreduce Hdfs Impala CSS

Job Posted By

About Organisation

HUQUO CONSULTING PVT. LTD.