This posting has been closed |
- Hadoop development and implementation.
- Loading from disparate data sets.Be able to handle batch/real time ingestions
- Good knowledge about MQs/Kafka Messages
- Pre-processing using Hive and Pig.
- Designing, building, installing, configuring and supporting Hadoop data flow jobs
- Translate complex functional and technical requirements into detailed design.
- Perform analysis of vast data stores and uncover insights.
- Abide by security and data privacy standards for data access
- Create scalable and high-performance web services for data tracking.
- High-speed querying.
- Managing and deploying HBase/Spark oriented solutions using Java/Scala/Python
- Being a part of a POC effort to help build new Hadoop clusters for new solutions – hands on hive scripts/shell scripts
- Test prototypes and oversee handover to operational teams.
- Propose best practices/standards for development and scheduling batch operations
Requirements:
- Bachelor’s degree or equivalent work experience.
- At least 6+ years of experience working in Big Data Technologies
- At least 11 years of experience in Data warehousing
- At least 4+ years of experience in core Java and its ecosystem
- At least 3+ years of experience on Oracle big data stack (Cloudera, ODI, Fusion,etc).