Hadoop Developer with 7+ Years of IT experience including 5 years in Big Data and Analytics field in Storage, Querying, Processing and Analysis for developing E2E Data pipelines. Expertise in designing scalable Big Data solutions, data warehouse models on large-scale distributed data, performing wide range of analytics.
• Expertise in all components of Hadoop/Spark Ecosystems - Spark, Hive, Pig, Flume, Sqoop, HBase, Kafka, Oozie, Impala, Stream sets, Apache NIFI, Hue, AWS.
• 3+ years of experience working in programming languages Scala/Python.
• Extensive knowledge on data serialization techniques like Avro, Sequence Files, Parquet, JSON and ORC.
• Acute knowledge on Spark architecture and real-time streaming using Spark.
• Hands on experience with Spark Core, Spark SQL and Data Frames/Data Sets/RDD API.
• Good knowledge on Amazon Web Services (AWS) cloud services like EC2, S3, EMR and VPC.
• Experienced in Data Ingestion, Data Processing, Data Aggregations, Visualization in Spark Environment.
• Hands on experience in working with large volume of Structured and Un-Structured data.
• Expert in migrating the code components from SVN repository to Bit Bucket repository.
• Experienced in building Jenkins pipelines for continuous code integration from Github into Linux machine. Experience in Object Oriented Analysis Design (OOAD) and development.
• Good understanding in end-to- end web applications and design patterns.
• Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
• Experience in implementing by using agile methodology. Well versed in using Software development methodologies like Agile Methodology and Waterfall processes.
• Experienced in handling databases: Netezza, Oracle and Teradata.
• Strong team player with good communication, analytical, presentation and inter-personal skills.