Uploaded File
add photo
Mirza
azharbaighddev@gmail.com
669-256-2364
9 years experience
0
Recommendations
Average rating
64
Profile views
Summary

Experience
Information Technology
Jun 2019 - present
* Hands on experience on importing and exporting data from different databases like Oracle, MySQL into HDFS and Hive using Sqoop.
• Worked extensively with HIVE DDLs and Hive Query language (HQLs)
• Having experience in writing Hive queries to parse and process HDFS data.
• Analyze the large data sets by using Pig Scripts
• Having good experience in integration between hive and pig and H-base using H- catalog.
• Good Work Experience on Apache Spark with Scala programming.
• I have good work experience to creating data frames and Rdd's using Scala.
• I have good work experience to writing regular expressions.
• I have good work experience in shark-spark - sql
• Hands on experience in creating RDD & applying operations transformations and Actions.
• Having good experience in creating maven projects in Scala IDE
• Good Knowledge on Flume, Kafka.
• Having good knowledge in creating hive tables in H-base
• Having good knowledge in integration of hive metastore to H-base
• Having good experience in ran spark-submit jobs in standalone mode.
• Great team player and quick learner with effective communication, motivation, and organizational skills combined with attention to details and business improvements.
No skills were added
Remove Skill
Hadoop Developer
Information Technology
Apr 2018 - present
Houston, TX
Responsibilities:
• Support, monitoring of production jobs and provides 365x24x7 technical support and incident Production resolution
• Expertise in Creating, Debugging, Scheduling and Monitoring jobs using Airflow and Oozie.
• Created Airflow Scheduling scripts in Python.
• Moved all files data from various sources to HDFS for further processing.
• Moved all log/text files generated by various products into HDFS location.
• Written the Sqoop commands for importing and exporting data from Relational Databases to HDFS and vice versa.
• Performed Puppet, Kibana, Elastic Search, and Tableau, Red Hat infrastructure for data ingestion, processing, and storage.
• Good Knowledge on Hadoop Ecosystem, HDFS, Hadoop, Spark Architectures.
• Expertise in working with Spark Framework using Spark SQL, Spark Streaming.
• Prepared, processed numerous customer input files parsed and reformatted the data to meet product requirements
• Experience in manipulating/analysing large datasets and finding patterns and insights within structured data.
• Kibana implementation for logging of communication requests.
• AWS IOT, data processing with spark, repository using s3, redshift, Athena, REPORTING using ELK (Elasticsearch, Logstash, and Kibana on EC2), Node-red Embedded Side.
• Worked on Reporting through ELK (Elastic Search, LogStash, Kibana) .
• Good Perception on Production/Application Support life cycle and Strong Analytical and Programming Skills. Created custom sized VPC, Subnets, Security group, Internet gateways, VPC Peering and route tables to ensure a secure zone for organization in cloud.
• Proficient in AWS services like EC2, VPC, S3, ELB, Auto Scaling, EBS, RDS, IAM, Cloud Formation, Route 53, Cloud Watch, and Cloud Front.
• Maintained edge location to cache data with CDN using Cloud Front to deliver data with less latency.
• Configured and maintained the monitoring and alerting of production servers using Cloud Watch.
• Stay updated with the latest features/capabilities of the AWS platform.
• Experience in writing PIG scripts to access HDFS data in Hadoop Systems
• Experience in writing of HIVE reports & Oozie scheduling
• Highly experienced in importing and exporting data between HDFS and Relational Database Management systems using Sqoop.
• Knowledge on Analyzing the Data using K-Means Algorithm with the help of MiLB.
• Proficient in Technologies like SQL, PL/SQL, HiveQl, HBase, Spark SQL.
• Experience in Implementing OOZIE workflows
• Hands on experience in VPN Putty WinSCP VNCviewer etc.
• Good Knowledge in Dealing with UNIX commands
• Having knowledge and experience on complete installation of jdk1.6.0, HDFS, PIG, Hive and eclipse.
• Good Knowledge on Python
• Having good knowledge & Experience in Other Utilities TOAD, SQL LOADER, SQL*PLUS.
• Strong experience in requirements gathering, analysis, conversion and implementation of business requirements into business requirement documents, high level and low-level design documents
• Knowledge on Machine Learning using Spark Mlib.
• Proficient in handling oracle Procedures, Functions and Packages.
• Good Experience in Handling End-to-End Projects.
• Written Hive queries to process the HDFS data.
• Loaded all the data in HIVE-External Tables in later Sprints.
• Involved in conversion of SQL to HQL and My SQL Procedures to Scala Code to run on Spark Engine.
• Creating Rdd to load the unstructured data.
• Used apache airflow for AUTOMATIC SLIDING DOOR SIDEWALL, NOSE CONE UNDERFRAME LUGGAGE RACK
• Regular expressions are used which can be used n using the. method.
• Using map & flat map transimitions to map data which was stored in Rdd.
• Creating data frames to get tables format
• Using show and count actions after creation of data frames
• Used Oozie to schedule jobs with workflow and coordinator xml files.
• Daily interaction with client to get inputs to know about new enhancements. Environment: Cloudera Hadoop (CDH), Hive, Sqoop, Oozie, HDFS, Spark.
Apache AWS Eclipse Hadoop Hadoop Developer Hbase HDFS Hive Machine Learning Oozie Oracle Pig Python Spark SQL SQL Loader Sqoop Tableau TOAD UNIX XML
Remove Skill
Information Technology
Jun 2019 - present
* Hands on experience on importing and exporting data from different databases like Oracle, MySQL into HDFS and Hive using Sqoop.
• Worked extensively with HIVE DDLs and Hive Query language (HQLs)
• Having experience in writing Hive queries to parse and process HDFS data.
• Analyze the large data sets by using Pig Scripts
• Having good experience in integration between hive and pig and H-base using H- catalog.
• Good Work Experience on Apache Spark with Scala programming.
• I have good work experience to creating data frames and Rdd's using Scala.
• I have good work experience to writing regular expressions.
• I have good work experience in shark-spark - sql
• Hands on experience in creating RDD & applying operations transformations and Actions.
• Having good experience in creating maven projects in Scala IDE
• Good Knowledge on Flume, Kafka.
• Having good knowledge in creating hive tables in H-base
• Having good knowledge in integration of hive metastore to H-base
• Having good experience in ran spark-submit jobs in standalone mode.
• Great team player and quick learner with effective communication, motivation, and organizational skills combined with attention to details and business improvements.
No skills were added
Remove Skill
Hadoop Developer
Information Technology
Apr 2018 - present
Houston, TX
Responsibilities:
• Support, monitoring of production jobs and provides 365x24x7 technical support and incident Production resolution
• Expertise in Creating, Debugging, Scheduling and Monitoring jobs using Airflow and Oozie.
• Created Airflow Scheduling scripts in Python.
• Moved all files data from various sources to HDFS for further processing.
• Moved all log/text files generated by various products into HDFS location.
• Written the Sqoop commands for importing and exporting data from Relational Databases to HDFS and vice versa.
• Performed Puppet, Kibana, Elastic Search, and Tableau, Red Hat infrastructure for data ingestion, processing, and storage.
• Good Knowledge on Hadoop Ecosystem, HDFS, Hadoop, Spark Architectures.
• Expertise in working with Spark Framework using Spark SQL, Spark Streaming.
• Prepared, processed numerous customer input files parsed and reformatted the data to meet product requirements
• Experience in manipulating/analysing large datasets and finding patterns and insights within structured data.
• Kibana implementation for logging of communication requests.
• AWS IOT, data processing with spark, repository using s3, redshift, Athena, REPORTING using ELK (Elasticsearch, Logstash, and Kibana on EC2), Node-red Embedded Side.
• Worked on Reporting through ELK (Elastic Search, LogStash, Kibana) .
• Good Perception on Production/Application Support life cycle and Strong Analytical and Programming Skills. Created custom sized VPC, Subnets, Security group, Internet gateways, VPC Peering and route tables to ensure a secure zone for organization in cloud.
• Proficient in AWS services like EC2, VPC, S3, ELB, Auto Scaling, EBS, RDS, IAM, Cloud Formation, Route 53, Cloud Watch, and Cloud Front.
• Maintained edge location to cache data with CDN using Cloud Front to deliver data with less latency.
• Configured and maintained the monitoring and alerting of production servers using Cloud Watch.
• Stay updated with the latest features/capabilities of the AWS platform.
• Experience in writing PIG scripts to access HDFS data in Hadoop Systems
• Experience in writing of HIVE reports & Oozie scheduling
• Highly experienced in importing and exporting data between HDFS and Relational Database Management systems using Sqoop.
• Knowledge on Analyzing the Data using K-Means Algorithm with the help of MiLB.
• Proficient in Technologies like SQL, PL/SQL, HiveQl, HBase, Spark SQL.
• Experience in Implementing OOZIE workflows
• Hands on experience in VPN Putty WinSCP VNCviewer etc.
• Good Knowledge in Dealing with UNIX commands
• Having knowledge and experience on complete installation of jdk1.6.0, HDFS, PIG, Hive and eclipse.
• Good Knowledge on Python
• Having good knowledge & Experience in Other Utilities TOAD, SQL LOADER, SQL*PLUS.
• Strong experience in requirements gathering, analysis, conversion and implementation of business requirements into business requirement documents, high level and low-level design documents
• Knowledge on Machine Learning using Spark Mlib.
• Proficient in handling oracle Procedures, Functions and Packages.
• Good Experience in Handling End-to-End Projects.
• Written Hive queries to process the HDFS data.
• Loaded all the data in HIVE-External Tables in later Sprints.
• Involved in conversion of SQL to HQL and My SQL Procedures to Scala Code to run on Spark Engine.
• Creating Rdd to load the unstructured data.
• Used apache airflow for AUTOMATIC SLIDING DOOR SIDEWALL, NOSE CONE UNDERFRAME LUGGAGE RACK
• Regular expressions are used which can be used n using the. method.
• Using map & flat map transimitions to map data which was stored in Rdd.
• Creating data frames to get tables format
• Using show and count actions after creation of data frames
• Used Oozie to schedule jobs with workflow and coordinator xml files.
• Daily interaction with client to get inputs to know about new enhancements. Environment: Cloudera Hadoop (CDH), Hive, Sqoop, Oozie, HDFS, Spark.
Apache AWS Eclipse Hadoop Hadoop Developer Hbase HDFS Hive Machine Learning Oozie Oracle Pig Python Spark SQL SQL Loader Sqoop Tableau TOAD UNIX XML
Remove Skill
Hadoop Developer
Information Technology
Oct 2017 - Mar 2018
Chicago, IL
Responsibilities:
• Responsible in Installation and Configuration of Hadoop Eco system components using CDH 5.2 Distribution.
• Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
• MySQL Database administration, Microsoft .Net, Java Script, RPA Tool, kibana reporting (Edge verve Technology)
• Expertise in architecture blueprints. Created AWS infrastructure for first move group application and its deployment.
• Create and Manage User Access Control using AWS IAM policy for distinct group of users.
• Assisted in setting up CI/CD pipeline in AWS.
• Amazon component alerting and monitoring through Cloud Watch.
• AWS cloud trail setup for auditing of various AWS accounts.
• Worked as AWS admin for creating backups/snapshots and maintaining the application.
• Hands on with EC2, ECS, ELB, EBS, S3, VPC, IAM, SQS, RDS, Lambda, Cloud Watch etc.
• Processed Multiple Data sources input to same Reducer using Generic Writable and Multi Input format.
• Worked Big data processing of clinical and non clinical data using Map Reduce.
• Visualize the HDFS data to customer using BI tool with the help of Hive ODBC Driver.
• Customized BI tool for manager team that perform Query analytics using HiveQL.
• Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
• Created Partitions, Buckets based on State to further process using Bucket based Hive joins.
• Created Hive Generic UDF's to process business logic that varies based on policy.
• Moved Relational Data base data using Sqoop into Hive Dynamic partition tables using staging tables.
• Experienced in Monitoring Cluster using Cloudera manager.
• Involved in Discussions with business users to gather the required knowledge.
• Capable of creating real time data streaming solutions and batch style large scale distributed computing applications using Apache Spark, Spark Streaming, Kafka and Flume.
• Analyzing the requirements to develop the framework.
• Designed and developed architecture for data services ecosystem spanning Relational, NoSQL and Big Data technologies.
• Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
• Developed Java Spark streaming scripts to load raw files and corresponding.
• Processed metadata files into AWS S3 and Elasticsearch cluster.
• Developed Python Scripts to get the recent S3 keys from Elasticsearch.
• Elaborated Python Scripts to fetch/get S3 files using Boto3 module.
• Implemented PySpark logic to transform and process various formats of data like XLSX, XLS, JSON, TXT.
• Built scripts to load PySpark processed files into Redshift Db and used diverse PySpark logics.
• Developed scripts to monitor and capture state of each file which is being through.
• Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources.
• Involved in scheduling Oozie workflow engine to run multiple Hives and pig jobs and used Oozie Operational Services for batch processing and scheduling workflows dynamically.
• Included migration of existing applications and development of new applications using AWS cloud services.
• Wrought with data investigation, discovery and mapping tools to scan every single data record from many sources.
• Implemented Shell script to automate the whole process.
• Integrated Apache Storm with Kafka to perform web analytics.
• Uploaded click stream data from Kafka to Hdfs, HBase, and Hive by integrating with Storm.
• Extracted data from SQL Server to create automated visualization reports and dashboards on Tableau.
• Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring and Troubleshooting, Managing and reviewing data backups & log files. Environment: AWS S3, Java, Maven, Python, Spark, Kafka, Elasticsearch, MapR Cluster, Amazon Redshift Db, Shell script, Boto3, pandas, Elasticsearch, certifi, PySpark, Pig, Hive, Oozie, JSON.
.NET Apache AWS Big Data Database Backups Flume Hadoop Hadoop Developer Hbase HDFS Hive Java JavaScript JSON MapReduce Maven Metadata MongoDB MySQL Oozie Pig Python RPA Spark SQL SQL Server Sqoop Tableau Shell Scripts PySpark Kafka
Remove Skill
Hadoop Developer
Information Technology
Oct 2017 - Mar 2018
Chicago, IL
Responsibilities:
• Responsible in Installation and Configuration of Hadoop Eco system components using CDH 5.2 Distribution.
• Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
• MySQL Database administration, Microsoft .Net, Java Script, RPA Tool, kibana reporting (Edge verve Technology)
• Expertise in architecture blueprints. Created AWS infrastructure for first move group application and its deployment.
• Create and Manage User Access Control using AWS IAM policy for distinct group of users.
• Assisted in setting up CI/CD pipeline in AWS.
• Amazon component alerting and monitoring through Cloud Watch.
• AWS cloud trail setup for auditing of various AWS accounts.
• Worked as AWS admin for creating backups/snapshots and maintaining the application.
• Hands on with EC2, ECS, ELB, EBS, S3, VPC, IAM, SQS, RDS, Lambda, Cloud Watch etc.
• Processed Multiple Data sources input to same Reducer using Generic Writable and Multi Input format.
• Worked Big data processing of clinical and non clinical data using Map Reduce.
• Visualize the HDFS data to customer using BI tool with the help of Hive ODBC Driver.
• Customized BI tool for manager team that perform Query analytics using HiveQL.
• Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
• Created Partitions, Buckets based on State to further process using Bucket based Hive joins.
• Created Hive Generic UDF's to process business logic that varies based on policy.
• Moved Relational Data base data using Sqoop into Hive Dynamic partition tables using staging tables.
• Experienced in Monitoring Cluster using Cloudera manager.
• Involved in Discussions with business users to gather the required knowledge.
• Capable of creating real time data streaming solutions and batch style large scale distributed computing applications using Apache Spark, Spark Streaming, Kafka and Flume.
• Analyzing the requirements to develop the framework.
• Designed and developed architecture for data services ecosystem spanning Relational, NoSQL and Big Data technologies.
• Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
• Developed Java Spark streaming scripts to load raw files and corresponding.
• Processed metadata files into AWS S3 and Elasticsearch cluster.
• Developed Python Scripts to get the recent S3 keys from Elasticsearch.
• Elaborated Python Scripts to fetch/get S3 files using Boto3 module.
• Implemented PySpark logic to transform and process various formats of data like XLSX, XLS, JSON, TXT.
• Built scripts to load PySpark processed files into Redshift Db and used diverse PySpark logics.
• Developed scripts to monitor and capture state of each file which is being through.
• Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources.
• Involved in scheduling Oozie workflow engine to run multiple Hives and pig jobs and used Oozie Operational Services for batch processing and scheduling workflows dynamically.
• Included migration of existing applications and development of new applications using AWS cloud services.
• Wrought with data investigation, discovery and mapping tools to scan every single data record from many sources.
• Implemented Shell script to automate the whole process.
• Integrated Apache Storm with Kafka to perform web analytics.
• Uploaded click stream data from Kafka to Hdfs, HBase, and Hive by integrating with Storm.
• Extracted data from SQL Server to create automated visualization reports and dashboards on Tableau.
• Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring and Troubleshooting, Managing and reviewing data backups & log files. Environment: AWS S3, Java, Maven, Python, Spark, Kafka, Elasticsearch, MapR Cluster, Amazon Redshift Db, Shell script, Boto3, pandas, Elasticsearch, certifi, PySpark, Pig, Hive, Oozie, JSON.
.NET Apache AWS Big Data Database Backups Flume Hadoop Hadoop Developer Hbase HDFS Hive Java JavaScript JSON MapReduce Maven Metadata MongoDB MySQL Oozie Pig Python RPA Spark SQL SQL Server Sqoop Tableau Shell Scripts PySpark Kafka
Remove Skill
Hadoop Developer
Information Technology
Oct 2016 - Sep 2017
Chicago, IL
Responsibilities:
• Experience with Hadoop Ecosystem components like HBase, Sqoop, ZooKeeper, Oozie, Hive and Pig with Cloudera Hadoop distribution.
• Conceived and designed custom POCs using Kafka 0.10 and the Twitter Stream in standalone mode architected the front-end near real-time data pub/sub non-blocking messaging system using the Kafka/Confluent.io Enterprise
• Developed PIG and Hive UDF's in java for extended use of PIG and Hive and wrote Pig Scripts for sorting, joining, filtering and grouping the data.
• Worked with NoSQL databases like HBase for creating HBase tables to load large sets of semi structured data coming from various sources.
• Expand programs in Spark based on the application for faster data processing than standard MapReduce programs.
• Custom Kafka broker design to reduce message retention from default 7 day retention to 30 minute retention - architected a light weight Kafka broker
• Elaborated spark programs using Scala, involved in creating Spark SQL Queries and Developed Oozie workflow for spark jobs.
• Prepared the Oozie workflows with Sqoop actions to migrate the data from relational databases like Oracle, Teradata to HDFS.
• Designed and Developed Real Time Stream processing Application using Spark, Kafka, Scala and Hive to perform Streaming ETL and apply Machine Learning.
• Developing Kafka producers and consumers in java and integrating with apache storm and ingesting data into HDFS and HBase by implementing the rules in storm.
• Built a prototype for real time analysis using Spark streaming and Kafka.
• Worked with Kafka for the proof of concept for carrying out log processing on a distributed system. Worked with NoSQL database Hbase to create tables and store data.
• Used the Spark - Cassandra Connector to load data to and from Cassandra. Real time streaming the data using Spark with Kafka.
• Creating Hive tables, dynamic partitions, buckets for sampling, and working on them using HiveQL.
• Used Sqoop to store the data into HBase and Hive.
• Enumerated Hive queries to do analysis of the data and to generate the end reports to be used by business users.
• Worked on scalable distributed computing systems, software architecture, data structures and algorithms using Hadoop, Apache Spark and Apache Storm etc. and ingested streaming data into Hadoop using Spark, Storm Framework and Scala.
• Responsible for developing data pipeline by implementing Kafka producers and consumers and configuring brokers.
• Imported required tables from RDBMS to HDFS using Sqoop and used Spark and Kafka to get real time streaming of data into HBase.
• Good experience with NOSQL databases like MongoDB.
• Managed mission-critical Hadoop cluster and Kafka at production scale, especially Cloudera distribution.
• Creating event processing data pipelines and handling messaging services using Apache Kafka.
• Written spark python for model integration layer.
• Experienced in handling large datasets using Spark in Memory capabilities, using broadcasts variables in Spark, effective & efficient joins, transformations and other capabilities.
• Elaborated Spark code and Spark-SQL/Streaming for faster testing and processing of data.
• Used Spark for interactive queries, processing of streaming data and integration with popular NoSQL database for huge volume of data.
• Developed a data pipeline using Kafka, HBase, Mesos Spark and Hive to ingest, transform and analyzing customer behavioral data. Environment: Hadoop, HDFS, CDH, Pig, Hive, Oozie, ZooKeeper, HBase, Spark, Storm, Spark SQL, NoSQL, Scala, Kafka, Mesos, Mango DB.
Apache Data Integration ETL Hadoop Hadoop Developer Hbase HDFS Hive Java Machine Learning MapReduce MongoDB Oozie Oracle Pig Python Spark SQL Sqoop Teradata Cassandra Kafka
Remove Skill
Hadoop Developer
Information Technology
Oct 2016 - Sep 2017
Chicago, IL
Responsibilities:
• Experience with Hadoop Ecosystem components like HBase, Sqoop, ZooKeeper, Oozie, Hive and Pig with Cloudera Hadoop distribution.
• Conceived and designed custom POCs using Kafka 0.10 and the Twitter Stream in standalone mode architected the front-end near real-time data pub/sub non-blocking messaging system using the Kafka/Confluent.io Enterprise
• Developed PIG and Hive UDF's in java for extended use of PIG and Hive and wrote Pig Scripts for sorting, joining, filtering and grouping the data.
• Worked with NoSQL databases like HBase for creating HBase tables to load large sets of semi structured data coming from various sources.
• Expand programs in Spark based on the application for faster data processing than standard MapReduce programs.
• Custom Kafka broker design to reduce message retention from default 7 day retention to 30 minute retention - architected a light weight Kafka broker
• Elaborated spark programs using Scala, involved in creating Spark SQL Queries and Developed Oozie workflow for spark jobs.
• Prepared the Oozie workflows with Sqoop actions to migrate the data from relational databases like Oracle, Teradata to HDFS.
• Designed and Developed Real Time Stream processing Application using Spark, Kafka, Scala and Hive to perform Streaming ETL and apply Machine Learning.
• Developing Kafka producers and consumers in java and integrating with apache storm and ingesting data into HDFS and HBase by implementing the rules in storm.
• Built a prototype for real time analysis using Spark streaming and Kafka.
• Worked with Kafka for the proof of concept for carrying out log processing on a distributed system. Worked with NoSQL database Hbase to create tables and store data.
• Used the Spark - Cassandra Connector to load data to and from Cassandra. Real time streaming the data using Spark with Kafka.
• Creating Hive tables, dynamic partitions, buckets for sampling, and working on them using HiveQL.
• Used Sqoop to store the data into HBase and Hive.
• Enumerated Hive queries to do analysis of the data and to generate the end reports to be used by business users.
• Worked on scalable distributed computing systems, software architecture, data structures and algorithms using Hadoop, Apache Spark and Apache Storm etc. and ingested streaming data into Hadoop using Spark, Storm Framework and Scala.
• Responsible for developing data pipeline by implementing Kafka producers and consumers and configuring brokers.
• Imported required tables from RDBMS to HDFS using Sqoop and used Spark and Kafka to get real time streaming of data into HBase.
• Good experience with NOSQL databases like MongoDB.
• Managed mission-critical Hadoop cluster and Kafka at production scale, especially Cloudera distribution.
• Creating event processing data pipelines and handling messaging services using Apache Kafka.
• Written spark python for model integration layer.
• Experienced in handling large datasets using Spark in Memory capabilities, using broadcasts variables in Spark, effective & efficient joins, transformations and other capabilities.
• Elaborated Spark code and Spark-SQL/Streaming for faster testing and processing of data.
• Used Spark for interactive queries, processing of streaming data and integration with popular NoSQL database for huge volume of data.
• Developed a data pipeline using Kafka, HBase, Mesos Spark and Hive to ingest, transform and analyzing customer behavioral data. Environment: Hadoop, HDFS, CDH, Pig, Hive, Oozie, ZooKeeper, HBase, Spark, Storm, Spark SQL, NoSQL, Scala, Kafka, Mesos, Mango DB.
Apache Data Integration ETL Hadoop Hadoop Developer Hbase HDFS Hive Java Machine Learning MapReduce MongoDB Oozie Oracle Pig Python Spark SQL Sqoop Teradata Cassandra Kafka
Remove Skill
Build and Release Engineer
Information Technology
Jun 2012 - Oct 2013
Responsibilities:
• As DevOps Engineer, I am responsible for day-to-day build and deployment into QA and pre-prod environments
• Automate the build and deploy of all the internal environments using various continuous integration tools and scripting languages.
• Integrated Subversion into uDeploy to automate the code check-out process
• Maintained and administered GIT source code tool.
• Developed processes, tools, automation for TFS (Team Foundation System) based software for build system and delivering SW Builds.
• Managed build results in uDeploy and deployed using workflows in uDeploy.
• Delivered specific versions of various components of an application into target environments using uDeploy.
• Maintain and track inventory using uDeployand set alerts when the servers are full and need attention.
• Modeled the structure for multi-tiered applications orchestrates the processes to deploy each tier.
• Experience in JIRA to capture, organize and prioritize issues. Experience in partially administering JIRA for issue management.
• Network administration and Network monitoring.
• Developed build and deployment scripts using ANT and MAVEN as build tools in Jenkins to move from one environment to other environments.
• Developing Information Security policies and coordinating the activities required for implementing them. Creating a compliance review plan and conduct periodic review to evaluate the compliance level.
• Implementation of TCP/IP& related Services-DHCP/DNS/WINS.
• Used Hudson/Jenkins for automating Builds and Automating Deployments.
• Troubleshot TCP/IP, layer 1, 2 problems and connectivity issues in multi-protocol Ethernet environment.
• Proficiency in using the iRules for redirection of HTTP based traffic to HTTPS traffic, HTTP acceleration iRule, HTTP header-insertion and modification.
• Used various plug-ins to extend the base functionality of Hudson/Jenkins to deploy, integrate tests and display reports.
• Configuring network services such as DNS/NFS/NIS/NTP for UNIX/Linux Servers.
• Owned build farm and produced effective multiple branch builds to support parallel development
• Owned Release to Production Process, gathered approvals, signoffs from stakeholders and QAs before going to PROD.
• Maintaining the VMware ESXi Servers through VMware Infrastructure Client (vSphere client).
• Managed the Release Communication and Co-ordination Process.
• Developed build scripts using ANT and MAVEN as the build tools for the creation of build artifacts like war or ear files.
• Maintained the Shell and Perl scripts for the automation purposes.
• Involved in editing the existing ANT/MAVEN files in case of errors or changes in the project requirements. Environment: Windows Solaris, UNIX, C++, Java, Eclipse 3.20, Ant, Jenkins, JBoss Application Server, CVS, Subversion, VTFS, Jira and Cygwin, IBMClearcase 7.0. Environment: Java, JSP, J2EE, Servlets, Hibernet, Java Beans, HTML, Reactjs, JavaScript, groovy, JDeveloper, Apache Tomcat, Web server, Oracle, JDBC, XML.
Ant Apache Tomcat C++ ClearCase Continuous Integration Devops DNS Eclipse Groovy HTML J2EE Java Java Servlet Javabeans JavaScript JBOSS BRMS JDBC Jenkins JIRA JSP Linux Maven Network Administration Network Monitoring NFS NIS Oracle Perl Scripting Solaris SVN TCP/IP UNIX VMWare vSphere Windows Apache Git Perl Scripts React.js
Remove Skill
Build and Release Engineer
Information Technology
Jun 2012 - Oct 2013
Responsibilities:
• As DevOps Engineer, I am responsible for day-to-day build and deployment into QA and pre-prod environments
• Automate the build and deploy of all the internal environments using various continuous integration tools and scripting languages.
• Integrated Subversion into uDeploy to automate the code check-out process
• Maintained and administered GIT source code tool.
• Developed processes, tools, automation for TFS (Team Foundation System) based software for build system and delivering SW Builds.
• Managed build results in uDeploy and deployed using workflows in uDeploy.
• Delivered specific versions of various components of an application into target environments using uDeploy.
• Maintain and track inventory using uDeployand set alerts when the servers are full and need attention.
• Modeled the structure for multi-tiered applications orchestrates the processes to deploy each tier.
• Experience in JIRA to capture, organize and prioritize issues. Experience in partially administering JIRA for issue management.
• Network administration and Network monitoring.
• Developed build and deployment scripts using ANT and MAVEN as build tools in Jenkins to move from one environment to other environments.
• Developing Information Security policies and coordinating the activities required for implementing them. Creating a compliance review plan and conduct periodic review to evaluate the compliance level.
• Implementation of TCP/IP& related Services-DHCP/DNS/WINS.
• Used Hudson/Jenkins for automating Builds and Automating Deployments.
• Troubleshot TCP/IP, layer 1, 2 problems and connectivity issues in multi-protocol Ethernet environment.
• Proficiency in using the iRules for redirection of HTTP based traffic to HTTPS traffic, HTTP acceleration iRule, HTTP header-insertion and modification.
• Used various plug-ins to extend the base functionality of Hudson/Jenkins to deploy, integrate tests and display reports.
• Configuring network services such as DNS/NFS/NIS/NTP for UNIX/Linux Servers.
• Owned build farm and produced effective multiple branch builds to support parallel development
• Owned Release to Production Process, gathered approvals, signoffs from stakeholders and QAs before going to PROD.
• Maintaining the VMware ESXi Servers through VMware Infrastructure Client (vSphere client).
• Managed the Release Communication and Co-ordination Process.
• Developed build scripts using ANT and MAVEN as the build tools for the creation of build artifacts like war or ear files.
• Maintained the Shell and Perl scripts for the automation purposes.
• Involved in editing the existing ANT/MAVEN files in case of errors or changes in the project requirements. Environment: Windows Solaris, UNIX, C++, Java, Eclipse 3.20, Ant, Jenkins, JBoss Application Server, CVS, Subversion, VTFS, Jira and Cygwin, IBMClearcase 7.0. Environment: Java, JSP, J2EE, Servlets, Hibernet, Java Beans, HTML, Reactjs, JavaScript, groovy, JDeveloper, Apache Tomcat, Web server, Oracle, JDBC, XML.
Ant Apache Tomcat C++ ClearCase Continuous Integration Devops DNS Eclipse Groovy HTML J2EE Java Java Servlet Javabeans JavaScript JBOSS BRMS JDBC Jenkins JIRA JSP Linux Maven Network Administration Network Monitoring NFS NIS Oracle Perl Scripting Solaris SVN TCP/IP UNIX VMWare vSphere Windows Apache Git Perl Scripts React.js
Remove Skill
Edit Skills
Non-cloudteam Skill
Education
Skills
Hive
2021
12
Oracle
2021
12
Pig
2021
12
Apache
2021
11
Hbase
2021
11
Java
2018
11
Python
2021
11
Eclipse
2021
10
Hadoop
2021
8
Hadoop Developer
2021
8
HDFS
2021
8
Oozie
2021
8
Spark
2021
8
SQL
2021
8
Sqoop
2021
8
Tableau
2021
8
UNIX
2021
8
JavaScript
2018
7
Maven
2018
7
AWS
2021
6
Machine Learning
2021
6
MySQL
2018
6
Javabeans
2013
5
JDBC
2013
5
MongoDB
2018
5
SQL Loader
2021
5
TOAD
2021
5
J2EE
2013
4
Kafka
2018
4
MapReduce
2018
4
XML
2021
4
.NET
2018
3
Apache Tomcat
2013
3
C++
2013
3
Cassandra
2017
3
Git
2013
3
JBOSS BRMS
2013
3
JIRA
2013
3
Linux
2013
3
Metadata
2018
3
Network Administration
2013
3
NFS
2013
3
NIS
2013
3
Perl
2013
3
Shell Scripts
2018
3
Solaris
2013
3
SQL Server
2018
3
SVN
2013
3
Teradata
2017
3
Windows
2013
3
Ant
2013
2
Big Data
2018
2
ClearCase
2013
2
Continuous Integration
2013
2
Data Integration
2017
2
Database Backups
2018
2
Devops
2013
2
DNS
2013
2
ETL
2017
2
Flume
2018
2
Groovy
2013
2
HTML
2013
2
Java Servlet
2013
2
Jenkins
2013
2
JSON
2018
2
JSP
2013
2
Network Monitoring
2013
2
Perl Scripts
2013
2
PySpark
2018
2
React.js
2013
2
RPA
2018
2
Scripting
2013
2
TCP/IP
2013
2
VMWare
2013
2
vSphere
2013
2
C
0
6
C
0
6
Apache Webserver
2013
5
Apache Webserver
2013
5
Java
2015
5
Data General
2015
4
Data General
2015
4
Eclipse
2015
4
Hive
2019
4
MySQL
2018
4
Oracle
2019
4
Pig
2019
4
Hbase
2019
3
Javabeans
2015
3
JavaScript
2015
3
JDBC
2015
3
Maven
2018
3
Pages
2015
3
Pages
2015
3
Python
2019
3
RHadoop
2019
3
RHadoop
2019
3
AJAX
2015
2
AJAX
2015
2
Hibernate
2015
2
Hibernate
2015
2
iWeb
2013
2
iWeb
2013
2
J2EE
2015
2
jQuery
2018
2
jQuery
2018
2
MVC
2015
2
MVC
2015
2
Oracle 11i
2015
2
Oracle 11i
2015
2
SAP Detailed Scheduling
2019
2
SAP Detailed Scheduling
2019
2
SCSS
2015
2
SCSS
2015
2
Spring
2015
2
Spring
2015
2
Stored Procedure
2015
2
Stored Procedure
2015
2
Struts
2015
2
Struts
2015
2
Tableau
2019
2
Triggers
2015
2
Triggers
2015
2
UNIX
2019
2
WebServices
2015
2
WebServices
2015
2
.NET
2018
1
Apache
2013
1
Apache Tomcat
2013
1
Application Support
2019
1
Application Support
2019
1
AWS EC2
2019
1
AWS EC2
2019
1
Business Requirements
2019
1
Business Requirements
2019
1
C++
2013
1
Cassandra
2017
1
CVS
2013
1
CVS
2013
1
Database Management
2019
1
Database Management
2019
1
Design Patterns
0
1
Design Patterns
0
1
Ethernet
2013
1
Ethernet
2013
1
Fiddler
2019
1
Fiddler
2019
1
Gateways
2019
1
Gateways
2019
1
Git
2013
1
IBM WebSphere
0
1
IBM WebSphere
0
1
Information Security
2013
1
Information Security
2013
1
JBOSS BRMS
2013
1
Jdeveloper
2013
1
Jdeveloper
2013
1
JIRA
2013
1
JScript
2018
1
JScript
2018
1
LAN
0
1
LAN
0
1
Linux
2013
1
Metadata
2018
1
Microsoft SMS Server
2018
1
Microsoft SMS Server
2018
1
MongoDB
2017
1
NANT
2013
1
NANT
2013
1
Network Administration
2013
1
NFS
2013
1
NIS
2013
1
ODBC (Open Database Connectivity)
2018
1
ODBC (Open Database Connectivity)
2018
1
OLAP
0
1
OLAP
0
1
Oracle SQL*Plus
2019
1
Oracle SQL*Plus
2019
1
Perl
2013
1
Problem Solving
0
1
Problem Solving
0
1
Puppet
2019
1
Puppet
2019
1
RDBMS
2017
1
RDBMS
2017
1
RedHat
2019
1
RedHat
2019
1
Requirements Gathering
2019
1
Requirements Gathering
2019
1
REST
0
1
REST
0
1
Shell Scripts
2018
1
SOAP
0
1
SOAP
0
1
Software Architect
2017
1
Software Architect
2017
1
Solaris
2013
1
SQL Loader
2019
1
SQL Server
0
1
SSIS
0
1
SSIS
0
1
SVN
0
1
Technical Support
2019
1
Technical Support
2019
1
Teradata
2017
1
TOAD
2019
1
Web Weaver
2013
1
Web Weaver
2013
1
Windows
2013
1
Windows 2000
0
1
Windows 2000
0
1
Windows 98
0
1
Windows 98
0
1
Windows XP
0
1
Windows XP
0
1
Winstall
2019
1
Winstall
2019
1
WSDL
0
1
WSDL
0
1
XSL
0
1
XSL
0
1
XSLT
0
1
XSLT
0
1