Uploaded File
add photo
Ravi
rteja8442@gmail.com
704-709-0266
Charlotte, NC 28299
Data Scientist
9 years experience C2C
0
Recommendations
Average rating
70
Profile views
Summary

Over 6 plus years of professional experience in Data Science, with sound knowledge in concepts related to Machine Learning, Data Mining and Statistical Analysis.

  • Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing data mining and reporting solutions that scale across a massive volume of structured and unstructured data. Machine learning algorithms. Python 3.5 (NumPy, Pandas, Matplotlib and Sci-kit-learn), Decision Tree, Random Forest, Naïve Bayes, Logistic Regression, Linear Regression, Multiple regression, Cluster Analysis, Neural Networks, KNN, SVM, k-means from Mathematical perspective.
  • Experience in implementing data analysis with various analytic tools, such as Anaconda 4.0 Jupiter Notebook 4.X, R (ggplot2, Caret, dplyr) and Excel.
  • Involved in the entire data science project life cycle and actively involved in all the phases including data extraction, data cleaning, statistical modelling, and data visualization with large data sets of structured and unstructured data.
  • Skilled in Advanced Regression Modelling, Multivariate Analysis, Model Building, Business Intelligence tools and application of Statistical Concepts.
  • Ability to write and optimize diverse SQL queries, working knowledge of RDBMS like SQL server 2008, NoSQL databases like MongoDB 3.2.
  • Experience with ELK (Elasticsearch) stack technologies
  • Extensive experience in Text Analytics, Statistical Machine learning and Data Mining, providing solutions to various business problems and generating data visualizations using R, Python.
  • Performed Information Extraction using NLP algorithms coupled with Deep learning (ANN and CNN, RNN, LSTM, Encoders, Embedders), Keras and Tensor Flow.
  • Trained convolutional neural network using Tenser flow, to collect image features from images of items provided by each seller.
  • Good nowledge in Computer vision and ROS development.
  • Good knowledge on use of Microservices FOR developing a service-oriented architecture.
  • Trained and tested various object detection models.
  • Excellent understanding of SDLC, Agile and Scrum development methodology.
  • Have good experience, working with ARIMA parametric time series models.
  • Created customs reports and dashboards using simple drag and drop functionality by use of Sageworks.
  • Assist in creation of forecasting tools and models using mathematical and statistical techniques, to determine continuous improvement opportunities.
  • Experience in tools and concepts such as Hadoop, MapReduce, Spark 1.6, PySpark, SparkSQL, HDFS, Hive 1.X., related to Big Data.
  • Extracted data from HDFS and prepared data for exploratory analysis through data munging.
  • Proficient in Predictive Modelling, methods related to Data Mining, Factor Analysis, ANOVA, Hypothetical Testing, Normal Distributions and other advanced statistical and econometric techniques.
  • Worked as a member of an experienced optimization team to design creative variants to be A/B tested. Focused on developing, improving, and testing the subscriptions landing pages as well as the crosswords page.
  • Hands on experience and in provisioning virtual clusters under Amazon Web Service (AWS) cloud which includes services like Elastic compute cloud (EC2), S3, and EMR.
  • Performing Map Reduce jobs in Hadoop and implemented spark analysis using Python, for performing machine learning & predictive analytics on AWS platform.
  • Experienced in Data Integration, Data Validation and Data Quality control for ETL process and Data Warehousing, using MS Visual Studio SSIS, SSAS, SSRS.
  • Worked with complex applications such as MATLAB and SPSS to develop neural networks and performed cluster analysis.
  • Strong experience and good knowledge in data visualization using tools such as Tableau for creating line and scatterplots, Bar-charts, Histograms, Pie-chart, Dot-charts, Boxplots, Timeseries, Error Bars, Multiple Charts types, Multiple Axes, subplots etc.
  • Automated recurring reports using SQL and Python and visualized them on BI platform like Tableau.
  • Experience in visualization tools like Tableau 9.X, 10.X for creating dashboards.
  • Used version control tools like Git 2.X and VM.
  • Passionate about gleaning insightful information from massive data assets and developing a culture of sound, data-driven decision making.
  • Experience in Visual Basic for Applications and VB programming languages to work with developing applications.
  • Taking responsibility for technical problem solving, creatively meeting product objectives and developing best practices
  • Excellent communication skills (verbal and written) to communicate with clients and team prepare + deliver effective presentations.
  • Ability to maintain a fun, casual, professional and productive team atmosphere.

Experience
Data Scientist
Mar 2018 - present
Providence, RI
Responsibilities:
  • Used pandas, NumPy, Seaborn, SciPy, matplotlib, sci-kit-learn in python for developing various machine learning algorithms.
  • Participated in all phases of datamining data collection, data cleaning, developing models, validation, visualization and performed Gap analysis.
  • Data Manipulation and Aggregation from a different source using Toad, Business Objects, Power BI, and SmartView.
  • Extracted company relations from unstructured news to form a relation map surrounding a company.
  • Managed with portfolio risk modeling and diversified system of mutual funds that are grouped together to provide an expected return with a corresponding amount of risk.
  • Performed the risk at funds which are invested in the model portfolio of payroll and taken the future contributions which are invested in the model portfolio from what we have chosen.
  • Maintained the model portfolio's balance and rebalanced regularly to return it to the target allocation.
  • Monitored the portfolio as often or as infrequently as we can wish via on-line access to the accounts to avoid risk by applying ML algorithms.
  • Applied distant supervision, CNN model and attention mechanism to obtain relation label for each company and achieved 91% accuracy in relation classification.
  • Performed Information Extraction using NLP algorithms coupled with Deep learning (ANN and CNN), Keras and TensorFlow.
  • Implemented Agile Methodology for building an internal application.
  • Programmed a utility in Python that used multiple packages (SciPy, NumPy, pandas)
  • Implemented Classification using supervised algorithms like Logistic Regression, Decision trees, KNN, Naive Bayes.
  • Worked on predictive and what-if analysis using R from HDFS and successfully loaded files to HDFS from Teradata and loaded from HDFS to HIVE.
  • Prepared data visualization reports for the management using R.
  • Experience in implementing data analysis with various analytic tools, such as Anaconda 4.0 Jupiter Notebook 4.X, R (ggplot2, Caret, dplyr) and Excel.
  • Experience working in the Python 2.7 and 3 environments using the following libraries Anaconda, NumPy, Pandas, Scipy, iPython etc. Excellence in at least one of these is highly desirable.
  • Loaded and saved the models by using the Pytorch and Reinforcement learning (RL)
  • Developed MapReduce/Spark, R modules for machine learning & predictive analytics in Hadoop on AWS. Implemented a R-based distributed random forest.
  • Knowledge on time series analysis using AR, MA, ARIMA, GARCH and ARCH model.
  • Data transformation from various resources, data organization, features extraction from raw and stored.
  • Experience with ELK (Elasticsearch) stack technologies.
  • Work with the capacity planning director in building and maintaining highly advanced capacity plans for the Individual Client Solutions team.
  • Handled importing data from various data sources, performed transformations using Hive, MapReduce.
  • Recommend / developed sensor and embedded image processing HW and overall system design & architecture for a multi-camera computer-vision based solution.
  • Project experience in data mining, Segmentation analysis, business forecasting and association rule mining using large data sets with Machine learning.
  • Worked as a member of the experience optimization team to design creative variants to be A/B tested Focused on developing, improving, and testing the subscriptions landing pages as well as the crosswords page.
  • Setup storage and data analysis tools in AWS cloud computing infrastructure.
  • Created financial package that supports 3-Year financial plan for all AWS cloud services infrastructure expenses.
  • Interaction with Business Analyst, SMEs, and other Data Architects to understand Business needs and functionality for various project solutions.
  • Developed normalized Logical and Physical database models for designing an OLTP application.
  • Researched, evaluated, architected, and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for the clients.
  • Identifying and executing process improvements, hands-on in various technologies such as Oracle, Informatica, Business Objects.
  • Used advanced statistical techniques coupled with a deep understanding of industry knowledge and market to improve sales, market share, and marketing ROI of the company.
  • Maintained communications with Marketing and reviews Marketing material for accuracy.
  • Prepared ROI analysis and statistical reports analyzed and interpreted data Developed ad hoc reports through a variety of data warehouse applications by pricing analysis.
  • Participated in all phases of Data-mining, Data-collection, Data cleaning, Developing-Models, validation, Visualization and performed Gap Analysis.
  • Designed 3NF data models for both ODS, OLTP systems and dimensional data models using Star and Snowflake Schemas. Have done Text analytics on historical email subject lines to retrieve effective Keywords and suggest them to the creative team to create a new subject line that would increase open rates and delivery rates.
  • Knowledge on time series analysis using AR, MA, ARIMA, GARCH and ARCH model.
  • Designed dashboards with Tableau and provided complex reports including summaries, charts, and graphs to interpret findings to team and stakeholders.
  • Identified process improvements that significantly reduce workloads and improved quality.
  • Analyzed the email user click history and third-party data for pattern recognition and to support and change targeting algorithms.
  • Supported client by developing Machine Learning Algorithms on Big Data using PySpark to analyze transaction fraud, Cluster Analysis etc.
  • Used GCC to compile C++ in order to make the system interpret the C code and convert it into a machine language that the system understands.
  • Created Revenue optimization algorithm to divert click traffic to different advertiser throughout the day to maximize Revenue.
AWS Business Analysis Data Analysis Data Architecture Data Cleansing Data Mining Data Visualization Data Warehousing Gap Analysis Hadoop HDFS Hive Informatica Logistic Regression Machine Learning MapReduce Oracle Pattern Recognition Spark Tableau Teradata TOAD
Remove Skill
Data Scientist
Nov 2016 - Feb 2018
Responsibilities:
  • Utilized Spark, HBase, MongoDB, Python (MLlib) and a broad variety of machine learning methods including classifications, regressions, dimensionally reduction etc. and utilized the engine to increase user lifetime by 45% and triple user conversations for target categories.
  • Performed Data Profiling to learn about behavior with various features such as traffic pattern, location, time, Date and Time etc.
  • Worked with Sales and Marketing team for Partner and collaborated with a cross-functional team to frame and answer important data questions.
  • Application of various machine learning algorithms and statistical Modelling like decision trees, regression models, neural networks, SVM, clustering to identify volume using Scikit-learn package in python.
  • Performed data visualization using Tableau and generated dashboards to present the findings.
  • Recommended and evaluated marketing approaches based on quality analytics of customer consuming behavior
  • Evaluated models using Cross Validation, Log loss function, ROC curves and used AUC for feature selection.
  • Prototyping and experimenting ML/DL algorithms and integrating into production system for different business needs.
  • Analyse traffic patterns by calculating autocorrelation with different time lags.
  • Addressed overfitting by implementing the algorithm regularization methods like L2 and L1.
  • Used Principal Component Analysis in feature engineering to analyze high dimensional data.
  • Created and designed reports that will use gathered metrics to infer and draw logical conclusions from past and future behavior.
  • Performed Logistic Regression, Random forest, Decision Tree, SVM to classify package is going to deliver on time for the new route.
  • Used Python and Spark to implement different machine learning algorithms including Generalized Linear Model, SVM, Random Forest and Neural Network
  • Used MLlib and Spark's Machine learning library to build and evaluate different models.
  • Implemented rule-based expertise system from the results of exploratory analysis and information gathered from the people from different departments.
  • Performed Data Cleaning, features scaling and features engineering using pandas and NumPy packages in python.
  • Communicated the results with operations team for taking best decisions.
  • Collected data needs and requirements by Interacting with the other departments.
  • Developed MapReduce pipeline for feature extraction using Hive.
  • Created Data Quality scripts using SQL and Hive to validate successful data load and ensured the quality of the data. Created various types of data visualizations using Python and Tableau.
  • Interacted with Business Analyst, SMEs and Data Architects to understand business needs and functionality for various project solutions. Environment: Python, R, Machine learning, CDH5, HDFS, Hive, AWS, Impala, Linux, Spark, Tableau, SQL Server 2012, Microsoft Excel, MATLAB, Spark SQL and PySpark.
AWS Business Analysis Data Architecture Data Cleansing Data Profiling Data Visualization Hbase HDFS Hive impala Logistic Regression Machine Learning MapReduce Spark SQL SQL Server Tableau
Remove Skill
Machine Learning Engineer
Information Technology
Feb 2015 - Oct 2016
Tampa, FL
Responsibilities:
  • Enhanced Data collection procedures to include information that is relevant for building analytic systems and created a value from data by performing advanced analytics and statistical techniques to determine deep insights and optimal solution architecture.
  • Maintained and developed complex SQL queries, stored procedures, views, functions, and reports that qualify customer requirements using Microsoft SQL Server 2008 R2.
  • Supported Sales, engaged in management planning and decision making on sales incentives through producing, developing and maintaining financial models and through reporting and sensitivity analysis by customer segment.
  • Worked with the ETL team to document the transformation rules for Data migration from OLTP to Warehouse environment, for reporting purposes.
  • Used Pandas, NumPy, Seaborn, SciPy, Matplotlib, Scikit-learn.
  • Worked on data modelling and produced data mapping and data definition documentation.
  • Developed test plans to ensure successful delivery of the project. Employed performance analytics predicated on high-quality data to develop reports and dashboards with actionable insights.
  • Precisely development and implementation several types of sub-reports, drill-down reports, summary reports, parameterized reports, and ad-hoc reports using SSRS through mailing server subscriptions & SharePoint server.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conduct data analysis.
  • Resolved data related issues through assessing data quality, data consolidation, evaluating existing data sources.
  • Generated ad-hoc reports using Crystal Reports 9 and SQL Server Reporting Services (SSRS).
  • Generated the reports and visualizations based on the insights, using Tableau and developed dashboards for the company insight teams.
  • Worked closely with data architects to review all the conceptual, logical and physical database design models with respect to functions, definition, maintenance review and support data analysis, Data quality and ETL design that feeds the logical data models.
  • Created financial package that support 3-Year financial plan for all AWS cloud services infrastructure expenses. Environment: SQL Server 2008R2/2005 Enterprise, SSRS, SSIS, Crystal Reports, Windows Enterprise Server 2000, DTS, SQL Profiler, Tableau, Qlik View and Django.
AWS Crystal Reports ETL SQL SQL Server SSIS Stored Procedure
Remove Skill
Python developer
Information Technology
Apr 2014 - Jan 2015
Bloomington, IL
Responsibilities:
  • Developed a new product for the unsecured loans featuring Term Business and personal loans.
  • Developed and implemented the user registration and login feature for the application process from scratch by extending Django user model.
  • Used Restful web service calls for the validation.
  • Developed a fully automated continuous integration system using Git, Gerrit, Jenkins, MySQL and custom tools developed in Python and Bash.
  • Implemented schema and data migration for the postgresdatabase using South migration tool.
  • Developed user-friendly modals for the form submissions using simplemodal.js, JQuery, Ajax and JavaScript.
  • Experience in building the war with help the putty and deployed into cloud environment using the cloud controller and experience in solving the cloud issue.
  • Worked closely with Client managers/Business Analysts of the bank to drive technical solutions, design and provide development estimates for schedule and effort.
  • Worked very closely with product owners, project managers and vendors to satisfy all the business needs.
  • Used Django framework for database application development.
  • Dynamic, hard-working, ability to work in-groups as well as independently with initiative to learn new technologies/tool quickly and emphasis on delivering quality services.
  • Have strong ability to build productive relationships with peers, management, and clients using strong communication, interpersonal, organizational, and planning skills. Environment: Python, Django, JSP, Oracle, Java, MySQL, Linux, HTML, CSS.
AJAX Continuous Integration Django Git HTML JavaScript Jenkins jQuery JSP Linux MySQL Oracle Project Management Python
Remove Skill
Python Developer
Information Technology
Feb 2013 - Mar 2014
Santa, ID
Responsibilities:
  • Involved in building database model, APIs and views utilizing Python, in order to build an interactive web-based solution.
  • Used data types like dictionaries, tuples and object -concepts based inheritance features for making complex algorithms of networks.
  • Worked on Python open stack API's.
  • Carried out various mathematical operations for calculation purpose using python libraries.
  • Managed large datasets using Panda data frames and MySQL.
  • Worked with JSON based REST Web services.
  • Involved in Agile Methodologies and SCRUM Process.
  • Creating unit test/regression test framework for working/new code.
  • Using Subversion version control tool to coordinate team-development.
  • Developed SQL Queries, Stored Procedures, and Triggers Using Oracle, SQL, PL/SQL.
  • Responsible for debugging and troubleshooting the web application.
  • Supported user groups by handling target-related software issues/service requests, identifying/fixing bugs.
  • Developed Views and Templates with Django view, controller and template language to create a user-friendly website interface.
  • Used Django APIs for database access.
  • Used Python for XML, JSON processing, data exchange and business logic implementation.
  • Used Python scripts to update the content in database and manipulate files.
  • Created UI using JavaScript and HTML5. Designed and developed data management system using MongoDB.
  • Proficient in Software Design and Development with a solid background in developing. Environment: Python, Django, JSP, Oracle, Java, MySQL, Linux, HTML and CSS.
Agile Methodology Django HTML JavaScript JSON JSP Linux MongoDB MySQL Oracle PL/SQL Python REST SQL Stored Procedure SVN Triggers UI XML
Remove Skill
Edit Skills
Non-cloudteam Skill
Education
Skills
AWS
2021
4
SQL
2018
4
Business Analysis
2021
3
Data Architecture
2021
3
Data Cleansing
2021
3
Data Visualization
2021
3
HDFS
2021
3
Hive
2021
3
Logistic Regression
2021
3
Machine Learning
2021
3
MapReduce
2021
3
Oracle
2021
3
Spark
2021
3
SQL Server
2018
3
Tableau
2021
3
Data Profiling
2018
2
Django
2015
2
Hbase
2018
2
HTML
2015
2
impala
2018
2
JavaScript
2015
2
JSP
2015
2
Linux
2015
2
MySQL
2015
2
Python
2015
2
Stored Procedure
2016
2
Agile Methodology
2014
1
AJAX
2015
1
Continuous Integration
2015
1
Crystal Reports
2016
1
Data Analysis
2021
1
Data Mining
2021
1
Data Warehousing
2021
1
ETL
2016
1
Gap Analysis
2021
1
Git
2015
1
Hadoop
2021
1
Informatica
2021
1
Jenkins
2015
1
jQuery
2015
1
JSON
2014
1
MongoDB
2014
1
Pattern Recognition
2021
1
PL/SQL
2014
1
Project Management
2015
1
REST
2014
1
SSIS
2016
1
SVN
2014
1
Teradata
2021
1
TOAD
2021
1
Triggers
2014
1
UI
2014
1
XML
2014
1
Big Data
0
1
Business Intelligence
0
1
Data Integration
0
1
Data Validation
0
1
Linear Regression
0
1
Statistical Analysis
0
1
Visual Studio
0
1