Hemanthsai Uppu
Email :-Phone Number : -
LinkedIn - https://www.linkedin.com/in/hemanth-sai-u-17b055131
Senior Data Engineer | Data Science
HIGHLIGHTS OF QUALIFICATIONS:
• Over 8 years solid experience in Data Science fields in Python, Numpy, pandas, Django, OOPS Programming,
Machine Learning and in Chatbots, Point of Sales Industry, Ecommerce, HealthCare Industry, Car Industry
(BMW). Also worked on some of theFreelance Projects like developing an application that can monitor
network and serversetc.
• Python API development with Flask, Django using concepts such as Class inheritance and more advanced
python concepts such as programming content generators and decorators.
• Strong knowledge of machine learning algorithms like regression, neural networks and techniques like pattern
creations, words and sentences grouping using NLTK, Spacy.
• Having profound knowledge on Data Processing, Application Development and End-to- End flow of data
science and machine learning like Preprocessing, EDA, Feature Engineering, Hyperparameters tuning, Model
Training, Prediction and Performance Metrics etc.
• Having Artificial Intelligence Graduation certificate from Georgian College
• Having particularly pleasant experience in Python, Machine Learning and Data Science with excellent
technological skills especially Python, AWS, AWS Lambda, Aws Glue,Machine Learning, Regression Analysis,
NLTK, Spacy, Sklearn, Keras, TensorFlow, Neural Networks, Pandas, Tableau, Hadoop, Microsoft Power BI,
ETL, Git, Jira, Confluence, Bitbucket, Terraform …
• Highly educated Artificial Intelligence Post Graduation which includes the latest and updated technologies
like Auto ML Vision, Teapot, Reinforcement Learning using MonteCarlo methods, Neural Networks etc.
SKILLS:
•
Python. Machine Learning, PySpark, Tensorflow, Linear and Logistic Regression and other
Algorithms, Nltk, Regex, Pandas, Keras, Jenkins, Scrapy, django, aws,Docker, lambda, aws,
glue,Hadoop, mongodb, postgres, sql,k-sql, stackdriver, Kubernetes, celery, rabbit-mq,
swagger-api, Prometheus Database
EDUCATION & CERTIFICATION:
PG-Artificial Intelligence - Georgian College, Canada
May.2021 – Dec.2021
B.E-Computer Science - Prathyusha Engineering College, India
April.2012 – June .2016
PROFESSIONAL EXPERIENCE:
Quantiphi , Toronto, Canada
Dec. 2022 – Current
Data Engineer / Data Scientist (AWS/Python/NLP)
• Working on manipulating the Sunlife Insurance Data using Pyspark and AWS Glue.
• Worked in the development Pipelines, Automation of Glue Jobs, ETL and Deployments.
• Led the implementation of Apache Airflow to orchestrate and automate ETL workflows in a cloud-based
insurance data migration project, enhancing operational efficiency and reducing data processing time by 30%.
• Worked on terraform to automate the glue jobs deployments as well.
• Work Includes debugging, Performance Improvement using different techniques in Glue Version 3 and Glue
Version 4.
• Worked on couple of analytics using naming conventions of insurances and also withTableau.
Anuvu , Montreal, Canada
Jan. 2022 – Nov. 2022
Python Developer (Flask/AWS/Python/Pandas)
• Product Based company which develops the softwares for the flights which will have the Wifi, movies and
•
•
•
•
•
•
•
•
•
other stuff. Especially we handle the WIFI service where users can buy the Wifi from the flights.
Worked in the development of new api services as micro services using Flask, Docker and Swagger UI etc.
Worked in postgres , redis and memcache. Using Pandas manipulated some of the documents.
Work also includes the deployment and CI/CD pipe-lining through jenkins and deploying using kubernetes.
Since it’s already developed project we develop new services with the existing repositories
through the inheritance of those classes.
Designed and configured Airflow DAGs (Directed Acyclic Graphs) to manage complex data pipelines, ensuring
seamless integration and real-time synchronization of legacy insurance systems with the new cloud
infrastructure
Also worked in different VSMU’s as part of the debugging and also worked with production servers which will
also have the access to ping the satellite.
Apart from that, by following some of the strategies like peer programming and group meetings the work
goes really fast and efficient. It always includes my inputs and plans to make this task or project to get to
work.
As a part of team member, contributed into development strategies by following codingstandards like writing
test-cases immediately with the development, peer reviewing approvals to do the merge with master etc.
Based on the sprint, we do debugging, solving the issues by tracking the environments like Dev, Staging and
Pre-Prod through the Grafana.
Scorg International Pvt Ltd
May. 2021 – July .2021
Client : COGNIZANT (BMW - M4 Systems), Hyderabad, India
Sr. Python Engineer (AWS/Python)
• Understand stakeholder’s business problems and identify possible technical solutions toeffectively provide
the project deliverables on time.
• Use Python into lambda and Glue jobs to create a better and efficient data processing and data manipulation.
• Creating a pipeline which included lambda, glue-job, redshift and finally into elastic search, Kibana for
dashboards and analytics. These whole pipelines will be created through the terraform.
• Creation of dashboards to visualize the analysis of the data using Kibana and Elasticsearch.
• Used terraform by using python Inheritance and object oriented programming to automate the data
processing in AWS.
• Leverage industry best practices to design, test, implement and support a solution to migrate all the current
codes to AWS using Lambda.
• Managing each task day-to-day to Create python codes with different packages/modules using spark, pandas
to do a variety of data analysis and create pipeline and deploy into production from end-to-end.
• Facilitate collaboration with other developers, product owners, product teams and designers to solve
interesting and challenging problems through delivery of various analytics solutions.
• Provide knowledge transfer to other team members and be a part of an agile cross- functional team.
UST Global Services (Health Care - Anthem), Trivandrum, India
Jan. 2020 – March. 2021
Sr. Software Engineer (AWS/Python/Django/Pandas)
• Working with Anthem (client), to manipulate huge health-care data and return the data
as expected Using pandas and pyspark as well as Numpy.
• Worked with the Kafka Streaming data will be taken and stored into AWS-SQL Databases. Worked to
create a Glue Job with spark, where I process the dataframesafter cleansing, manipulating and
transformations (as json).
• Not only the Kafka but also worked with the structured and Un-Structured data and
preparation of data into the required format using pandas and numpy etc.
• Provide knowledge transfer to other team members and assisting them to understandthe project and
terminologies that are being used.
• Been a crucial collaborator to develop and handling some of the modules or featuresfrom end-to-end.
• Working with the deliverables and improvements to understand and automate thecurrent existing
•
•
•
•
environment.
Assigning the day-to-day tasks and back-tracking with the team as well as collaboratinginto different issues
facing by each-others to solve the blockers to move forward easily.
Deployment and evaluating the solutions in staging and production environments.
Implemented sharding and replication techniques to distribute data and ensure high availability.
Used Django and MongoDB into ORM as well as used django-queryset for processing and
prometheus for handling Time-series data,
EY – Trivandrum, India
June .2019 – Dec .2019
Client : LABGLO - (Nokia Tax Systems)
Data Scientist & Python Developer (AWS/Python/Django)
• Working with different emails to scrape the data using NLTK and preparation of data,into specified data
format that is requested by the Tax Filing Team.
• Developing a Django Application to process, visualize and make available the data tocollect and analyze
for other teams.
• Collecting and processing the data from various sources like SAP, Emails, Excel and processing these data
based on the various kinds of data and other related information.
• Deployment of application onto the Kubernetes and docker to deploy them into threestages like
development, staging, production.
• Automating the Ticket Creation SNOW based the data processed from the sources where this Snow
Tickets allow teams for filing the taxes and addressing the issues ofdata related that must be solved.
• Collaborating with the other teams to understand the requirements and automate themas per their
requirements.
• Analyzing the data and visualizing these data using Power BI and preparation ofdocuments based
on these analytics.
Labglo Technologies LLP -India (Ecommerce – Price.com),
Nov. 2018 – Dec.2019
Data Engineer and Application Developer (Python/Scrapy/Django/Numpy)
• Developed an application from end-to-end as a Single Collaborator and Crucial one too.
• Worked with scrapy to scrap different E-commerce websites and preparation as well as
manipulation of the data, then publishing into Price.com.
• Scraping different websites using scrapy and processing with REGEX to process thescraped website
data and arrange them with proper format of database.
• Worked with Elastic-search services by indexing the data with postgres and to provide afeature like search
and visualize the price differences with the other websites using Kibana. Also used Numpy to perform
Dimensional Operations on Indexing data.
• Regression Analysis on prices and predicted the prices to understand the prices on thewebsites for
products by analytics.
• Worked on managing and processing SQL DB queries through pyspark.
• Used different SQL DB in using various SQL clauses (SELECT, FROM, WHERE, JOIN, GROUP BY, HAVING, ORDER
BY) to perform data operations effectively.
• Developed different web services that can provide data regarding unique features tomaintain and
process the data including the scrapers. The Application was fully developed using DJANGO.
• Deployment of these services into AWS environment and providing the access to publicenvironment.
• Creating a CI/CD Pipelines using GIT and Docker. Also worked on day-to-dayImprovements of
tasks and had a crucial part of releasing different versions.
Deliverect -Belgium (Delivery – POS Systems)
June .2018 – Sep .2018
Application Developer (Python/Flask)
• Worked as a python developer for integrating Delivery platform data with POS Systems.working with the
tools like Docker, celery etc.
• Played as a prominent collaborator in the development and Integration of code.
• Worked with the cloud environments like Google cloud engine integrated thestackdriver.
•
•
•
Celery was used to automate each service that executes automatically as per designedin docker.
Swagger API was used to automate and maintain the API Services.
Implementation of StackDriver into Docker and initiated as a service to monitor theapplication server.
UST Global Services - (Chatbot Designer), Trivandrum
Nov.2016 – June .2018
Machine Learning Engineer (Python)
• Worked on development of chatbot designer (LIKE RASA), where we are creatingdifferent chatbots
using trained NER (Natural Entity Recognition).
• Used Regex to manipulate the utterances with Object Oriented Programming withpython.
• Also worked on document processing project as a part of this chatbot using deeplearning.
• Worked on training NER (Natural Entity Recognition) by implementing Flask, Gunicorn asan API Services.
• Training the different utterances to NER and evaluating the application predictions livelythrough the
Chatbot UI Chatbox.
• Developed a scraper to get the live data like Climate and other Information from someof the sources and
google based on the client requirement.
• Manipulation and processing the data by some of the NLTK tools like Tokenization,Lemmatization etc.
• Developed some of the features like story shifting while having a conversation with theBOT’s etc.
FREELANCE PROJECTS:
SURVEY SNEAK(FREELANCE)
March.2019 – Oct .2019
Developed the application using Django. Modeling the Database based on the front-end design.Architectural Design for
the whole Application as well as UI Mock-up Screens. Worked on API Services that are to be called onto each page as
well as security using token system.
CEBURU(FREELANCE)
May.2018 – Dec. 2018
Developed an application to analyze the bandwidth data, network data. Continuous Monitoringof server data and
other system related data through python API Services. Analyzed the data ofsystem data to monitor health of system
on every hour through scheduled Services.