Hemanthsai

Hemanthsai

$20/hr
Web dev, GenAI apps, scraping, cloud data, docs handling, Python automation & deployment
Reply rate:
-
Availability:
Hourly ($/hour)
Age:
30 years old
Location:
Trivandrum, Kerala, India
Experience:
9 years
Hemanthsai Uppu Email :-Phone Number : - LinkedIn - https://www.linkedin.com/in/hemanth-sai-u-17b055131 Senior Data Engineer | Data Science HIGHLIGHTS OF QUALIFICATIONS: • Over 8 years solid experience in Data Science fields in Python, Numpy, pandas, Django, OOPS Programming, Machine Learning and in Chatbots, Point of Sales Industry, Ecommerce, HealthCare Industry, Car Industry (BMW). Also worked on some of theFreelance Projects like developing an application that can monitor network and serversetc. • Python API development with Flask, Django using concepts such as Class inheritance and more advanced python concepts such as programming content generators and decorators. • Strong knowledge of machine learning algorithms like regression, neural networks and techniques like pattern creations, words and sentences grouping using NLTK, Spacy. • Having profound knowledge on Data Processing, Application Development and End-to- End flow of data science and machine learning like Preprocessing, EDA, Feature Engineering, Hyperparameters tuning, Model Training, Prediction and Performance Metrics etc. • Having Artificial Intelligence Graduation certificate from Georgian College • Having particularly pleasant experience in Python, Machine Learning and Data Science with excellent technological skills especially Python, AWS, AWS Lambda, Aws Glue,Machine Learning, Regression Analysis, NLTK, Spacy, Sklearn, Keras, TensorFlow, Neural Networks, Pandas, Tableau, Hadoop, Microsoft Power BI, ETL, Git, Jira, Confluence, Bitbucket, Terraform … • Highly educated Artificial Intelligence Post Graduation which includes the latest and updated technologies like Auto ML Vision, Teapot, Reinforcement Learning using MonteCarlo methods, Neural Networks etc. SKILLS: • Python. Machine Learning, PySpark, Tensorflow, Linear and Logistic Regression and other Algorithms, Nltk, Regex, Pandas, Keras, Jenkins, Scrapy, django, aws,Docker, lambda, aws, glue,Hadoop, mongodb, postgres, sql,k-sql, stackdriver, Kubernetes, celery, rabbit-mq, swagger-api, Prometheus Database EDUCATION & CERTIFICATION: PG-Artificial Intelligence - Georgian College, Canada May.2021 – Dec.2021 B.E-Computer Science - Prathyusha Engineering College, India April.2012 – June .2016 PROFESSIONAL EXPERIENCE: Quantiphi , Toronto, Canada Dec. 2022 – Current Data Engineer / Data Scientist (AWS/Python/NLP) • Working on manipulating the Sunlife Insurance Data using Pyspark and AWS Glue. • Worked in the development Pipelines, Automation of Glue Jobs, ETL and Deployments. • Led the implementation of Apache Airflow to orchestrate and automate ETL workflows in a cloud-based insurance data migration project, enhancing operational efficiency and reducing data processing time by 30%. • Worked on terraform to automate the glue jobs deployments as well. • Work Includes debugging, Performance Improvement using different techniques in Glue Version 3 and Glue Version 4. • Worked on couple of analytics using naming conventions of insurances and also withTableau. Anuvu , Montreal, Canada Jan. 2022 – Nov. 2022 Python Developer (Flask/AWS/Python/Pandas) • Product Based company which develops the softwares for the flights which will have the Wifi, movies and • • • • • • • • • other stuff. Especially we handle the WIFI service where users can buy the Wifi from the flights. Worked in the development of new api services as micro services using Flask, Docker and Swagger UI etc. Worked in postgres , redis and memcache. Using Pandas manipulated some of the documents. Work also includes the deployment and CI/CD pipe-lining through jenkins and deploying using kubernetes. Since it’s already developed project we develop new services with the existing repositories through the inheritance of those classes. Designed and configured Airflow DAGs (Directed Acyclic Graphs) to manage complex data pipelines, ensuring seamless integration and real-time synchronization of legacy insurance systems with the new cloud infrastructure Also worked in different VSMU’s as part of the debugging and also worked with production servers which will also have the access to ping the satellite. Apart from that, by following some of the strategies like peer programming and group meetings the work goes really fast and efficient. It always includes my inputs and plans to make this task or project to get to work. As a part of team member, contributed into development strategies by following codingstandards like writing test-cases immediately with the development, peer reviewing approvals to do the merge with master etc. Based on the sprint, we do debugging, solving the issues by tracking the environments like Dev, Staging and Pre-Prod through the Grafana. Scorg International Pvt Ltd May. 2021 – July .2021 Client : COGNIZANT (BMW - M4 Systems), Hyderabad, India Sr. Python Engineer (AWS/Python) • Understand stakeholder’s business problems and identify possible technical solutions toeffectively provide the project deliverables on time. • Use Python into lambda and Glue jobs to create a better and efficient data processing and data manipulation. • Creating a pipeline which included lambda, glue-job, redshift and finally into elastic search, Kibana for dashboards and analytics. These whole pipelines will be created through the terraform. • Creation of dashboards to visualize the analysis of the data using Kibana and Elasticsearch. • Used terraform by using python Inheritance and object oriented programming to automate the data processing in AWS. • Leverage industry best practices to design, test, implement and support a solution to migrate all the current codes to AWS using Lambda. • Managing each task day-to-day to Create python codes with different packages/modules using spark, pandas to do a variety of data analysis and create pipeline and deploy into production from end-to-end. • Facilitate collaboration with other developers, product owners, product teams and designers to solve interesting and challenging problems through delivery of various analytics solutions. • Provide knowledge transfer to other team members and be a part of an agile cross- functional team. UST Global Services (Health Care - Anthem), Trivandrum, India Jan. 2020 – March. 2021 Sr. Software Engineer (AWS/Python/Django/Pandas) • Working with Anthem (client), to manipulate huge health-care data and return the data as expected Using pandas and pyspark as well as Numpy. • Worked with the Kafka Streaming data will be taken and stored into AWS-SQL Databases. Worked to create a Glue Job with spark, where I process the dataframesafter cleansing, manipulating and transformations (as json). • Not only the Kafka but also worked with the structured and Un-Structured data and preparation of data into the required format using pandas and numpy etc. • Provide knowledge transfer to other team members and assisting them to understandthe project and terminologies that are being used. • Been a crucial collaborator to develop and handling some of the modules or featuresfrom end-to-end. • Working with the deliverables and improvements to understand and automate thecurrent existing • • • • environment. Assigning the day-to-day tasks and back-tracking with the team as well as collaboratinginto different issues facing by each-others to solve the blockers to move forward easily. Deployment and evaluating the solutions in staging and production environments. Implemented sharding and replication techniques to distribute data and ensure high availability. Used Django and MongoDB into ORM as well as used django-queryset for processing and prometheus for handling Time-series data, EY – Trivandrum, India June .2019 – Dec .2019 Client : LABGLO - (Nokia Tax Systems) Data Scientist & Python Developer (AWS/Python/Django) • Working with different emails to scrape the data using NLTK and preparation of data,into specified data format that is requested by the Tax Filing Team. • Developing a Django Application to process, visualize and make available the data tocollect and analyze for other teams. • Collecting and processing the data from various sources like SAP, Emails, Excel and processing these data based on the various kinds of data and other related information. • Deployment of application onto the Kubernetes and docker to deploy them into threestages like development, staging, production. • Automating the Ticket Creation SNOW based the data processed from the sources where this Snow Tickets allow teams for filing the taxes and addressing the issues ofdata related that must be solved. • Collaborating with the other teams to understand the requirements and automate themas per their requirements. • Analyzing the data and visualizing these data using Power BI and preparation ofdocuments based on these analytics. Labglo Technologies LLP -India (Ecommerce – Price.com), Nov. 2018 – Dec.2019 Data Engineer and Application Developer (Python/Scrapy/Django/Numpy) • Developed an application from end-to-end as a Single Collaborator and Crucial one too. • Worked with scrapy to scrap different E-commerce websites and preparation as well as manipulation of the data, then publishing into Price.com. • Scraping different websites using scrapy and processing with REGEX to process thescraped website data and arrange them with proper format of database. • Worked with Elastic-search services by indexing the data with postgres and to provide afeature like search and visualize the price differences with the other websites using Kibana. Also used Numpy to perform Dimensional Operations on Indexing data. • Regression Analysis on prices and predicted the prices to understand the prices on thewebsites for products by analytics. • Worked on managing and processing SQL DB queries through pyspark. • Used different SQL DB in using various SQL clauses (SELECT, FROM, WHERE, JOIN, GROUP BY, HAVING, ORDER BY) to perform data operations effectively. • Developed different web services that can provide data regarding unique features tomaintain and process the data including the scrapers. The Application was fully developed using DJANGO. • Deployment of these services into AWS environment and providing the access to publicenvironment. • Creating a CI/CD Pipelines using GIT and Docker. Also worked on day-to-dayImprovements of tasks and had a crucial part of releasing different versions. Deliverect -Belgium (Delivery – POS Systems) June .2018 – Sep .2018 Application Developer (Python/Flask) • Worked as a python developer for integrating Delivery platform data with POS Systems.working with the tools like Docker, celery etc. • Played as a prominent collaborator in the development and Integration of code. • Worked with the cloud environments like Google cloud engine integrated thestackdriver. • • • Celery was used to automate each service that executes automatically as per designedin docker. Swagger API was used to automate and maintain the API Services. Implementation of StackDriver into Docker and initiated as a service to monitor theapplication server. UST Global Services - (Chatbot Designer), Trivandrum Nov.2016 – June .2018 Machine Learning Engineer (Python) • Worked on development of chatbot designer (LIKE RASA), where we are creatingdifferent chatbots using trained NER (Natural Entity Recognition). • Used Regex to manipulate the utterances with Object Oriented Programming withpython. • Also worked on document processing project as a part of this chatbot using deeplearning. • Worked on training NER (Natural Entity Recognition) by implementing Flask, Gunicorn asan API Services. • Training the different utterances to NER and evaluating the application predictions livelythrough the Chatbot UI Chatbox. • Developed a scraper to get the live data like Climate and other Information from someof the sources and google based on the client requirement. • Manipulation and processing the data by some of the NLTK tools like Tokenization,Lemmatization etc. • Developed some of the features like story shifting while having a conversation with theBOT’s etc. FREELANCE PROJECTS: SURVEY SNEAK(FREELANCE) March.2019 – Oct .2019 Developed the application using Django. Modeling the Database based on the front-end design.Architectural Design for the whole Application as well as UI Mock-up Screens. Worked on API Services that are to be called onto each page as well as security using token system. CEBURU(FREELANCE) May.2018 – Dec. 2018 Developed an application to analyze the bandwidth data, network data. Continuous Monitoringof server data and other system related data through python API Services. Analyzed the data ofsystem data to monitor health of system on every hour through scheduled Services.
Get your freelancer profile up and running. View the step by step guide to set up a freelancer profile so you can land your dream job.