DATA ENGINEER
G Venkatachala
Email:-Phone: -
Professional Summary
Experienced IT professional with 5.6 years in the industry, including over 3 years specializing in Azure cloud technologies, data engineering, and analysis. Proficient in Azure Data Factory, Azure Databricks, PySpark, SQL, and Snowflake. Demonstrated ability in ETL processes, development, maintenance, testing, and documentation, with a strong background in optimizing and automating data workflows for seamless cloud integration and high data quality.
Technical Skills
Cloud Platforms: Azure Data Factory, Azure Databricks
Programming Languages: PySpark, Python, SQL, PL/SQL
Databases: Snowflake, Oracle 10g/11g
Tools: Toad, SVN, SQL Impact, SQL*Loader, PL/SQL Developer
Operating Systems: Windows Family
Professional Experience
Acro Business Solutions Pvt Ltd, Bangalore
Data Engineer
January 2019 - August 2023
Developed ADF pipelines incorporating control flow activities like Get Metadata, If Condition, ForEach, Delete, and Validation.
Created and scheduled pipelines for executing Databricks Notebook activities.
Debugged and fixed issues in data pipelines, using Event Trigger and Schedule Trigger for automation.
Developed Spark applications using Spark SQL for data extraction, transformation, and aggregation.
Configured and scheduled Databricks jobs for automated and periodic data processing.
Optimized PySpark jobs using broadcast variables and caching mechanisms.
Integrated Databricks with Azure services and data lakes for big data analytics.
Writer Corporation, Hyderabad
Senior Executive (Systems & Operations)
June 2017 - May 2018
Developed stored procedures, functions, packages, and triggers in SQL.
Utilized PL/SQL concepts like cursors, exceptions, and dynamic SQL for backend coding.
Debugged and resolved issues in SQL programs, ensuring efficient and error-free code.
Conducted test cases, code reviews, unit testing, impact analysis, and documentation.
Projects
Fund Accounting (FA)
Organization:
Duration: April 2021 - August 2023
Technology: SQL, Forms and Reports 10g/11g, Azure Databricks, Azure Data Factory (ADF)
Platform: Windows XP/7
Role: Azure Data Engineer
Connected on-premises data sources using Azure Data Factory Copy Activity and Self-Hosted Integration runtime, copying data into Azure Data Lake Store Gen2.
Created and scheduled multiple Databricks PySpark notebooks for validating, processing, and reading external JSON data.
Debugged data pipelines and scheduled job activities in Databricks and Azure ADF.
Developed and modified Oracle forms and reports per functional requirements.
Used PL/SQL programming for developing stored procedures and database triggers.
Prepared technical documents and reviewed developed code.
Propstay
Duration: January 2019 - March 2021
Technology: SQL, PL/SQL, XML Publisher
Platform: Windows XP/7
Project Type: Development/Enhancement
Designed and modeled relational databases for efficient structure and performance.
Defined and created database objects like tables, views, indexes, and constraints.
Wrote and optimized SQL queries for efficient data retrieval and manipulation.
Supported client issues, attended weekly status meetings, and provided solutions to recurring issues.
Developed and modified existing XML publisher reports per client requirements.
Certifications
Databricks Certification
Additional Skills
Debugging and Troubleshooting: Skilled in identifying and resolving data pipeline and SQL program issues.
Optimization: Enhanced performance of data processing jobs through code optimization and efficient resource utilization.
Documentation and Testing: Proficient in creating comprehensive documentation and conducting thorough testing to ensure software quality and maintainability.
Soft Skills
Strong problem-solving and analytical abilities.
Effective communication and collaboration with team members and stakeholders.
Capable of managing multiple tasks and projects simultaneously, ensuring timely delivery and high-quality outcomes.