ANCY VARGHESE
Data Engineer--www.linkedin.com/in/ancy-varghese-a-
Results-oriented and highly skilled Data Engineer with 10+ years of extensive experience in designing, developing, and maintaining scalable ETL processes by following Agile Methodologies. Hands on experience in Data Warehousing and Business process automation projects. Expertise in ETL tools such as Talend Enterprise Edition for Big Data Integration, Oracle Data Integrator (ODI), Pentaho Kettle and cloud-based platforms like Amazon Web Services, Azure DevOps, and Data warehouse technologies. Adept at leveraging advanced technical acumen and strong analytical skills to solve complex data challenges.
WORK EXPERIENCE
Tata Consultancy Services - Data Engineer
Feb 2016 – Jan 2024
• Designed, developed and maintained scalable ETL jobs using Talend Enterprise Edition for Big Data Integration (Integrating with AWS S3, Hive, PostgreSQL, Snowflake, MySQL, Netezza, AS400, Flat files, and Rest APIs), Oracle Data Integrator(11g/12c) and Pentaho Kettle to handle 2 TBs of daily data from diverse data sources
• Developed and maintained CI/CD pipelines using Azure DevOps and Jenkins
• Led the migration from Data Stage to Talend Open Studio, resulting in annual cost savings of $278,000 and an increase in performance of 28%
• Created and scheduled a single ETL job to re-run failed jobs because of server failure, which reduced manual workload by 64% daily
• Conducted performance tuning for ETL processes resulting in 20% improvement in data processing speed and efficiency
• Created comprehensive documentation and implemented best practices for data engineering workflows, ETL processes, and data security, improving team productivity by 15%
• Studied Python in generating data extracts to integrate with existing ETL processes which can be transformed and used in Data Analytics
• Learned Tableau and built a Tableau Dashboard to visualize customer demographics and purchasing patterns, improving data accessibility for stakeholders
• Worked with data scientists, data analysts, and business stakeholders to understand data requirements and deliver data solutions that support business intelligence
• Onboarded new resources for the team and provided necessary training
Wipro Technologies - ETL Developer
Aug 2010 – Jan 2016
• Designed and developed over 200+ ETL mappings/interfaces and load plans, handling data volumes exceeding 1TB daily across various data sources (Oracle, SQL Server, and flat files)
• Led a data migration project that moved over 10 million records to data warehouses, achieving a 99.9% data accuracy rate
• Reduced data integration failures by 50% through the implementation of automated error handling.
• Reduced ETL processing time by 30% through performance tuning and optimization of workflows and queries
PROJECTS
DOVE - Inventory DevOps - Senior Data Engineer Jan 2024
• Maintaining and adding new features to Petronas Trading Corporation (PETCO) dashboard as per the Business requirement
• Analise the requirements to prioritize Sprint tasks and create user stories/tasks in Azure DevOps for each Sprint
• Develop Talend jobs by extracting data from AWS S3 buckets, transforming, and loading into PostgreSQL and Hive (by following the best practices such as loading Master and Local contexts (in MySQL), usage of Joblets and Routines, maintaining separate DQ, EL, and Transformation jobs
• Create and manage Execution plans in Talend Administration Center (TAC) for automation
• Perform Unit Testing, User Acceptance Testing and Support QA team for their testing activities
• Designed and developed ETL jobs and transformations using Pentaho Kettle to load data from multiple sources into a centralized data warehouse.
• Optimized ETL performance by refactoring slow transformations and managing memory utilization.
Talend Cloud Migration - Data Engineer Dec 2022
• Collaborate as an expert in planning, designing, development, and deployment of Cloud jobs
• Re-design the existing On-premises ETL jobs to cloud by pushing data into AWS S3 buckets and Snowflake
• Deploy the ETL jobs in Talend Management console (TMC) then to higher environments using Jenkins and automate the Execution plans
• Perform Unit Testing, User Acceptance Testing and Support QA team for their testing activities
• Conducted over 200 code reviews providing constructive feedback that improve code quality, readability, and maintainability
GDPR - Customer Search - Data Engineer Sep 2021
• Facilitate meetings with business subject matter experts, decision makers, systems owners, and end users to define goals and requirements for ETL design
• Analise the existing data and develop Talend jobs to perform Fuzzy, Exact, Wildcard, Multiformat and combination searches for a given attribute in AS400, Oracle, Snowflake, DB2, Netezza, and Rest API
• Deploy the ETL jobs to Talend Administration Center (TAC) and automate the Execution plans
• Perform Unit Testing, User Acceptance Testing and Support QA team for their testing activities
Talend Migration - Data Engineer Apr 2020
• Analize and redesign the existing 10K Data stage jobs using Talend Open Studio, Oracle, Netezza, AS400 and other data warehouse tools.
• Deploy the ETL jobs to Talend Administration Center (TAC) and automate the Execution plans
• Perform Unit Testing, User Acceptance Testing and Support QA team for their testing activities
Koperasi Angkatan Tentera Malaysia Berhad (KATMB) - Data Engineer Aug 2018
• Collaborate as an expert in the planning, design, development, and deployment. Provide architectural guidance, technology, and process improvement recommendations to the business
• Create Global(reusable) mappings, variables, procedures, packages. Design and develop the load plans to automate the data flow
• Provide effortless communication between the customer and the offshore team to complete Phase I and Phase II Go-Live
• Receive and submit sign off from every business unit as part of business requirement
• Conduct technical sessions related to ETL execution and Production monitoring to the Malaysian data team
Mazda Marketing Implementation - ETL Developer Jan 2016
• Participate in the Team meetings to gather business requirements and develop the specification documents
• Develop and customize ODI interfaces to fetch the data from Siebel, Mainframe systems to DWH.
• Create and maintain shell scripts to trigger the ETL jobs
• Perform Unit Testing, User Acceptance Testing and Support QA team for their testing activities
Education
• M.Tech in Software Engineering – Birla Institute of Technology and Science, INDIA
• B.Sc Computer Science – Prajyoti Niketan College, INDIA
Technical SKILLS
• Talend Enterprise Edition for Big Data Integration
• Oracle Data Integrator 11g/12C
• Amazon Web Services - S3
• Azure DevOps
• Oracle, SQL Server, AS400, Netezza, MySQL, Hive, PostgreSQL and Snowflake
• Basic Tableau
• Data Warehousing
• Shell Scripting and Basic Python
Achivements
• Received PETRONAS Customer appreciation for end-to-end development of re-run facility in TAC for every PETCO job in case of server failure
• Recognized as TCS Contextual Masters for GDPR Customer Search
• Received Technical Excellence Award for extraordinary contribution while working with Western Union
• Received Koperasi Tentera Customer appreciation for contribution to KATMB production release
• Received Outstanding performer of the year Award for the contributions while working with Mazda