Arun Mathew

Arun Mathew

$35/hr
Data engineer |Data Analyst | Python | Data Specialist | Azure DE | AWS DE | SnowFlake
Reply rate:
-
Availability:
Hourly ($/hour)
Age:
38 years old
Location:
Woodbridge, Ontario, Canada
Experience:
6 years
  Arun Mathew   Contact   Address Woodstock, ON N4T 0N6 Phone - E-mail -  Technical Profile   Programming and Scripting: Python (Pandas, PySpark), Java, powershell, SQL Data Management: Oracle, PostgreSQL, Snowflake, MongoDB BI tools: Tableau, Power BI ETL tools: Informatica, DBT, DataStage, AWS Glue Source Code Management: Git, GitLab, Bitbucket Azure Cloud tools: Azure SQL Database, Azure Data Lake, Azure Data Factory, Azure Databricks, Azure Synapse Analytics, Pyspark Orchestration and Container Management: Docker, Kubernetes, Jenkin Data Analyst Activities: Data Cleansing, Data Reconciliation, Data Mapping, Data Governance, Data profiling, Data Lineage, Data Modeling Other Tools and process: Jira, Confluence, Qtest, ALM, Agile/Scrum, Kanban, Jupyter Notebook Performance testing tools: Load runner, JMeter, Dynatrace, APPD, Splunk, Fiddler, Performance Center   Competencies   Technology leadership work streams Risk analysis Data analysis   Data engineer with 5 years of experience, with expertise in ETL pipelines, Data Cleansing, Data Reconciliation, Data Mapping, Data profiling, Data Lineage and data modeling. With a proven history of demonstrating data gathering, transformation and optimization, I am looking to leverage my technical and problem-solving skills to drive positive business outcomes for my next organization. Well versed with all stages of software development life cycle (SDLC) and software testing life cycle Professional experience in Data Analysis and Business Intelligence. Extensive experience in all phases of Software development which includes Gathering Requirements, Analysis, Design, Development, Data Analytics, Visualization and Data Warehousing Master’s degree in data science. Experience in continuous integration and continuous deployment methodologies using Git, Bitbucket, Bamboo and Jenkins. Knowledge of Banking and Financial, Insurance and HRIS Expertise in ETL Pipelines, PostgreSQL, Snowflake, Azure, DBT and Power BI Experience in Data Cleansing, Data Reconciliation, Data Mapping, Data Governance, Data profiling, Data Lineage, Data Modeling. Developed and maintained stored procedures, views, and functions in SQL server to optimize data extract, transform and load (ETL) process. Automated ETL data validation for daily batch processing in test environment. Building of data warehouse and Development of Reports in Power BI. Hands-on experience with Cloud services such as AWS and Azure. Excellent communication skills with the ability to work independently as well as in a team. AWS Certified Solutions Architect Associate AWS Certified Cloud Practitioner Oracle Database SQL certified Expert Work History     2019-01 - 2022-05   Data Engineer Tata Consultancy Services Optimized data processing by implementing efficient ETL pipelines and streamlining database design. Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability. Enhanced data quality by performing thorough cleaning, validation, and transformation tasks. Automated routine tasks using Python scripts, increasing team productivity and reducing manual errors. Develop CI/CD pipelines and testing automation. Utilized Azure Data Lake Storage to manage data storage. Set up data integration solutions using Azure Data Factory. Migrated large-scale data warehouse to Azure Synapse Analytics. Migrated legacy systems to modern big-data technologies, improving performance and scalability while minimizing business disruption. Fine-tuned query performance and optimized database structures for faster, more accurate data retrieval and reporting. Led end-to-end implementation of multiple high-impact projects from requirements gathering through deployment and post-launch support stages. Designed scalable and maintainable data models to support business intelligence initiatives and reporting needs. Streamlined complex workflows by breaking them down into manageable components for easier implementation and maintenance. Evaluated various tools, technologies, and best practices for potential adoption in the company’s data engineering processes. Established robust monitoring processes to detect system anomalies. Collaborated with cross-functional teams for seamless integration of data sources into the company’s data ecosystem. Conducted extensive troubleshooting to identify root causes of issues and implement effective resolutions in a timely manner. Provided technical guidance and mentorship to junior team members, fostering a collaborative learning environment within the organization. Increased efficiency of data-driven decision making by creating user-friendly dashboards that enable quick access to key metrics. Developed and delivered business information solutions. Designed data models for complex analysis needs. Developed database architectural strategies at modeling, design, and implementation stages to address business or industry requirements. Gathered, defined and refined requirements, led project design and oversaw implementation. Planned and installed upgrades of database management system software to enhance database performance. Reviewed project requests describing database user needs to estimate time and cost required to accomplish projects. Established and secured enterprise-wide data analytics structures. 2011-01 - 2019-01   ETL Tester/ETL Test Lead Tata Consultancy Services Design, develop, and execute test plans and test cases for ETL processes focused on data integration and data conversion Automate test scripts and processes using industry-standard tools and frameworks to validate ETL pipelines, ensuring they meet business and technical requirements. Performed comprehensive testing of data extraction, transformation, and loading (ETL) processes to ensure data accuracy, completeness, and integrity across different data sources. Collaborated closely with data engineers, developers, and business analysts to understand data requirements and translate them into effective test strategies. Validated data conversion processes during migrations and integrations, ensuring that data is accurately transformed and loaded into target systems. Worked in Data Cleansing, Data Reconciliation, Data Mapping, Data Governance, Data profiling, Data Lineage, Data Modeling. Identify, document, and track defects, working closely with the development team to ensure timely resolution. Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability Developed, implemented and maintained data validation protocols, standards, and documentation. Analyzed complex data and identified anomalies, trends, and risks to provide useful insights to improve internal controls. Generated detailed studies on potential third-party data handling solutions, verifying compliance with internal needs and stakeholder requirements. Standardized ETL mapping document and centralized mapping document repository. Reviewed project requests describing database and ETL changes to estimate time and cost required to accomplish projects. Produce detailed test reports, including defect summaries and recommendations for improving data quality and ETL processes. Automated Daily regression data validation thus reducing the effort and cost for regression test. Worked as part of Database migration to PostgreSQL from Oracle. Worked as part of migration from Informatica to python framework ETL. Improvised the ETL testing process to make it more efficient. 2022-06 - 2023-10   Performance Test Engineer High Line Software Corporation, Markham, ON . Conducted load, stress and endurance tests using JMeter scripts to simulate realistic user activities. Used monitoring tools Appd and Dynatrace tools on frequent basis Developed and deployed test load scripts with JMeter. Developed and deployed industrial automation systems, increasing efficiency and reducing costs. Reviewed scalability, performance and load balancing of each application. Coordinated and monitored work of co-located and remote teams. Mentoring and hands on training provided in an ongoing basis. Installed and maintained testing tools including LoadRunner, JMeter. Have been part of the following Performance Test planning activities: Performance Test plan preparation based on NFR, and SLA. Test plan review meeting. Design Workload model. Target TPS and response time calculation based on Production volume/throughput from the Client application. Baseline the existing application before new changes are made. Prepared Scripts in VUGEN Scripting tool for Load Runner in the following protocols: HTTP/HTML Web Services True client (Browser-based) Verifying whether the scripts are created with proper checkpoints/validation, Correlation, Parameterization, Think time, and Transaction name with the standard naming convention. Test data set up using automation/load runner scripts. Created dashboard for KPI’s in APPD and Dynatrace Coordinate with Dynatrace team to Instrument the Server with Dynatrace agents. Execute Load test, Stress Test/Scalability test, Endurance test. Test analysis and performance test report sharing. Raising the Performance defects and following up with concerned teams. Call out the risks and variance in response time/resource utilization compared to baseline/SLA. Conduct Test Result review meeting with Stakeholders. Analyzing the Load test results and sharing the result with the concerned team. Providing server analysis for test duration from Dynatrace and Splunk Tools. Execute Ad hoc performance testing request. Analyzing Logs for Exceptions and Error. Work with Development and middleware team to tune the resource utilization like Memory, CPU, and GC for new applications without impacting response time. Preparing sign-off Document with all the test results, Defects, and risks identified during the test. Created live monitoring for JMeter test using Grafana and influx dB Analyze Performance of application in production. Compare the QA region performance results with Production performance results. Migration of Load runner scripts to JMeter Scripts. Education     2020-11 - 2023-05   Master of Technology: Data Science Birla Institute of Technology and Science   2005-05 - 2009-05   Bachelor of Technology: Electronics and Communication Government Engineering College, Trivandrum   Certifications     2023-12   Certified AWS Solutions Architect   2023-12 2015-10 2012- 08   Certified AWS cloud Practitioner ISTQP certified Oracle Database SQL Certified Expert .
Get your freelancer profile up and running. View the step by step guide to set up a freelancer profile so you can land your dream job.