I am a results-driven Software Developer with expertise in Python, PySpark, SQL, and cloud-based data processing. With a strong foundation in data engineering, ETL pipeline development, and distributed computing, I am passionate about building scalable and efficient solutions for complex data challenges.
Currently, I work as a PySpark Developer at Tata Consultancy Services (TCS), where I design and optimize ETL pipelines to process and transform large datasets. My work involves integrating PySpark jobs with Azure Cloud Storage, ensuring seamless data ingestion, transformation, and retrieval. Additionally, I have developed Python-based data validation scripts to detect missing values, anomalies, and duplicates, improving data quality and accuracy.
Previously, I gained valuable industry experience as a Software Engineer at TCS, where I developed real-time and historical financial data extraction solutions using Apache Spark SQL. My role involved working closely with business teams to translate requirements into optimized data solutions, debugging production issues, and ensuring high data integrity for analytics and reporting.
During my internship at Intel, I worked on GPU software development, where I implemented new functionalities for Display Flip features and developed automated test cases using Python and JSON to validate GFX Display Driver performance. My contributions helped enhance system stability and compliance with Intel’s test scenarios.
In addition to my industry experience, I have a strong academic background with an M.Tech in Computer Science from NIT Goa, where my thesis focused on task allocation strategies in Fog Computing. I designed a framework to compare various optimization algorithms like Particle Swarm Optimization (PSO), Ant Colony Optimization (ACO), and Firefly Algorithm, demonstrating that PSO outperforms others in task efficiency and latency reduction.
Beyond my technical skills, I am certified in Microsoft Azure Fundamentals (AZ-900) and have a keen interest in generative AI, having completed Google’s Introduction to Generative AI course. My expertise spans data engineering, distributed computing, cloud storage integration, automation, and software development.
I am passionate about solving complex data problems, optimizing workflows, and contributing to cutting-edge innovations in big data, cloud computing, and AI-driven solutions. I am always eager to learn new technologies and collaborate with like-minded professionals to drive impactful projects.