Ritwick Kumar

Ritwick Kumar

$100/hr
Hands on Bigdata Developer and Architect
Reply rate:
50.0%
Availability:
Hourly ($/hour)
Age:
45 years old
Location:
Gurgaon, Haryana, India
Experience:
14 years
RITWICK KUMAR Bigdata Technical Architect/Bigdata Solution Development and Implementation: Offering 13+ years’ experience Hands-on experience in architecting and implementing bigdata solutions for data analytics and business intelligence Domain Expertise entails Retail, Banking and Healthcare - http://bigdatamakingscience.com - Profile Summary 14years of IT industry experience including 6years of experience in Bigdata. Currently working as Technical Lead with Boston Scientific Services Pvt Ltd and responsible for architecting & implementation of Bigdata projects. Built motivated resource teams by cultivating a culture valuing excellence, respect and fun Architected and Implemented complex Bigdata solutions for Retail client (including Recommendation Engine), Banking client (central data lake, data and application security, customer management and risk management) Member of core gate review team responsible for architectural reviews of solution blueprinting of bigdata projects in the organization and defining the Best Practices and Standards for the Bigdata solutions. Gained international exposure through 2 years of onsite assignments and completing specialized training/Certification from LOMA for Insurance domain. Experience in managing large project estimation, costing, schedule preparation & tracking, resource planning, execution to closure. Good Experience in collaborating with the third-party vendors Successfully conducted various customer workshops with business clients for requirements gathering. Translated Business requirements in to Technical specifications Ability to handle multiple tasks and tight deadlines. Conceptual thinking and adapt to new technologies quickly. Good Management Skills, communication and Presentation skills. A team player - proven ability to lead a medium size team Experience with Cloudera Distribution for Hadoop as well as the Hortonworks Distribution Platform. Technical skills entail Bigdata platforms (CDH, HDP), UNIX shell scripting, Programming Languages (Scala, Mapreduce, Hive, Pig, Impala, Java, Python, R), Source Control Tools (PVCS, Visual Source Safe, SVN) Core Competencies Bigdata Platforms: Cloudera and Hortonworks Programming Languages: Scala, Hive, R Mapreduce, Pig, Impala, Java, Python Data Ingestion: Flume, Sqoop, WebHDFS, HttpFS, Kafka, Streamsets Data Transformation: Trifacta, Streamsets, Mapreduce, Hive User Experience: Hadoop User Experience (HUE) File Formats: Avro, Parquet, ORC, Text, Sequence Workflow Management: Oozie Databases: NoSQL (HBase, Cassandra), RDBMS (Netezza, SQL Server, Oracle) Security: Linux Access Control Lists, Apache Sentry, Kerberos, Firewalls Notable Accomplishments across the Career Domain Expertise: Retail, Banking, Insurance and Healthcare Solution Architecting: Owned sourcing, transformation and archival of data in Enterprise Data Hub as part of Enterprise Data Management efforts on multiple projects, ensured compliance to data policies, and established data lineages. Process Enhancement Initiatives: Implemented various value improvement projects and processes for automatic the manual activities in hadoop platform maintenance and temp directory cleaning, data ingestion framework for delta load and generating the control information for each job (ingestion as well as transformation on the cluster) People Leadership Skills: Imparted trainings and performed knowledge enhancement sessions for improving the productivity. Navigated seamlessly through the high attrition crises period, resulting in being asked to perform multiple roles like team lead, developer, support, stakeholder management, etc. Skill Set Academic Details Master of Engineering in Computer Science and Engineering from Institute of Technology, Agra University (University Topper) Bachelor of Engineering in Computer Science and Engineering from Maharishi Dayanand University, Rohtak (Second Rank in the University) Change Agent Certifications Recommendation Systems (MOOC, Coursera) LOMA 290- Introduction to Insurance, Underwriting (LOMA, USA) LOMA 280- Principles of Insurance (LOMA, USA) Organizational Experience Personal Details Annexure Organization: Boston Scientific India Period: May’ 16 – Present Project: Stormer Position: Technical Architect Environment: Cloudera Distribution for Hadoop 5.4.7, StreamSets, Thingsworx, Hive, Pig, Sqoop, Mapreduce, Unix shell scripting, Microsoft Active Directory, Kerberos, Cloudera Search, Spark, Scala, Python, Impala Description: Stormer’s objective is to implement an Internet of Things (IoT) solution, Thingworx, which enables the collection and exchange of data to Neuromodulation’s Clinician Programmer (CP). The solution allows Neuromodulation to collect CP data in order to drive analytics, and increase frequency of data movement. Before this implementation the data ingestion used to happen once a quarter with no provision of in time analytics leaving the leadership without the useful insight for fact based decision making. This implementation also allows for data ingestion from Enterprise Integrations in raw format (oil) and in a user-specified format (gasoline, diesel). The Enterprise Data Hub can now support data analytics through raw ingest of data and transformation into a new schema. Key Responsibilities  Interact with the Business partners and work with them to understand the requirements  Architect end to end solution blueprints starting from data mapping, ingestion to feeding downstream systems.  Implement and deliver the solutions to the business partners (giving them support and guidance during UAT)  Manage all four hadoop clusters and handle the upgrades  Suggest and implement value improvement ideas in the current process  Incorporate the approved client application standards into the hadoop cluster implementation  Performance Benchmarking for various data ingestion and processing use cases  Recruit and build a team of Bigdata experts for the organization  Liaise with internal technical teams for integration development and testing  Manage offshore development and support teams  Business stakeholder management  Scope definition and management  Weekly reporting to the India Head and the business  Risk & Issues management  Resolving business and system issues  People and Team Management  Risk analysis and mitigation Client: Royal Bank of Scotland, UK Organization: Accenture Services Pvt Ltd Period: Oct’ 14– May’ 16 Project: Bigdata Implementation Position: Technical Architect / Offshore Lead Environment: Cloudera Distribution for Hadoop 5.4.3, Trifacta, Hive, Pig, Sqoop, Mapreduce, Unix shell scripting, Microsoft Active Directory, Kerberos Description: Client is looking to harness the power of the enormous amount of data they have and a few important use cases include Customer Management (Use the predictive knowledge to track the customer, create customer segments and to increase customer retention and customer loyalty), Ris k Management (Identify the risk of a portfolio or even a customer to drive higher loan volumes and maintain a target risk for the portfolio) and Call Center Analysis (Applying text and sentiment analysis on user complaints with an emphasis to reduce customer churn, and drive up-sell, cross-sell, customer monitoring alerts and fraud detection). Key Responsibilities:  Architect the security envelop for the Hadoop Clusters  Define the process of metadata management with internal SIEM tool  Define and Implement the data ingestion from Informatica Power Exchange to the Kerberized hadoop cluster  Implement the Bigdata hadoop environment for 2000+ internal business users  Incorporate the approved client application standards into the hadoop cluster implementation  Recruit and build a team of Bigdata experts for the organization  Liaise with internal technical teams for integration development and testing  Manage offshore development and support teams  Business stakeholder management        Scope definition and management Weekly reporting to the business Risk & Issues management Resolving business and system issues People and Team Management Risk analysis and mitigation Project Tracking Client: Coupons.com, USA Organization: Accenture Services Pvt Ltd Period: Nov’ 11– Oct’ 14 Project: ReceiptIQ Position: Technical Lead Environment: Apache Solr, Apache Hadoop, Cassandra, Hive, Sqoop, Netezza, MicroStrategy, Java, Unix shell scripting Description: A big data solution where retail customers are provided with real time offers based on their purchase behavior, basket items and offer scores. A multitenant system storing data from large retailers on No SQL database and generating recommendations from a customized in house recommendation engine (Accenture Recommendation Engine - ARE). The solution includes integration with analytical engine to produce wide variety of reports and dashboards depicting the multi dimensional image of customers, products, offer and coupons. Responsibilities:  Define and implement the data ingestion strategy  Customize mapreduce algorithms  Plan and support the client deployments  Manage offshore development team (15+ members)  Recruit and build a team of Bigdata experts  Conduct monthly Steering committee meetings with the key business stakeholders  Business stakeholder management  Scope definition and management  Weekly reporting to the business  Risk & Issues management  Resolving business and system issues  People and Team Management  Risk analysis and mitigation  Project Tracking Client: Target, USA Organization: Accenture Services Pvt Ltd Period: Jan’ 11– Nov’ 11 Project: Accenture Customer Insight Position: Technical Lead Environment: Apache Solr, Apache Hadoop, Hive, Sqoop, Netezza, MicroStrategy, Java, Unix shell scripting. Description: Accenture Customer Insight (ACI) product in The AI Retail Management (R&D) products portfolio is a high performance analytical engine that puts relevant information about customer behaviors in the hands of non-technical users. It is uniquely designed to run against large data volumes and draw reports on transactional, point-of-sale, third-party and customer loyalty data. ACI supports a large number of users from inside and outside of the organization and allows them to seamlessly run hundreds of reports each day. Responsibilities:  Define and implement the data ingestion strategy  Work with clients on requirement analysis.  Customize java based web UI for customer requirements.  Design and customize reports and dashboards for various lines of business  Plan and support the client deployments  Manage offshore development team (7 members)  Business stakeholder management  Scope definition and management  Weekly reporting to the business  Risk & Issues management  Resolving business and system issues  People and Team Management  Risk analysis and mitigation Client: Northwestern Mutual, USA Organization: Infosys Technologies Pvt Ltd Period: Dec’ 09– Jan’ 11 Project: WebSphere Version 7 Application Migration Position: Team Lead Environment: Java 1.4/1.6, WAS 6.0/7.0, MQ 6.0/7.0 , RSA 7.5 , IBM UDB/DB2 9.1, Sybase v 15, IBM MQ Workflow Description: Over 250 corporate applications are now on WebSphere Application Server, including many of the top tier applications. WebSphere application server versions have finite lifecycles. Migration of large suite of applications to newer versions requires early planning, quicker development and deployment. Migrations now take two or more years to accomplish. In addition, WAS v6 infrastructure is nearing capacity, causing to either add to an older platform or devote our attention to the newest one. This project involves migrating these enterprise applications to WAS 7 environment. Applications are compiled to Java 1.6; underlying frameworks are upgraded to compatible versions and set up on a brand new infrastructure and integration points are tested. Responsibilities:  Define Roadmap/approach for migration  Understand AS-IS architecture/Identify changes and their impact  Prepare migration cookbooks  Prepare and update Traceability matrix  Help other teams to upgrade/resolve technical issues  Lead a team of 1 onsite developer and 12 offshore developers Client: MetLife Insurance, USA Organization: Infosys Technologies Pvt Ltd Period: Mar’ 09– Dec’ 09 Project: Common Validator Framework Position: Team Lead Environment: Java 1.4, WAS 6.0, RSA 7.0, IBM UDB/DB2 9.1. Description: Common Validator Framework is a replacement for the industry standard iLog rule validator. Renewing iLog license every year was not a feasible option due to budget constraints. CVF was then built as in house replacement for iLog. Rules for application validation were formulated into a database up-loadable format and the core rule validator engine was built to process those rules for different applications types in New Business area. New insurance or insurance change applications are scanned for the input and corresponding rules are run by the rule engine to validate the input . Responsibilities:  Requirement workshops with the customer.  Scope definition and management.  Weekly reporting to the clients  Calls with clients for project status and issues  Traceability matrix  Risk & Issues management  Resolving business and system issues.  People and Team Management  Risk analysis and mitigation Client: Northwestern Mutual, USA Organization: Infosys Technologies Pvt Ltd Period: Feb’ 08– Feb’ 09 Project: WebSphere Version 7 Application Migration Position: Program Analyst Environment: Java 1.4, WAS 6.0, MQ 6.0, RAD 6.0 , IBM UDB/DB2 9.1, Sybase v 15, IBM MQ Workflow. Description: One of the big initiatives as Northwestern Mutual was to introduce Long Term Care insurance to customers and sell directly through own Financial Representatives. Earlier the business line was managed through other companies. As part of this initiative business processes reengineering was done for Long Term Care domain. It involved analysis of data flow interactions with upstream and downstream systems built with heterogeneous technologies on multiple databases. A core team was set up to define a roadmap to redesign the system in mainstream technologies and also identify the business improvement opportunities. There were multiple tracks to identify the current state business requirements, define future state requirements, solution architecture for the new system.. Responsibilities:  Conducting requirement workshops with the customer and develop functional and Technical specifications  Define re-engineering process for the business rules in legacy systems  Developing system requirements, system blue print and high level design.  Requirement analysis and development effort estimation.  Traceability matrix  Reviewing Detail Design documents, code and unit test results.  Resolving system issues.  Risk analysis and mitigation Client: Northwestern Mutual, USA Organization: Infosys Technologies Pvt Ltd Period: Nov’ 05– Jan’ 08 Project: Downtown E-Files Position: Program Analyst Environment: Java 1.4.2, JEE 1.4, EJB 2.0, Struts 1.2, MQ 6.0, MQ Workflow 3.4, IBM CM 8.1, AIX, HP Unix, MQ Broker 6.0, IBM UDB 8.0, Sybase 11.0, Visual Basic 6.0, PL I, COBOL, z/OS, Pervasive SQL Btrieve, RAD 6.0. Description: The E-Files solution was defined to eliminate paper-based processing in New Business, DI & TC-TUP (in addition to the existing Franklin solution). NM signed an Enterprise Software and Services option (ESSO) with IBM in 2004 that enables unrestricted use of several IBM products for the next three years. The ESSO includes IBM Content Manager and NM has utilized this opportunity to re-align its ECM strategy to the org-wide objective of “Faster, Better, and Cheaper”. NM has made a strategic decision to move from Documentum to IBM Content Manager (ICM) as its Enterprise Content Management (ECM) system. This resulted in IBM CM Migration project. Responsibilities:  Requirement capturing and requirement analysis  Developing system requirements, system blue print and high level design.  Reviewing Detail Design documents, code and unit test results.  Creating traceability matrix between requirements, detail design, code changes and Unit test cases.  Resolving business and system issues.  Performance testing of the components using Wily Interscope. Client: Riverdeep Interactive Learning Inc., USA Organization: NIIT Limited Period: May’ 05– Oct’ 05 Project: Learning Village Position: Software Engineer Environment: Java 1.4.2, Struts 1.1, IBM UDB 6.0, WSAD 5.0 Description: Riverdeep Learning Village enables district administrators to develop, align and distribute standards-based curriculum to all schools and all levels of teachers. It is a flexible framework which can be adapted to a variety of instructional methodologies and to all areas within a district’s curriculum. With Riverdeep Learning Village, a district can align standards to their curriculum, making it accessible to all stakeholders within a school district, including administrators, teachers, students and parents. Riverdeep Learning Village has been designed to adapt to how a district functions rather than forcing a district to conform to a predefined construct. School districts find immediate benefit using Riverdeep Learning Village as it enables personalization to meet a district’s needs and align curriculum content to ever-changing standards. Riverdeep Learning Village was designed for ease of use for district and local administrators to manage curriculum services from their own desktop. Responsibilities:  Requirement analysis  Coding, Unit testing, Unit Test Planning.  Code review and unit test result review.  Creating traceability matrix between requirements, detail design, code changes and Unit test cases.  Coordinating with the offsite team (based in India) and assigning the task to the team Client: Reliance Infocomm limited Organization: HTS India Period: Jun’ 04– Apr’ 05 Project: Reliance Grievance Tracking System Position: Jr. J2EE Architect Environment: Java 1.4.2, Struts 1.1, IBM UDB 6.0, WSAD 5.0, Weblogic 8.1 Description: Objective of this project is to allow the client’s customer care to provide with the capabilities like receiving complaints from customers, task assignment and workflow reporting for the problems assigned, resolved, pending. The Grievance Tracking System helped customer care department in improving their work and process efficiency. Responsibilities:  To study the business requirements and identify the changes needed in the system.  Coding, Unit testing, functional testing and integration testing for all the components.
Get your freelancer profile up and running. View the step by step guide to set up a freelancer profile so you can land your dream job.