RITWICK KUMAR
Bigdata Technical Architect/Bigdata Solution Development and
Implementation: Offering 13+ years’ experience
Hands-on experience in architecting and implementing bigdata solutions for
data analytics and business intelligence
Domain Expertise entails Retail, Banking and Healthcare
-
http://bigdatamakingscience.com
-
Profile Summary
14years of IT industry experience including 6years of experience in Bigdata. Currently working as Technical Lead
with Boston Scientific Services Pvt Ltd and responsible for architecting & implementation of Bigdata projects.
Built motivated resource teams by cultivating a culture valuing excellence, respect and fun
Architected and Implemented complex Bigdata solutions for Retail client (including Recommendation Engine),
Banking client (central data lake, data and application security, customer management and risk management)
Member of core gate review team responsible for architectural reviews of solution blueprinting of bigdata projects
in the organization and defining the Best Practices and Standards for the Bigdata solutions.
Gained international exposure through 2 years of onsite assignments and completing specialized
training/Certification from LOMA for Insurance domain.
Experience in managing large project estimation, costing, schedule preparation & tracking, resource planning,
execution to closure. Good Experience in collaborating with the third-party vendors
Successfully conducted various customer workshops with business clients for requirements gathering. Translated
Business requirements in to Technical specifications
Ability to handle multiple tasks and tight deadlines. Conceptual thinking and adapt to new technologies quickly.
Good Management Skills, communication and Presentation skills. A team player - proven ability to lead a medium
size team
Experience with Cloudera Distribution for Hadoop as well as the Hortonworks Distribution Platform.
Technical skills entail Bigdata platforms (CDH, HDP), UNIX shell scripting, Programming Languages (Scala,
Mapreduce, Hive, Pig, Impala, Java, Python, R), Source Control Tools (PVCS, Visual Source Safe, SVN)
Core Competencies
Bigdata Platforms: Cloudera and
Hortonworks
Programming Languages: Scala, Hive, R
Mapreduce, Pig, Impala, Java, Python
Data Ingestion: Flume, Sqoop, WebHDFS,
HttpFS, Kafka, Streamsets
Data Transformation: Trifacta, Streamsets,
Mapreduce, Hive
User Experience: Hadoop User Experience
(HUE)
File Formats: Avro, Parquet, ORC, Text,
Sequence
Workflow Management: Oozie
Databases: NoSQL (HBase, Cassandra),
RDBMS (Netezza, SQL Server, Oracle)
Security: Linux Access Control Lists, Apache
Sentry, Kerberos, Firewalls
Notable Accomplishments across the Career
Domain Expertise: Retail, Banking, Insurance and Healthcare
Solution Architecting: Owned sourcing, transformation and
archival of data in Enterprise Data Hub as part of Enterprise Data
Management efforts on multiple projects, ensured compliance to
data policies, and established data lineages.
Process Enhancement Initiatives: Implemented various value
improvement projects and processes for automatic the manual
activities in hadoop platform maintenance and temp directory
cleaning, data ingestion framework for delta load and generating the
control information for each job (ingestion as well as transformation
on the cluster)
People Leadership Skills: Imparted trainings and performed
knowledge enhancement sessions for improving the productivity.
Navigated seamlessly through the high attrition crises period,
resulting in being asked to perform multiple roles like team lead,
developer, support, stakeholder management, etc.
Skill Set
Academic Details
Master of Engineering in Computer Science and Engineering from
Institute of Technology, Agra University (University Topper)
Bachelor of Engineering in Computer Science and Engineering from
Maharishi Dayanand University, Rohtak (Second Rank in the University)
Change Agent
Certifications
Recommendation Systems (MOOC, Coursera)
LOMA 290- Introduction to Insurance, Underwriting (LOMA, USA)
LOMA 280- Principles of Insurance (LOMA, USA)
Organizational Experience
Personal Details
Annexure
Organization: Boston Scientific India
Period: May’ 16 – Present
Project: Stormer
Position: Technical Architect
Environment: Cloudera Distribution for Hadoop 5.4.7, StreamSets, Thingsworx, Hive, Pig, Sqoop,
Mapreduce, Unix shell scripting, Microsoft Active Directory, Kerberos, Cloudera Search, Spark, Scala,
Python, Impala
Description: Stormer’s objective is to implement an Internet of Things (IoT) solution, Thingworx,
which enables the collection and exchange of data to Neuromodulation’s Clinician Programmer (CP).
The solution allows Neuromodulation to collect CP data in order to drive analytics, and increase
frequency of data movement. Before this implementation the data ingestion used to happen once a
quarter with no provision of in time analytics leaving the leadership without the useful insight for fact
based decision making. This implementation also allows for data ingestion from Enterprise
Integrations in raw format (oil) and in a user-specified format (gasoline, diesel). The Enterprise Data
Hub can now support data analytics through raw ingest of data and transformation into a new
schema.
Key Responsibilities
Interact with the Business partners and work with them to understand the requirements
Architect end to end solution blueprints starting from data mapping, ingestion to feeding
downstream systems.
Implement and deliver the solutions to the business partners (giving them support and guidance
during UAT)
Manage all four hadoop clusters and handle the upgrades
Suggest and implement value improvement ideas in the current process
Incorporate the approved client application standards into the hadoop cluster implementation
Performance Benchmarking for various data ingestion and processing use cases
Recruit and build a team of Bigdata experts for the organization
Liaise with internal technical teams for integration development and testing
Manage offshore development and support teams
Business stakeholder management
Scope definition and management
Weekly reporting to the India Head and the business
Risk & Issues management
Resolving business and system issues
People and Team Management
Risk analysis and mitigation
Client: Royal Bank of Scotland, UK
Organization: Accenture Services Pvt Ltd
Period: Oct’ 14– May’ 16
Project: Bigdata Implementation
Position: Technical Architect / Offshore Lead
Environment: Cloudera Distribution for Hadoop 5.4.3, Trifacta, Hive, Pig, Sqoop, Mapreduce, Unix
shell scripting, Microsoft Active Directory, Kerberos
Description: Client is looking to harness the power of the enormous amount of data they have and a
few important use cases include Customer Management (Use the predictive knowledge to track the
customer, create customer segments and to increase customer retention and customer loyalty), Ris k
Management (Identify the risk of a portfolio or even a customer to drive higher loan volumes and
maintain a target risk for the portfolio) and Call Center Analysis (Applying text and sentiment analysis
on user complaints with an emphasis to reduce customer churn, and drive up-sell, cross-sell,
customer monitoring alerts and fraud detection).
Key Responsibilities:
Architect the security envelop for the Hadoop Clusters
Define the process of metadata management with internal SIEM tool
Define and Implement the data ingestion from Informatica Power Exchange to the Kerberized
hadoop cluster
Implement the Bigdata hadoop environment for 2000+ internal business users
Incorporate the approved client application standards into the hadoop cluster implementation
Recruit and build a team of Bigdata experts for the organization
Liaise with internal technical teams for integration development and testing
Manage offshore development and support teams
Business stakeholder management
Scope definition and management
Weekly reporting to the business
Risk & Issues management
Resolving business and system issues
People and Team Management
Risk analysis and mitigation
Project Tracking
Client: Coupons.com, USA
Organization: Accenture Services Pvt Ltd
Period: Nov’ 11– Oct’ 14
Project: ReceiptIQ
Position: Technical Lead
Environment: Apache Solr, Apache Hadoop, Cassandra, Hive, Sqoop, Netezza, MicroStrategy, Java,
Unix shell scripting
Description: A big data solution where retail customers are provided with real time offers based on
their purchase behavior, basket items and offer scores. A multitenant system storing data from large
retailers on No SQL database and generating recommendations from a customized in house
recommendation engine (Accenture Recommendation Engine - ARE). The solution includes integration
with analytical engine to produce wide variety of reports and dashboards depicting the multi dimensional image of customers, products, offer and coupons.
Responsibilities:
Define and implement the data ingestion strategy
Customize mapreduce algorithms
Plan and support the client deployments
Manage offshore development team (15+ members)
Recruit and build a team of Bigdata experts
Conduct monthly Steering committee meetings with the key business stakeholders
Business stakeholder management
Scope definition and management
Weekly reporting to the business
Risk & Issues management
Resolving business and system issues
People and Team Management
Risk analysis and mitigation
Project Tracking
Client: Target, USA
Organization: Accenture Services Pvt Ltd
Period: Jan’ 11– Nov’ 11
Project: Accenture Customer Insight
Position: Technical Lead
Environment: Apache Solr, Apache Hadoop, Hive, Sqoop, Netezza, MicroStrategy, Java, Unix shell
scripting.
Description: Accenture Customer Insight (ACI) product in The AI Retail Management (R&D)
products portfolio is a high performance analytical engine that puts relevant information about
customer behaviors in the hands of non-technical users. It is uniquely designed to run against large
data volumes and draw reports on transactional, point-of-sale, third-party and customer loyalty data.
ACI supports a large number of users from inside and outside of the organization and allows them to
seamlessly run hundreds of reports each day.
Responsibilities:
Define and implement the data ingestion strategy
Work with clients on requirement analysis.
Customize java based web UI for customer requirements.
Design and customize reports and dashboards for various lines of business
Plan and support the client deployments
Manage offshore development team (7 members)
Business stakeholder management
Scope definition and management
Weekly reporting to the business
Risk & Issues management
Resolving business and system issues
People and Team Management
Risk analysis and mitigation
Client: Northwestern Mutual, USA
Organization: Infosys Technologies Pvt Ltd
Period: Dec’ 09– Jan’ 11
Project: WebSphere Version 7 Application Migration
Position: Team Lead
Environment: Java 1.4/1.6, WAS 6.0/7.0, MQ 6.0/7.0 , RSA 7.5 , IBM UDB/DB2 9.1, Sybase v 15,
IBM MQ Workflow
Description: Over 250 corporate applications are now on WebSphere Application Server, including
many of the top tier applications. WebSphere application server versions have finite lifecycles.
Migration of large suite of applications to newer versions requires early planning, quicker development
and deployment. Migrations now take two or more years to accomplish. In addition, WAS v6
infrastructure is nearing capacity, causing to either add to an older platform or devote our attention to
the newest one. This project involves migrating these enterprise applications to WAS 7 environment.
Applications are compiled to Java 1.6; underlying frameworks are upgraded to compatible versions
and set up on a brand new infrastructure and integration points are tested.
Responsibilities:
Define Roadmap/approach for migration
Understand AS-IS architecture/Identify changes and their impact
Prepare migration cookbooks
Prepare and update Traceability matrix
Help other teams to upgrade/resolve technical issues
Lead a team of 1 onsite developer and 12 offshore developers
Client: MetLife Insurance, USA
Organization: Infosys Technologies Pvt Ltd
Period: Mar’ 09– Dec’ 09
Project: Common Validator Framework
Position: Team Lead
Environment: Java 1.4, WAS 6.0, RSA 7.0, IBM UDB/DB2 9.1.
Description: Common Validator Framework is a replacement for the industry standard iLog rule
validator. Renewing iLog license every year was not a feasible option due to budget constraints. CVF
was then built as in house replacement for iLog. Rules for application validation were formulated into
a database up-loadable format and the core rule validator engine was built to process those rules for
different applications types in New Business area. New insurance or insurance change applications are
scanned for the input and corresponding rules are run by the rule engine to validate the input .
Responsibilities:
Requirement workshops with the customer.
Scope definition and management.
Weekly reporting to the clients
Calls with clients for project status and issues
Traceability matrix
Risk & Issues management
Resolving business and system issues.
People and Team Management
Risk analysis and mitigation
Client: Northwestern Mutual, USA
Organization: Infosys Technologies Pvt Ltd
Period: Feb’ 08– Feb’ 09
Project: WebSphere Version 7 Application Migration
Position: Program Analyst
Environment: Java 1.4, WAS 6.0, MQ 6.0, RAD 6.0 , IBM UDB/DB2 9.1, Sybase v 15, IBM MQ
Workflow.
Description: One of the big initiatives as Northwestern Mutual was to introduce Long Term Care
insurance to customers and sell directly through own Financial Representatives. Earlier the business
line was managed through other companies. As part of this initiative business processes reengineering was done for Long Term Care domain. It involved analysis of data flow interactions with
upstream and downstream systems built with heterogeneous technologies on multiple databases.
A core team was set up to define a roadmap to redesign the system in mainstream technologies and
also identify the business improvement opportunities. There were multiple tracks to identify the
current state business requirements, define future state requirements, solution architecture for the
new system..
Responsibilities:
Conducting requirement workshops with the customer and develop functional and Technical
specifications
Define re-engineering process for the business rules in legacy systems
Developing system requirements, system blue print and high level design.
Requirement analysis and development effort estimation.
Traceability matrix
Reviewing Detail Design documents, code and unit test results.
Resolving system issues.
Risk analysis and mitigation
Client: Northwestern Mutual, USA
Organization: Infosys Technologies Pvt Ltd
Period: Nov’ 05– Jan’ 08
Project: Downtown E-Files
Position: Program Analyst
Environment: Java 1.4.2, JEE 1.4, EJB 2.0, Struts 1.2, MQ 6.0, MQ Workflow 3.4, IBM CM 8.1, AIX,
HP Unix, MQ Broker 6.0, IBM UDB 8.0, Sybase 11.0, Visual Basic 6.0, PL I, COBOL, z/OS, Pervasive
SQL Btrieve, RAD 6.0.
Description: The E-Files solution was defined to eliminate paper-based processing in New Business,
DI & TC-TUP (in addition to the existing Franklin solution).
NM signed an Enterprise Software and Services option (ESSO) with IBM in 2004 that enables
unrestricted use of several IBM products for the next three years. The ESSO includes IBM Content
Manager and NM has utilized this opportunity to re-align its ECM strategy to the org-wide objective of
“Faster, Better, and Cheaper”. NM has made a strategic decision to move from Documentum to IBM
Content Manager (ICM) as its Enterprise Content Management (ECM) system. This resulted in IBM CM
Migration project.
Responsibilities:
Requirement capturing and requirement analysis
Developing system requirements, system blue print and high level design.
Reviewing Detail Design documents, code and unit test results.
Creating traceability matrix between requirements, detail design, code changes and Unit test cases.
Resolving business and system issues.
Performance testing of the components using Wily Interscope.
Client: Riverdeep Interactive Learning Inc., USA
Organization: NIIT Limited
Period: May’ 05– Oct’ 05
Project: Learning Village
Position: Software Engineer
Environment: Java 1.4.2, Struts 1.1, IBM UDB 6.0, WSAD 5.0
Description: Riverdeep Learning Village enables district administrators to develop, align and
distribute standards-based curriculum to all schools and all levels of teachers. It is a flexible
framework which can be adapted to a variety of instructional methodologies and to all areas within a
district’s curriculum. With Riverdeep Learning Village, a district can align standards to their curriculum,
making it accessible to all stakeholders within a school district, including administrators, teachers,
students and parents. Riverdeep Learning Village has been designed to adapt to how a district
functions rather than forcing a district to conform to a predefined construct.
School districts find immediate benefit using Riverdeep Learning Village as it enables
personalization to meet a district’s needs and align curriculum content to ever-changing standards.
Riverdeep Learning Village was designed for ease of use for district and local administrators to
manage curriculum services from their own desktop.
Responsibilities:
Requirement analysis
Coding, Unit testing, Unit Test Planning.
Code review and unit test result review.
Creating traceability matrix between requirements, detail design, code changes and Unit test cases.
Coordinating with the offsite team (based in India) and assigning the task to the team
Client: Reliance Infocomm limited
Organization: HTS India
Period: Jun’ 04– Apr’ 05
Project: Reliance Grievance Tracking System
Position: Jr. J2EE Architect
Environment: Java 1.4.2, Struts 1.1, IBM UDB 6.0, WSAD 5.0, Weblogic 8.1
Description: Objective of this project is to allow the client’s customer care to provide with the
capabilities like receiving complaints from customers, task assignment and workflow reporting for the
problems assigned, resolved, pending. The Grievance Tracking System helped customer care
department in improving their work and process efficiency.
Responsibilities:
To study the business requirements and identify the changes needed in the system.
Coding, Unit testing, functional testing and integration testing for all the components.