Giri Kumar Contact No: +91-9840988636
Email:[email protected]
Experience Summary:
Having 8+ years of IT experience in IT with focus on Application Development, ETL
and Data visualization
o 3 years of experience in Big Data using Spark, Sqoop, Hive and HDFS.
o 5 years of experience in ETL DataStage, Oracle PL/SQL and Unix Scripting
Experience in core Spark (batch processing) and Spark SQL using Scala language.
Experience in manipulating/analyzing large datasets and finding patterns and insights
within structured and unstructured data
Exposure in design and development of solutions for Big Data using the Hadoop and
Spark eco system technologies (Hive, Sqoop, Spark, Scala, HDFS)
Process large volume of data from varies data sources like DB, files.
Experience in migrating the data using Sqoop from HDFS to Relational Database System
and vice-versa according to client's requirement.
Written Hive queries for data analysis to meet the business requirements
Performance tuning of Hive queries using Partitioning/Bucketing.
Performance tuning of Spark Jobs using DAG.
Having knowledge in HBase.
Good experience and knowledge in CI (Continuos Integration) and CD (Continuous
Deployment) Methodologies with Jenkins
Experience in Moody’s tool
Having knowledge in trouble shooting of context creation problems, GL Recon breaks,
Check errors and ALM/LIQ Stress process issues.
Having knowledge in investigating LCR process related issues and provide solution.
Well versed with Extraction, Transformation and Loading of Data using ETL tool –
DataStage with different versions like IBM InfoSphere DataStage and Qualitystage 8.1,
8.7 &11.3.
Expertise in working with different data sources like Flat Files, RDBMS
Good knowledge in SQL
Having Knowledge in UNIX Commands and Shell Scripting
Having Knowledge in scheduling the jobs and to automate the ETL process
Strong analytic skills, self-driven, keen to learn and adopt new technologies and
processes
Flexibility and Adaptability in regards with new technologies and environments willing to
work as a key player in challenging and creative environment
Employment Summary:
Currently working as Big Data developer in Standard Chartered GBS, Chennai from
Nov’16 to till date.
Educational Qualification:
B.Tech (Computers) From J.N.T.U, Anantapur.
Technical Skills:
Hadoop/Big Data HDFS, Hive, Sqoop, HBase, Spark, Scala
ETL Tool IBM InfoSphere DataStage and Qualitystage 8.1, 8.7&11.3
RDBMS Oracle 10g, Netezza 7
Languages SQL, Shell Scripting
Operating Systems Windows 2000 / 2003 Server/ XP, UNIX
Other Tools Control-M 9.0, Git, Bit Bucket, Jenkins, Autosys 11.1, Serena 4.5,
service now, Star Team 14.0
Projects Handled:
Project Name RCO
Client Standard Chartered Bank
Role Bigdata Developer
Technologies used Hadoop, HDFS, HIVE, Sqoop, Scala, Oracle and Spark
Project Tenure Feb 2019 – Till Date
Project Description Under Basel III there is a requirement for Banks to maintain an adequate level of
unencumbered, high-quality liquid assets that can be converted into cash to meet its
liquidity needs for a 30-calendar day time horizon under a significantly severe liquidity
stress scenario specified by supervisors. This is a continuous requirement (i.e.daily) and
given the uncertain timing of outflows and inflows, banks and supervisors are also
expected to be aware of any potential mismatches within the 30-day period and ensure that
enough liquid assets are available to meet any cash flow gaps throughout the period.
Responsibilities Extracted the data from Oracle into HDFS using Sqoop
Built Spark scripts by utilizing Scala depending on the requirement
Created Hive tables to store the processed results in a tabular format
Identified data issues and provided recommendations for resolution to ensure
optimal performance
Involved in gathering requirements, designing, development and testing
Involved in writing unit test cases
Responsible for Coding and ensuring timely deliveries of work requests
Coordinated with Testers and Developers in identifying and fixing issues when
involved in the SIT and UAT Testing team
Responsible for providing the walk thru for the changes going for production and
also able to provide the technical solutions during deployment issues and
production bugs
Project Name ALM Fermat
Client Standard Chartered Bank
Role ETL DataStage Developer
Technologies used Data Stage 8.5, Control M, OBIEE, Linux, Windows 7, Moody’s Fermat
Project Tenure Nov 2016 – Jan 2019
Project Description The Fermat ALM solutions provide all the necessary features to perform risk and
profitability management. In terms of risk management Fermat ALM enables to handle
liquidity risk. Notably Fermat provides an in-depth analysis of the interest rate risk taking
in to account embedded options and economic indexes such as inflation. In terms of
profitability, Fermat ALM can at first level estimate the Net Interest income (NII) on the
balance sheet. It can also consider commissions assumptions and premium /discounts
deferments to estimate the net banking income (NBI) as well as the whole income
statement with in user defined GAAP.
Project Name SAMI
Client TIAA-CREF, USA
Role ETL DataStage Developer
Technologies used DataStage 8.1&11.3, Oracle 10g, Windows XP/UNIX, Tableau
Project Tenure Oct 2014 – Nov 2016
Project Description SAMI is Sales Analytics Market Intelligence. It’s used for the Field marketing which will
give the sales people analytical capability. It helps to find how a fund is working, how
much profit making compared other companies. Data from the SAMI is called competitive
data. We get prospective client info from the contacts from the real estate business they
had. Companies like Bloomberg sells data to companies like TIAA.
We implemented the Affiliated Data Repository that leverages the current Affiliated Sales
Data Structure to create a single Robust, in house platform to support Sales and Reporting
Business Intelligence initiatives for all affiliated flows, assets and relationships, contact &
client information. This is for Sales Pension & brokerage.
Project Name NORTH AMERICA CREDIT CARD DW
Client TD Bank - Banking, Canada
Role ETL DataStage Developer
Technologies used DataStage 8.7, Netezza 7.0, Windows XP/UNIX, SAS
Project Tenure Sep 2013 to Jul 2014
Project Description It is a strategic initiative to design and implement a Data Warehouse solution to store and
maintain the Product, Offers, Customer and Account centric data. This Data Warehouse
will be primarily used by the business users for analytical and reporting purposes and no
Operational system takes data from this Data Warehouse. This solution is aimed at
optimizing the current business processes and other dependencies. Successful
implementation of this internal solution would ensure easy access to the data for reporting
and analytical needs
Current Phase (Wave 1) of this project will involve 34 different source interfaces (42
source files) from various source systems. Focus will be on to make these files available in
the ‘iHUB’ Environment in an applicable format for further processing to populate the
data warehouse