Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
4 views3 pages

GiriKumar BigData Developer

Giri Kumar has over 8 years of IT experience, specializing in Application Development, ETL, and Data Visualization, with significant expertise in Big Data technologies such as Spark, Sqoop, and Hive. He is currently a Big Data developer at Standard Chartered GBS, Chennai, and has worked on various projects involving data extraction, transformation, and analysis for banking and financial applications. His technical skills include proficiency in Hadoop, DataStage, SQL, and Unix, along with a strong background in performance tuning and CI/CD methodologies.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views3 pages

GiriKumar BigData Developer

Giri Kumar has over 8 years of IT experience, specializing in Application Development, ETL, and Data Visualization, with significant expertise in Big Data technologies such as Spark, Sqoop, and Hive. He is currently a Big Data developer at Standard Chartered GBS, Chennai, and has worked on various projects involving data extraction, transformation, and analysis for banking and financial applications. His technical skills include proficiency in Hadoop, DataStage, SQL, and Unix, along with a strong background in performance tuning and CI/CD methodologies.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 3

Giri Kumar Contact No: +91-9840988636

Email:[email protected]

Experience Summary:

 Having 8+ years of IT experience in IT with focus on Application Development, ETL


and Data visualization
o 3 years of experience in Big Data using Spark, Sqoop, Hive and HDFS.
o 5 years of experience in ETL DataStage, Oracle PL/SQL and Unix Scripting
 Experience in core Spark (batch processing) and Spark SQL using Scala language.
 Experience in manipulating/analyzing large datasets and finding patterns and insights
within structured and unstructured data
 Exposure in design and development of solutions for Big Data using the Hadoop and
Spark eco system technologies (Hive, Sqoop, Spark, Scala, HDFS)
 Process large volume of data from varies data sources like DB, files.
 Experience in migrating the data using Sqoop from HDFS to Relational Database System
and vice-versa according to client's requirement.
 Written Hive queries for data analysis to meet the business requirements
 Performance tuning of Hive queries using Partitioning/Bucketing.
 Performance tuning of Spark Jobs using DAG.
 Having knowledge in HBase.
 Good experience and knowledge in CI (Continuos Integration) and CD (Continuous
Deployment) Methodologies with Jenkins
 Experience in Moody’s tool
 Having knowledge in trouble shooting of context creation problems, GL Recon breaks,
Check errors and ALM/LIQ Stress process issues.
 Having knowledge in investigating LCR process related issues and provide solution.
 Well versed with Extraction, Transformation and Loading of Data using ETL tool –
DataStage with different versions like IBM InfoSphere DataStage and Qualitystage 8.1,
8.7 &11.3.
 Expertise in working with different data sources like Flat Files, RDBMS
 Good knowledge in SQL
 Having Knowledge in UNIX Commands and Shell Scripting
 Having Knowledge in scheduling the jobs and to automate the ETL process
 Strong analytic skills, self-driven, keen to learn and adopt new technologies and
processes
 Flexibility and Adaptability in regards with new technologies and environments willing to
work as a key player in challenging and creative environment
Employment Summary:

 Currently working as Big Data developer in Standard Chartered GBS, Chennai from
Nov’16 to till date.

Educational Qualification:

 B.Tech (Computers) From J.N.T.U, Anantapur.

Technical Skills:
Hadoop/Big Data HDFS, Hive, Sqoop, HBase, Spark, Scala

ETL Tool IBM InfoSphere DataStage and Qualitystage 8.1, 8.7&11.3

RDBMS Oracle 10g, Netezza 7


Languages SQL, Shell Scripting
Operating Systems Windows 2000 / 2003 Server/ XP, UNIX
Other Tools Control-M 9.0, Git, Bit Bucket, Jenkins, Autosys 11.1, Serena 4.5,
service now, Star Team 14.0

Projects Handled:
Project Name RCO
Client Standard Chartered Bank
Role Bigdata Developer
Technologies used Hadoop, HDFS, HIVE, Sqoop, Scala, Oracle and Spark
Project Tenure Feb 2019 – Till Date
Project Description Under Basel III there is a requirement for Banks to maintain an adequate level of
unencumbered, high-quality liquid assets that can be converted into cash to meet its
liquidity needs for a 30-calendar day time horizon under a significantly severe liquidity
stress scenario specified by supervisors. This is a continuous requirement (i.e.daily) and
given the uncertain timing of outflows and inflows, banks and supervisors are also
expected to be aware of any potential mismatches within the 30-day period and ensure that
enough liquid assets are available to meet any cash flow gaps throughout the period.

Responsibilities  Extracted the data from Oracle into HDFS using Sqoop
 Built Spark scripts by utilizing Scala depending on the requirement
 Created Hive tables to store the processed results in a tabular format
 Identified data issues and provided recommendations for resolution to ensure
optimal performance
 Involved in gathering requirements, designing, development and testing
 Involved in writing unit test cases
 Responsible for Coding and ensuring timely deliveries of work requests
 Coordinated with Testers and Developers in identifying and fixing issues when
involved in the SIT and UAT Testing team
 Responsible for providing the walk thru for the changes going for production and
also able to provide the technical solutions during deployment issues and
production bugs

Project Name ALM Fermat


Client Standard Chartered Bank
Role ETL DataStage Developer
Technologies used Data Stage 8.5, Control M, OBIEE, Linux, Windows 7, Moody’s Fermat
Project Tenure Nov 2016 – Jan 2019
Project Description The Fermat ALM solutions provide all the necessary features to perform risk and
profitability management. In terms of risk management Fermat ALM enables to handle
liquidity risk. Notably Fermat provides an in-depth analysis of the interest rate risk taking
in to account embedded options and economic indexes such as inflation. In terms of
profitability, Fermat ALM can at first level estimate the Net Interest income (NII) on the
balance sheet. It can also consider commissions assumptions and premium /discounts
deferments to estimate the net banking income (NBI) as well as the whole income
statement with in user defined GAAP.

Project Name SAMI


Client TIAA-CREF, USA
Role ETL DataStage Developer
Technologies used DataStage 8.1&11.3, Oracle 10g, Windows XP/UNIX, Tableau
Project Tenure Oct 2014 – Nov 2016
Project Description SAMI is Sales Analytics Market Intelligence. It’s used for the Field marketing which will
give the sales people analytical capability. It helps to find how a fund is working, how
much profit making compared other companies. Data from the SAMI is called competitive
data. We get prospective client info from the contacts from the real estate business they
had. Companies like Bloomberg sells data to companies like TIAA.
We implemented the Affiliated Data Repository that leverages the current Affiliated Sales
Data Structure to create a single Robust, in house platform to support Sales and Reporting
Business Intelligence initiatives for all affiliated flows, assets and relationships, contact &
client information. This is for Sales Pension & brokerage.

Project Name NORTH AMERICA CREDIT CARD DW


Client TD Bank - Banking, Canada
Role ETL DataStage Developer
Technologies used DataStage 8.7, Netezza 7.0, Windows XP/UNIX, SAS
Project Tenure Sep 2013 to Jul 2014
Project Description It is a strategic initiative to design and implement a Data Warehouse solution to store and
maintain the Product, Offers, Customer and Account centric data. This Data Warehouse
will be primarily used by the business users for analytical and reporting purposes and no
Operational system takes data from this Data Warehouse. This solution is aimed at
optimizing the current business processes and other dependencies. Successful
implementation of this internal solution would ensure easy access to the data for reporting
and analytical needs
Current Phase (Wave 1) of this project will involve 34 different source interfaces (42
source files) from various source systems. Focus will be on to make these files available in
the ‘iHUB’ Environment in an applicable format for further processing to populate the
data warehouse

You might also like