Sujatha Sudam
Hadoop Administrator
443-525-2153
[email protected]
PROFESSIONAL SUMMARY
Over 8+ Years of extensive IT experience with 4 +years of experience as a Hadoop Administrator and 4
years of experience as a UNIX/Linux Administrator along with SQL developer in designing and
implementing Relational Database models as per business needs in different domains.
Expertise in Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, and
Troubleshooting.
Hands-on experience in installation, configuration, supporting, and managing Hadoop Clusters using
Apache, Cloudera (CDH5.x), and Yarn distributions.
Hands-on experience in backup configuration and Recovery from a Name Node failure.
Decommissioning and commissioning the Node on running the Hadoop cluster.
Extensive experience in installation, configuration, management, and deployment of Big Data
components and the underlying infrastructure of the Hadoop Cluster.
Involved in benchmarking Hadoop/HBase cluster file systems various batch jobs and workloads
Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage, and network
Good experience in designing, configuring, and managing the backup and disaster recovery for
Hadoop data.
Experience in Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning servers
for optimal performance of the cluster.
As an admin involved in Cluster maintenance, troubleshooting, Monitoring, and following proper
backup and Recovery strategies.
Experience in HDFS data storage and support for running map-reduce jobs.
Good working knowledge of importing and exporting data from different databases namely MySQL,
PostgreSQL, Oracle into HDFS, and Hive using Sqoop.
Extensive experience in NoSQL and real-time analytics.
Strong knowledge of yarn terminology and the High-Availability Hadoop Clusters.
Hands-on experience in analyzing Log files for Hadoop and ecosystem services and finding the root
cause.
Experience in Chef, Puppet, or related tools for configuration management.
Expertise in Installing, Configuration, and Managing RedHat Linux 5, 6.
Good experience in scheduling cron jobs in Linux.
Proactively maintain and develop all Linux infrastructure technology to maintain a 24x7x365 uptime
service
Maintain best practices for managing systems and services across all environments
Fault finding, analysis, and logging information for reporting performance exceptions
Manage, coordinate, and implement software upgrades, patches, and hotfixes on servers,
workstations, and network hardware
Provide input on ways to improve the stability, security, efficiency, and scalability of the environment
Install and maintain all server hardware and software systems and administer all server performance
and ensure availability for the same.
Perform tests on all new software and maintain patches for management services and perform an
audit on all security processes
Collaborate with other teams and team members to develop automation strategies and deployment
processes
Provided root cause analysis of incident reports during any downtime issues
Provided customers with administrative support on a UNIX-based platform historical query database
serving many users.
TECHNICAL EXPERTISE
Hadoop HDFS, Map Reduce, Python, Pig, Hive, Sqoop, HBase, Spark, Zookeeper, Oozie, Kafka, Storm,
Framework Tez, Impala, Flume
OS Redhat Linux, UNIX, Windows 2000/NT/XP, Sun Solaris
Languages C, C++, SAS, PL/SQL
Scripting Unix, Perl Script
Languages
Database Oracle 10g/11g, SQL server, Teradata
Database Tools Oracle SQL Developer, SQL Plus
Version Control CVS, SVN, Git
PROFESSIONAL EXPERIENCE:
Client: USAA, San Antonio, TX Aug ’21 – Present
Role: Sr. Hadoop Administrator
Responsibilities:
Installation of Cloudera Services not found in Hortonworks.Installation of Cloudera Services not found
in Hortonworks.
Worked on migrating existing clusters from Hortonworks to Cloudera.
Updated admin scripts to be both Hortonworks and Cloudera compatible.
Scheduled various admin scripts using Control-M.
Using the monitoring tool Data dog, built custom monitors for our processes.
Worked on deploying various code changes across environments using GIT as source version control
and UCD as deploying tool.
Applied vendor provided hotfixes to resolve application related issues.
Worked extensively to address user related requests and queries.
Performed POC to test out spark dynamic allocation.
Worked on creating pipelines to load data from various sources into our Hadoop data lake.
Performed addition of new fields to hive tables and loading of historical data for them.
Built utility scripts in python to automate alter table generation scripts when adding new fields.
Developed script in python to automate table creation if not present by pulling DDL from GIT.
Built automated pipelines to load data from source files into snowflake tables and views.
Performed performance tuning on existing queries to make them more efficient.
Client: Capital One, Richmond, VA Apr ’18 – Aug’ 21
Role: Hadoop Administrator
Responsibilities:
Installation of various Hadoop Ecosystems and Hadoop Daemons.
Responsible for Disaster Recovery setup of production cluster.
Production is supported for all applications that are running on the cluster.
Created jobs to take periodic backups of configuration files and create HDFS snapshots.
Developed Scripts and Batch jobs to schedule various platform engineering tasks like snapshot
cleanups, logs archival, etc.
Involved in the start-to-end process of the Hadoop cluster setup in installation, configuration, and
monitoring of the Hadoop Cluster.
Built automated setup for cluster monitoring and issue escalation process.
Responsible for Cluster maintenance, commissioning, and decommissioning of Data nodes, Cluster
Monitoring, Troubleshooting, Managing, and reviewing data backups, Hadoop log files
Monitoring systems and services, architecture design and implementation of Hadoop deployment,
configuration management, backup, and disaster recovery systems and procedures.
Configured various properties to improve the performance of the cluster.
Monitored multiple Hadoop cluster environments using Cloudera Manager and other third-party
tools. Monitored workload, job performance, and capacity planning
Expertise in recommending hardware configuration for Hadoop cluster
Installing, Upgrading, and Managing Hadoop Cluster on Cloudera distribution
Managing and reviewing Hadoop log files to investigate any job failures.
Experience with Unix or Linux, including shell scripting
Environment: Hadoop, HDFS, Map Reduce, Shell Scripting, spark, solr, Pig, Hive, HBase, Sqoop, Flume, Oozie,
Zookeeper, cluster health, monitoring security, RedhatLinux, impala, Cloudera Manager
Client: Comcast, Philadelphia, PA May’17– Mar ’18
Role: Hadoop Administrator
Responsibilities:
Responsible for loading the customer's data and event logs from Oracle database, Teradata into HDFS
using Sqoop
End-to-end performance tuning of Hadoop clusters and Hadoop MapReduce routines against very
large data sets.
Performance tuning of Hadoop and Kafka cluster.
Diligently teaming with the infrastructure, network, database, application, and business intelligence
teams to guarantee high data quality and availability
Installed and configured various components of the Hadoop ecosystem and maintained their integrity
Designed, configured, and managed the backup and disaster recovery for HDFS data.
Experience with Unix or Linux, including shell scripting
Installing, Upgrading, and Managing Hadoop Cluster on Cloudera distribution.
Commissioned Data Nodes when data grew and decommissioned when the hardware degraded
Responsible for cluster planning and expansion strategy.
Expertise in recommending hardware configuration for Hadoop cluster
Installing, Upgrading, and Managing Hadoop Cluster on Cloudera distribution
Installed and configured Hadoop HDFS, MapReduce, Pig, Hive, and Sqoop.
Responsible for upgrades and patching of Hadoop clusters.
Managing and reviewing Hadoop and HBase log files.
Writing shell scripts to automate admin tasks of monitoring the cluster.
Environment: Hadoop, Map Reduce, Shell Scripting, spark, Pig, Hive, Cloudera Manager, CDH 5.4.3, HDFS,
Yarn, Hue, Sentry, Oozie, Zookeeper, Impala, Solr, Kerberos, cluster health, Puppet, Ganglia, Nagios, Flume,
Sqoop, Storm, Kafka, KMS
Client: Bed Bath & Beyond, Northern, NJ Jan’16 – Apr ’17
Role: Hadoop/ UNIX Administrator
Responsibilities:
Design and maintain all system tools for all scripts and automation processes and monitor all capacity
planning.
Integrate all required software and resolve all issues across various technologies and design require
enterprise servers and provide backup support.
Evaluate all documents according to system requirements and evaluate all designs and perform tests
on all development activities and administer all complex methodologies.
Develop an infrastructure to provide support to all business requirements and perform regular
troubleshooting on a system to resolve all issues.
Implemented and setup Virtualization environments for AIX LPARs, HP Integrity VMs; and Solaris
Zones and Logical Domains
Updated and created provisioning scripts to setup new operating systems and software for supported
platforms
Consolidated servers at numerous smaller remote data centers to three central data centers
Stabilized system by disk replacement, firmware upgrade in SAN storage, Solaris Volume
Management, and clustering environment on scheduled maintenance hours.
Enhanced business continuity procedure by adding critical middleware servers and identified through
power-down test activity.
Resolved issues, planned requests as point-of-contact to vendors, oversaw developers, and business
users, followed change control procedures, and reported results Monitor everyday systems and
evaluated the availability of all server resources and perform all activities for Linux servers.
Maintain and monitor all patch releases and design various patch installation strategies and maintain
all systems according to NIST standardization.
Administer all performance for various resources and ensure optimization for the same and provide
support to all applications and ensure the optimal level of customer service.
Maintain and monitor all system frameworks and provide after-call support to all systems and
maintain optimal Linux knowledge.
Wrote and maintained shell scripts using Perl and Bash.
Monitored, troubleshot, and resolved issues involving operating systems.
Applied ITIL approach to incident and problem management.
Developed and maintained a troubleshooting journal for the incident management team.
Participated in an on-call rotation to provide 24×7 technical support
Tested numerous software and hardware configurations during the development stages to recreate
the operating environments utilized by customers in an effort to avoid the distribution of releases
with bugs and/or erroneous documentation
Wrote utility scripts using BASH and KORN shell
Configured UNIX systems to use Active Directory, KERBEROS, NTPD, XDMCP, LDAP, SSH, FTP, TFTP,
and DNS
Performed problem diagnosis, corrected discrepancies, developed user and maintenance
documentation, provided user assistance, and evaluate system performance
Installed, and configured third-party applications, hardened new and existing servers and desktops
Environment: KERBEROS, monitoring tool HP OpenView ITO (OVO), Redhat Linux, Windows, FTP, Solaris,
HP_UX with Oracle, Sybase
Client: YASH Technologies, Hyderabad Sep’14 – Aug’15
Role: Hadoop Administrator
Responsibilities:
Responsible for loading the customer's data and event logs from Oracle database, Teradata into HDFS
using Sqoop
We are involved in initiating and successfully completing Proof of Concept on SQOOP for Pre-
Processing, Increased Reliability, and Ease of Scalability over traditional Oracle databases.
End-to-end performance tuning of Hadoop clusters and Hadoop MapReduce routines against very
large data sets.
Involved in loading data from the LINUX file system to HDFS.
Wrote and maintained shell scripts using Perl and Bash.
Monitored, troubleshot, and resolved issues involving operating systems.
Applied ITIL approach to incident and problem management.
Developed and maintained a troubleshooting journal for the incident management team.
Importing and exporting data into HDFS and Hive using Sqoop and Flume.
Proficient in using Cloudera Manager, an end-to-end tool to manage Hadoop operations.
Wrote MapReduce jobs to generate reports for the number of activities created on a particular day,
during a dump from multiple sources, and the output was written back to HDFS
Reviewed the HDFS usage and system design for future scalability and fault tolerance.
Installed and configured Hadoop HDFS, MapReduce, Pig, Hive, and Sqoop.
Cluster coordination services through Zookeeper.
Written the Spouts and Bolts after collecting the real stream customer data from the Kafka broker to
process and store into HBASE.
Analyze the log files and process them through Flume
Experience in the optimization of the MapReduce algorithm using combiners and partitions to deliver
the best results and worked on Application performance optimization.
Environment: UNIX Scripting, Hadoop, MapReduce, HDFS, Pig, Sqoop, Hive, Oracle, Teradata, and Eclipse
Client: Airtel, Hyderabad, India Nov’12 – Aug’14
Role: UNIX Administrator
Responsibilities:
Maintain and monitor all patch releases and design various patch installation strategies and maintain
all systems according to NIST standardization.
Administer all performance for various resources and ensure optimization for the same and provide
support to all applications and ensure the optimal level of customer service.
Maintain and monitor all system frameworks and provide after-call support to all systems and
maintain optimal Linux knowledge.
Assisted developers with troubleshooting custom software, and services such as ActiveSync, CalDav,
CardDav, and PHP
Top-level customer service and implementation for DKIM, SPF, and custom SSL/TLS security
Implemented and performed initial Nimble Storage CS460G-X2 array configuration and migrated data
from legacy BlueArc Titan storage array. Converted access from NFS to iSCSI
Assigned to selected projects and successfully defined hardware and software needs to complete
them.
Served as Project lead on updating hardware and software for the backup schema on both Windows
and UNIX/LINUX-based development networks.
Troubleshoot any errors found in code using simple PERL scripts.
Documented troubleshooting guide for administrators to be used for on-call pager duty.
Attended team meetings and handled light managerial duties in the absence of the team lead.
Environment: UNIX, Solaris, HP_UX, Red Hat Linux, Windows, FTP, SFTP