1. PRAVEEN REDDY GAJJALA
Professional Summary :
Having 2 years of overall IT experience.
2 years experience in Big Data and Hadoop and its components HDFS, Map
Reduce, Pig, Hive, Sqoop.Oozie and Hbase.
Good knowledge on Map Reduce Programming in Java programming language..
Extensive knowledge in Hive installation, configuration, setting up Metastore
and UDF Creation.
Good in integrating computer skills, customer support and education to meet
client needs and make customer delight.
Involved in Schema Design and installing the Hive software.
Involved in writing the Pig scripts to reduce the job execution.
Exceptional ability to learn new concepts.
Hard working and enthusiastic.
Knowledge on FLUME and NO-SQL
Experience in different operating Systems UNIX, LINUX, and WINDOWS.
Working as a Hadoop Developer in WIPRO , Hyderabad from Dec 2012 to Till
Academic profile :
Bachelor of Engineering in Electrical and Electronics (2008-2012),
KORM College Of Engineering, kadapa. (JNTU ANANTAPUR)
Technical skills :
Languages Core Java, Map Reduce, Pig, Sqoop, Hive, Hbase.
Web Technologies HTML, Hadoop API.
Java IDEs Eclipse.
Databases Oracle, My SQL
Operating Systems Windows7, Windows XP, 2000, 2003, Unix and Linux
2. Project :
Client : Sears
Environment : Hadoop , Apache Pig, Hive, HBase,Oozie. Java, Linux ,SQL
Eclipse, Oracle 10g, MapReduce, HDFS
Duration : Jan 2013 to till Date.
Role : Hadoop Developer
The purpose of the project is to improve customer's shopping
experience with Sears. Collect click stream data from Sears websites and mobile apps and
analyze the shopping patterns with these application and customize customer facing
applications to make it user friendly for the customer in reaching the products in fewer
clicks and also personalizing the shopping experience. Hadoop is used to collect and store
data from various data points and perform various analyses on these data using
MapReduce jobs, Pig and Hive Scripts. Aggregated results are then exported over to
downstream RDBMS for Business Intelligence reporting.
Roles and Responsibilities:
Worked on a live Hadoop production CDH3 cluster with 35 nodes
Worked with highly unstructured and semi structured data of 25 TB in size
Good experience in benchmarking Hadoop cluster.
Used Sqoop to import data from DB2 system in to HDFS
Worked on custom Map Reduce programs using Java
Designed and developed PIG data transformation scripts to work against
unstructured data from various data points and created a base line.
Worked on creating and optimizing Hive scripts for data analysts based on the
Good experience in working with Sequence files and compressed file formats.
Worked with performance issues and tuning the Pig and Hive scripts.
Exported the analyzed data to the relational databases using Sqoop for
visualization and to generate reports for the BI team.
Writing the script files for processing data and loading to HDFS
Worked with the infrastructure and the admin teams to set up monitoring probes
to track the health of the nodes
Created and maintained Technical documentation for launching Hadoop Clusters
and for executing Hive queries and Pig Scripts.