Sharad Patel

  • Hadoop Developer
  • Atlanta, GA
  • Member Since Jun 14, 2023

Candidates About

 

Sharad Patel

SUMMARY : 

Experienced Hadoop developer with strong foundation in distributed file systems like HDFS, HBase in big data environment. Excellent understanding of the complexities associated with big data with experience in developing modules and codes in MapReduce, Hive, Pig, Sqoop, Flume and Spark to address those complexities 

CORE QUALIFICATIONS

§  Hands on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, PIG, HIVE, HBASE, ZOOKEEPER, SQOOP, HUE, JSON. 

§  Good knowledge of Hadoop Development and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and Map Reduce concepts. 

§  Experience in installation, configuration, Management, supporting and monitoring Hadoop cluster using various distributions such as Apache and Cloudera. 

§  Good knowledge in programming Spark using Scala. 

§  Good understanding in processing of real-time data using Spark. 

§  Efficient in writing MapReduce Programs and using Apache Hadoop Map Reduce API for analyzing the structured and unstructured data. 

§  Experience in managing and reviewing Hadoop log files. 

§  Experience in writing external Pig Latin scripts. 

§  Experience in writing UDF's in java for Hive and pig. 

§  Experience in working with Flume/Kafka to load the log data from different sources into HDFS. 

§  Experience in using Apache Sqoop to import and export data to from HDFS and external RDBMS databases. 

§  Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and scheduling Hadoop jobs. 

§  Experience in using Hcatalog for Hive, Pig and Hbase. 

                                                                                                               

WORK EXPERIENCE

Feb 2017-July 2017                                                                                                         Hadoop Developer- AT&T -Atlanta, GA

Description: This project mainly deal with finding the churn prediction. It includes proper digging and analysis of both structured and unstructured data to get deeper insights into customer behaviour, their service usage patterns, preferences, and interests real-time. It minimizes customer defection by predicting which customers are likely to cancel a subscription to a service. Along with prediction it involves a lot much data engineering like acquisition, data cleaning, Data Munging and Data Transformation. 
 
Responsibilities: 

§  Involved in the process of data acquisition, data pre-processing and data exploration of telecommunication project in scala. 

§  As a part data acquisition in, used sqoop and flume to inject the data from server to hadoop using incremental import. 

§  In pre-processing phase used spark to remove all the missing data and data 
transformation to create new features. 

§  In data exploration stage used hive and impala to get some insights about the customer 
data. 

§  Used flume, sqoop, hadoop, spark and oozie for building data pipeline. 

§  Installed and configured Hadoop Map Reduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and Processing. 

§  Importing and exporting data into HDFS and Hive using Sqoop 

§  Experienced in defining job flows 

§  Experienced in managing and reviewing Hadoop log files. 

§  Load and transform large sets of structured, semi structured and unstructured data 

§  Responsible to manage data coming from different sources 

§  Cluster coordination services through Zookeeper. 

§  Involved in loading data from UNIX file system to HDFS. 

§  Installed and configured Hive and also written Hive UDFs. 

 

02/2014-11/2016                                                                                                                         Jumia Kenya- Hadoop Developer

Responsibilities: 

§  Involved in the process of data acquisition, data pre-processing and data exploration of  scala. 

§  As a part data acquisition in, used sqoop and flume to inject the data from server to hadoop using incremental import. 

§  In pre-processing phase used spark to remove all the missing data and data 
transformation to create new features. 

§  In data exploration stage used hive and impala to get some insights about the customer 
data. 

§  Used flume, sqoop, hadoop, spark and oozie for building data pipeline. 

§  Installed and configured Hadoop Map Reduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and Processing. 

§  Importing and exporting data into HDFS and Hive using Sqoop 

§  Experienced in defining job flows 

§  Experienced in managing and reviewing Hadoop log files. 

§  Load and transform large sets of structured, semi structured and unstructured data 

§  Responsible to manage data coming from different sources 

§  Cluster coordination services through Zookeeper. 

§  Involved in loading data from UNIX file system to HDFS. 

§  Installed and configured Hive and also written Hive UDFs. 

§  Involved in creating Hive tables, loading with data and writing hive queries which will 
run internally in map way 

§  Automated all the jobs, for pulling data from FTP server to load data into Hive tables, 
using Oozie workflows. 

Environment: Hadoop, Big Data, HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Flume, LINUX, Java, Eclipse, Hadoop Distribution of Cloudera., Windows, UNIX Shell Scripting, and Eclipse.

 

 

02/2013-02/2014                                                                                          Kabarak University:- Software Developer

§  Developed application logic using C#.

§  Written stored procedures, triggers using SQL in SQL SERVER 2005.

§  Worked intensely on the User Interface.

§  Developed WebForms using C# and ASP.NET.

§  Used AJAX in some webforms.

§  Worked on Cascading Style Sheets and XML-Schemas.

§  Developed Web Services for user authentication and file ransfer 

§  Used Xpath , XmlNode to access files and XMLDataDocument to synchronize with Datasets.

§  Compared the XML files sent by the Toll Authorities with XML Schemas.

§  Ensured Security to all the portals by creating Tampered proofed URLs.

§  Wrote triggers for sending Emails after to clients after any kind of transaction has been made

§  Extensively used GridViews sorting and paging

§  Implemented Template Columns for Custom Nested GridViews

§  Worked with Infragistic Controls extensively.

§  Worked on exporting reports to Excel from Gridviews and Ultrawebgrids.

§  Deployment of Application on Test and Production Server.

§  Handled many production issues and enhancement to the existing live portals

§  Environment

ASP.NET2.0, C#.NET, ADO.NET, Microsoft Visual Stuidio.NET 2012, IIS 5.0, SQL,XML, XSL, HTML, SQL Server 2000 Enterprise Manager, SQL Server 2005 Query Analyzer, .NETFramework 2.0, and Windows 2003 server 

 

EDUCATION

DeVry University, Decatur, USA

  • MS. Computer Network & Comm Mgmt with concentration in Information Security

Kabarak University, Kabarak, Kenya (GPA 4.0)

§  Bsc.Computer Science with concentration in Software Engineering course work and research included NFC mobile payment system, Title Deed Information System, Bluetooth enabled weight-capture system, Courier Management System and Study of network congestion at Kabarak University

Melvin Jones Lions Academy, Nakuru, Kenya

§  Cambridge IGCSE

 

TECHNICAL SKILLS

Big Data / Hadoop HDFS, MapReduce, HBase, Kafka, PIG, HIVE, Sqoop, Impala and Flume 
Real time/Stream Processing Apache Storm, Apache Spark 
Operating Systems Windows, Unix and Linux 
Programming Language Java, Scala, SQL 
Data Base Oracle 9i/10g, SQL Server, MS Access 
Web Technologies HTML, XML, JavaScript 
IDE Development Tools Eclipse, NetBeans 
Methodologies Agile, Scrum and Waterfall