manish_resume

6
Manish Bajaj [email protected] +91 8904109668 PROFESSIONAL SUMMARY Having a holistic experience of 1 year in the field of BigData, especially on Hadoop framework and its ecosystem. As a BigData Engineer, have a strong leverage on Hadoop stack with interest in solving problems related to BigData. My work so far has been focused on ETL based solutions, data management, algorithms on Mapreduce framework. 1 Year of experience in developing JAVA applications using Java technologies in the Banking and Financial domain, creating standalone java tools using core java. Proficient in Core Java, Hadoop, MapReduce,Pig, Hive, Flume, Oozie. Good experience of remote debugging with eclipse IDE. Capable of learning new technologies and software and understanding them quickly. Excellent analytical, interpersonal and customer relational skills. Highly motivated team player with excellent communication, presentation and interpersonal skills, always willing to work in challenging and cross-platform environment. Sound knowledge of the business domain, the functionality of which has been developed by me. My personal projects can found in the following link: https://github.com/bajaj1989/ Skill Set: Technical: Programming Language - Java, HiveQL, Pig script Frameworks - Big Data, Hadoop, MapReduce Cluster - HDFS Tools / IDEs - Eclipse, SQL Developer Database - Oracle, MYSQL Configuration Tools - Tortoise Git, GitHub

Upload: manish-bajaj

Post on 14-Jan-2017

132 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Manish_resume

Manish [email protected] +91 8904109668

PROFESSIONAL SUMMARY Having a holistic experience of 1 year in the field of BigData, especially on Hadoop framework and its

ecosystem. As a BigData Engineer, have a strong leverage on Hadoop stack with interest in solving prob-

lems related to BigData. My work so far has been focused on ETL based solutions, data management, algo-

rithms on Mapreduce framework.

1 Year of experience in developing JAVA applications using Java technologies in the Banking and Finan-

cial domain, creating standalone java tools using core java.

Proficient in Core Java, Hadoop, MapReduce,Pig, Hive, Flume, Oozie.

Good experience of remote debugging with eclipse IDE.

Capable of learning new technologies and software and understanding them quickly.

Excellent analytical, interpersonal and customer relational skills.

Highly motivated team player with excellent communication, presentation and interpersonal skills, always

willing to work in challenging and cross-platform environment.

Sound knowledge of the business domain, the functionality of which has been developed by me.

My personal projects can found in the following link:

https://github.com/bajaj1989/

Skill Set: Technical:

Programming Language - Java, HiveQL, Pig script Frameworks - Big Data, Hadoop, MapReduce Cluster - HDFS Tools / IDEs - Eclipse, SQL Developer Database - Oracle, MYSQL Configuration Tools - Tortoise Git, GitHub

Work Experience: Organization: Tata Consultancy Services Mar2013 - Till Date

Designation: Systems Engineer

PROJECTS UNDERTAKEN :

Client – TCS Internal Jun 2014 – Till Date

CBK TF TEAM( Business Module Development Team)Client Name : TCS InternalProject Type : Analysis/Development/Performance EnhancementTechnologies Used : Hadoop, MapReduce, Pig, Hive, Java, Sqoop, SQL

Page 2: Manish_resume

Manish Bajaj

IDE : HDFS, Oracle SQL developer Role : Team Member

Brief description of the project:

The project involves gathering customer related data, data related to transactions via credit/debit card, entire raw financial data from different sources and converts this into data with financial standards.

Roles and Responsibilities:

My current role involves extraction, transformation and loading data and analyzing it. I have used mapreduce algorithm and pig script for data transformation and also used sqoop scripts to ingest data from RDBMS(MySQL) and dump it into HDFS cluster. Our cluster consist of 30 nodes running on YARN and a stand-by namenode as per High Availability implementation on HDFS. The approximate size of the data is 40 TB. I have written HIVE queries to create external tables with partition on top of data present and analyze it.

Client – Commercial Bank Of Kuwait Jul2013– jun2014

CBK TF TEAM( Business Module Development Team)

Application : TCS BancsClient Name : Commercial Bank Of Kuwait (CBK)Project Type : Optimization/Development/Performance EnhancementTechnologies Used : Java , SQL, WebServices, Spring Batch FrameworkIDE : Oracle SQL developer Server : WebLogicRole : Team Member

Brief description of the project:

The project involved designing and developing the Trade Finance module for the core banking product ad customize it according to the requirement of the client. It is the financing of international trade. Trade finance includes such activities as lending, issuing letters of credit, factoring, export credit and insurance. Companies involved with trade finance include importers and exporters, banks and financiers, insurers and export credit agencies, as well as other service providers. Trade finance is of vital importance to the global economy, with the World Trade Organization estimating that 80 to 90% of global trade is reliant on this method of financing.

My Role:

Analyzing the business module by going through the Functional Specification Document and prepare a de-sign for the requirement.

Developing the functionality as required. Created complete flows for the Documentory Collection module, right from lodgement of bills to the pay-

ment of the bills. Developed EOD(end of day) spring batches for certain specific requirements. Involved in SIT defect fixing and support of benchmarking tests.

Page 3: Manish_resume

Manish Bajaj

Client – Deutsche Bank Mar 2013 – Jun 2013

BORIS ( TCS Bancs Implementation for DB)

Application : DB CoralClient Name : Deutsche Bank GermanyProject Type : DevelopmentTechnologies Used : Java, J2EE, SQLIDE : Eclipse , Oracle SQL developer Server : WebLogicRole : Team Member

Brief description of the project:

The project involves enhancement of TCS Bancs product as per the requirements of Deutsche Bank. The key areas of work involve the development of the Inventory Management module of the product. The Inventory included drafts and cheque books to be maintained internally by the bank.

My Role:

Analyzed the functional specification document . Created complete flows for Inventory Management module (both backend coding) . Also developed a design for the transfer of Inventory from higher level chief cashiers to lower level tellers

of the bank.. Identified and made changes in the code to close unclosed connection objects, statement objects and result

sets. Involved in SIT defect fixing and support of benchmarking tests .

PERSONAL PROJECTS UNDERTAKEN:

Twitter Sentiment Analysis:

The project involved gathering data on some specific popular current topics from twitter and determine the sentiment of the data. We configured apache flume to gather data from twitter and dump it into hdfs. The raw json data was then stored in to hive table using json serde. We then developed a Hive UDF for determining the sentiment of the data by marking certain keywords and emoticons as 'positive', 'negative' and 'neutral' and then based on the count of these keywords determined the sentiment.The github link to the project:https://github.com/bajaj1989/Twitter_sent_analysis/tree/master

ClickStream Data Analysis:

The project involved analysis of ClickStream Data generated from an E-commerce website. The volume of data generated per day on the site is approximately 2GB. My role involved refining the dataset and load it into HDFS. Using MapReduce framework, first job was to enhance the dataset and then refining the data based on sessionID and timestamp of a particular user. Later we dump this data into HIVE and using HIVE we performed various analysis on the dataset.

Web-Scrapping:

Page 4: Manish_resume

Manish Bajaj

This project involved gathering various forms of datasets directly from web pages. We used sources like IMDB and gathered the information of all movies since 1900 and refined the data to get the rating, name and year of the movie. Then we used pandas dataframes and matplotlib to plot a graph showing the rating of movies. We also gathered data from wikipedia and refined it and loaded the data into HDFS and performed various sorts of analysis on those datasets.

Training and Certification:

Attended training on Java, Hadoop Framework, Pig, HiveQL. Attended training on basic UNIX commands . Attended training on Python, Web Scrapping (CS-109). Undergoing courses on R and data science on Coursera. Have uploaded the Twitter sentiment analysis project on github(link mentioned above). Have participated in the Big data contest conducted by Impetus on Techgig.com

Academic Profile:

Secondary School Certificate from I.C.S.E board with 90% aggregate in 2006. Higher Secondary Certificate from C.B.S.E board with 82% aggregate in 2008. B.Tech in Electronics and Communication Engineering - From West Bengal University Of Technology,

Kolkata, with a DGPA of 8.77 in 2012.

DECLARATION

I do hereby declare that all the statements furnished above are true to the best of my knowledge and belief.

Manish Bajaj