Big Data and Hadoop (2.0) Developer Training at Bangalore

Hi, just a reminder that you're receiving this email because you have expressed an interest in KnowledgeWorks IT Consulting Pvt. Ltd.,. Don't forget to add workshop@knowledgeworks.co.in to your address book so we'll be sure to land in your inbox!
 
You may unsubscribe if you no longer wish to receive our emails.





Big Data & Hadoop (2.0) Developer Training

Dates: 20th, 21st & 22nd Jun 2014 at Bangalore

3 Days Instructor Led Hands-On Training

 

 

Hurry Register Now! 

 


KnowledgeWorks Event Album: Click Here 



Overview:


 

Apache Hadoop enables organizations to analyze massive volumes of structured and unstructured data and is currently very hot trend across the software tech industry. Hadoop will be adopted as default enterprise data hub by most of the enterprise soon. Hence Hadoop is being tagged by many as one of the most desired tech skills for 2014 and coming years.
 
This course will provide you an excellent kick start in building your fundamentals in developing big data solutions using hadoop platform and its ecosystem tools. The course is well balanced between theory and hands-on lab (more than 15 lab exercises) spread on real world uses cases like retail data analysis, sentiment analysis, log analysis, real time trend analysis etc.

 


Who Should Attend:

Prerequisite:


 

Architects and developers, who wish to write, build and maintain Apache Hadoop jobs.

The participants should have basic knowledge of java, SQL and Linux. It is advised to refresh these skills to obtain maximum benefit from this workshop.

 

 

What participants will learn?

 

 

The attendees will learn below topics through lectures and hands-on exercises

- Understand Big Data, Hadoop 2.0 architecture and it's Ecosystem

- Deep Dive into HDFS and YARN Architecture

- Writing map reduce algorithms using java APIs

- Advanced Map Reduce features & Algorithms

- How to leverage Hive & Pig for structured and unstructured data analysis

- Data import and export using Sqoop and Flume and create workflows using Oozie

- Hadoop Best Practices, Sizing and capacity planning

- Creating reference architectures for big data solutions

 


Course Content:


 

* What is Big Data & Why Hadoop?

    * Big Data Characteristics, Challenges with traditional system

 * Hadoop Overview & it's Ecosystem

    * Anatomy of Hadoop Cluster, Installing and Configuring Hadoop

    * Setting up hadoop lab (Hortonworks HDP 2.0 sandbox VM)

 * HDFS and YARN

     * HDFS Architecture, Name Nodes, Data Nodes and Secondary Name Node

     * Understanding HDFS HA and Federation architecture

     * YARN Architecture, Resource Manager, Node Manager and Application Master

    * Hands-On Exercise

* Map Reduce Anatomy

    * How Map Reduce Works?

    * Writing Mapper, Reducer and Driver using Java APIs,

    * Understanding Hadoop Data Type, Input & Output Formats

    * Hands On Exercises

* Developing Map Reduce Programs

    * Setting up Eclipse Development Environment, Creating Map Reduce Projects, Debugging and Unit Testing

    * Developing a map reduce algorithm on real world scenario

   * Hands On Exercises

* Advanced Map Reduce Concepts

    * Combiner, Partitioner, Counter, Setup and cleanup, Distributed Cache

    * Passing parameters, Multiple Inputs, Chaining multiple jobs

 

    * Applying Compression, Speculative Execution, Zero Reducers

    * Handling small files and bad records

    * Handling Binary data like images, documents etc.

    * Map and Reduce Side Joins, data partitioning

    * Hands On Exercises

* Sqoop

    * Importing and Exporting data using Sqoop & Flume

    * Hands On Exercise

* Hive

    * Hive Architecture, Internal & External Tables, Partitioning, Buckets

    * Writing queries - Joins, Union, Dynamic partitioning, Sampling

    * Writing UDFs, reading different data formats and best practices

    * Hands On Exercise

* Pig

    * Pig Basics, Loading data files

    * Writing queries - SPLIT, FILTER, JOIN, GROUP, SAMPLE, ILLUSTRATE etc.

    * Writing UDFs and best practices

    * Hands On Exercise

* Hadoop Best Practices, Advanced Tips & Techniques

    * Managing HDFS and YARN

    * Hadoop Cluster sizing, capacity planning and optimization

    * Hadoop Deployment options

 


Faculty Profile: KnowledgeWorks Accredited Trainer


 

He has about 15+ years of industry experience working on enterprise Java, SOA and Cloud computing platforms. He has worked with TCS, HP, and iGATE Patni and worked on large scale projects for customers like Motorola, Home Depot, CKWB Bank, P&G in the roles of solution and technical architect. He is a freelance who provides consulting and training on Cloud Computing, Big data & Hadoop. He has been teaching Hadoop for 2 years and has trained more than 500 people in Hadoop from large MNCs like EMC, CISCO, HP, YODLEE, YAHOO, SAMSUNG, VeriSign, Success Factors etc.

 

 

 

Fee Details:

Rs. 21,000.00 + 12.36% Service Tax (Per Participant)

Subject to availability of seats.  Terms & Conditions 

 

Hurry Register Now! 

 

Time: 09:30 am to 05:30 pm

 

Venue Details:

KnowledgeWorks IT Consulting Pvt. Ltd.,

No. 65, Sri Vinayaka Tower, 3rd Floor, 8th B Main, 27th Cross,

Jayanagar 4th Block, Bangalore - 560011

 

For any clarifications contact:

Mr. Sudhindra D N: +91 9886221314 | T: +91 80 26630622, 22459941, 41533451

E: sudhi@knowledgeworks.co.in| W: www.knowledgeworks.co.in 

 

 





Like us on Facebook     Follow us on Twitter     View our profile on LinkedIn     Find us on Google+     Visit our blog
This email was sent to seekajob@outlook.com by workshop@knowledgeworks.co.in |  
KnowledgeWorks IT Consulting Pvt. Ltd., | No.65, Vinayaka Tower, 3rd Floor, | 8th B Main, 27th Cross, | Jayanagar 4th Block | Bangalore | Karnataka | 560011 | India

No comments:

Post a Comment