Hadoop Online Training with real time experienced certified trainers Apache HADOOP training with Hive and Hbase Attend a demo on Apache Hadoop online Training
3. Contents
What is Hadoop.
Why do we need Hadoop.
How Hadoop works.
HDFS Architecture.
What is Map – Reduce.
Hadoop Cluster.
Hadoop Processes.
Topology of a Hadoop Cluster.
Distinction of Hadoop Framework .
Prerequisites to learn hadoop.
4. What is Hadoop
Hadoop is an open Source Framework.
Developed by Apache Software Foundation.
Used for distributed processing of large date
sets.
It works across clusters of computers using a
simple programming model (Map-Reduce).
5. Why do we need Hadoop
Data is growing faster.
Need to process multi petabytes of data.
The performance of traditional applications
is decreasing.
The number of machines in a cluster is not
constant.
Failure is expected, rather than exceptional.
6. How Hadoop Works
The Hadoop core consists of two modules :
Hadoop Distributed File System (HDFS)
[Storage].
Map Reduce [Processing].
Mapper
Reducer
8. What is Map – Reduce
Map Reduce plays a key role in hadoop framework.
Map Reduce is a Programming model for writing
applications that rapidly process large amount of
data.
Mapper – is a function that processes input data to
generate intermediate output data.
Reducer – Merges all intermediate data from all
mappers and generate final output data.
9. Hadoop Cluster
A Hadoop Cluster consist of multiple
machines Which can be classified into 3
types
Namenode
Secondary Namenode
Datanode
10. Hadoop Processes
Below are the daemons (Processes) Which
runs in a cluster.
Name node (Runs on a master machine)
Job Tracker (Runs on a master machine)
Data node (Runs on slave machines)
Task Tracker (Runs on slave machines)
12. Distinction
Simple – Hadoop allows users to quickly
write efficient parallel code.
Reliable – Because Hadoop runs on
commodity hardware, it can face frequent
automatically handle such failures.
Scalable – we can increase or decrease the
number of nodes (machine) in hadoop
cluster.
13. Prerequisites
Linux bases operating system (Mac OS, Redhat,
ubuntu)
Java 1.6 or higher version
Disk space ( To hold HDFS data and it’s
replications )
Ram (Recommended 2GB)
A cluster of computers.
You can even install Hadoop on single machine.
For more information visit this link: hadoop online
training
14. Newyorksys.com
NewyorkSys is one of the leading Training
Company in US. We have certified trainers. We
provide Online Training, Fast Track online training,
with job assistance. We are providing excellent
Training in all courses. We also help you in resume
preparation and provide job assistance till you get
job.
For more details Visit : http://www.newyorksys.com
15 Roaring Brook Rd, Chappaqua, NY 10514.
USA: +1-718-313-0499 , 718-305-1757
Email : enquiry@newyorksys.us
15. About Newyorksys
Newyorksys , the leading IT online training company is
going to conduct its exclusive Hadoop online training
programs. We have a comprehensive and engaging
curriculum for delivering the hadoop concepts
Labels: hadoop online training, hadoop training, hadoop
course, hadoop online course, hadoop training in usa,
hadoop technologies, hadoop training online, hadoop
classes in usa, hadoop training and development, online
hadoop training, apache hadoop training, apache
hadoop course in usa