Hadoop Technology

Atul Kushwaha
Presented by,
Atul Kushwaha
B.Tech, 3rd
Year(IT)
1109113033
Introduction
Big Data:
•Big data is a term used to describe the voluminous
amount of unstructured and semi-structured data a
company creates.
•Data that would take too much time and cost too much
money to load into a relational database for analysis.
• Big data doesn't refer to any specific quantity, the term
is often used when speaking about petabytes and
exabytes of data.
• The New York Stock Exchange generates about one terabyte of new
trade data per day.
• Facebook hosts approximately 10 billion photos, taking up one
petabyte of storage.
• The Internet Archive stores around 2 petabytes of data, and is
growing at a rate of 20 terabytes per month.
• The Large Hadron Collider near Geneva, Switzerland, produces about
15 petabytes of data per year.
Facts!!
So What Is The Problem?
 The transfer speed is around 100 MB/s
 A standard disk is 1 Terabyte
 Time to read entire disk= 10000 seconds or 3 Hours!
 Increase in processing time may not be as helpful because
• Network bandwidth is now more of a limiting factor
• Physical limits of processor chips have been reached
So What do We Do?
•The obvious solution is that
we use multiple processors to
solve the same problem by
fragmenting it into pieces.
•Imagine if we had 100 drives,
each holding one hundredth
of the data. Working in
parallel, we could read the
data in under two minutes.
Distributed Computing Vs
Parallelization
 Parallelization- Multiple processors or CPU’s
in a single machine
 Distributed Computing- Multiple computers
connected via a network
Problems In Distributed Computing
• Hardware Failure:
As soon as we start using many pieces of
hardware, the chance that one will fail is fairly
high.
• Combine the data after analysis:
Most analysis tasks need to be able to combine
the data in some way; data read from one
disk may need to be combined with the data
from any of the other 99 disks.
To The Rescue!
Apache Hadoop is a framework for running applications on large
cluster built of commodity hardware.
A common way of avoiding data loss is through replication:
redundant copies of the data are kept by the system so that in the
event of failure, there is another copy available. The Hadoop
Distributed Filesystem (HDFS), takes care of this problem.
The second problem is solved by a simple programming model-
Mapreduce. Hadoop is the popular open source implementation
of MapReduce, a powerful tool designed for deep analysis and
transformation of very large data sets.
What is ? ?
• It is an open source project by the Apache Foundation to handle
large data processing
• It was inspired by Google’s MapReduce and Google File System
(GFS) papers in 2003 and 2004
• It was originally conceived by Doug Cutting in 2005 and first used by
Yahoo! in 2006
• It is named after his son’s pet elephant incidentally
• It is basically a distributed file system which is written in Java.
Hadoop Approach to Distributed
Computing
 The theoretical 1000-CPU machine would cost a very
large amount of money, far more than 1,000 single-CPU.
 Hadoop will tie these smaller and more reasonably priced
machines together into a single cost-effective computer
cluster.
 Hadoop provides a simplified programming model which
allows the user to quickly write and test distributed
systems, and its’ efficient, automatic distribution of data
and work across machines and in turn utilizing the
underlying parallelism of the CPU cores.
Hadoop Components
HDFSHDFS
Storage
Self-healing
high-bandwidth
clustered storage
MapReduceMapReduce
Processing
Fault-tolerant
distributed
processing
Hadoop MapReduce
 MapReduce is a programming model
 Programs written in this functional style are automatically parallelized and
executed on a large cluster of commodity machines
 MapReduce is an associated implementation for processing and generating
large data sets.
The Programming Model Of MapReduce
 Map, written by the user, takes an input pair and produces a set of intermediate
key/value pairs. The MapReduce library groups together all intermediate values
associated with the same intermediate key I and passes them to the Reduce
function.
 The Reduce function, also written by the user, accepts an intermediate key I and a set of values
for that key. It merges together these values to form a possibly smaller set of values
 Filesystems that manage the storage across a network of machines
are called distributed filesystems.
 Hadoop comes with a distributed filesystem called HDFS, which
stands for Hadoop Distributed Filesystem.
 HDFS, the Hadoop Distributed File System, is a distributed file
system designed to hold very large amounts of data (terabytes or even
petabytes), and provide high-throughput access to this information.
 The Hadoop distributed file system (HDFS) is a distributed,
scalable, and portable file-system written in Java for the Hadoop
framework.
HADOOP DISTRIBUTED
FILESYSTEM (HDFS)
HDFS
 It manages storage on the cluster by breaking
incoming files into pieces, called blocks
 Stores each of the blocks redundantly across the
pool of servers
 It stores three complete copies of each file by
copying each piece to three different servers
Namenodes and Datanodes
 A HDFS cluster has two types of node operating in a master-slave
pattern: a namenode (the master) and a number of datanodes (slave).
 The namenode manages the filesystem namespace. It maintains the
filesystem tree and the metadata for all the files and directories in the
tree.
 Datanodes are the work horses of the filesystem.It manages storage
attached to the nodes that they run on.
 HDFS exposes a file system namespace and allows user data to be
stored in files.
 Internally, a file is split into one or more blocks and these blocks are
stored in a set of DataNodes.
Hadoop Technology
Who Uses Hadoop?
Advantages Over RDBMS
• Scalable: It can reliably store and process petabytes.
• Economical: It distributes the data and processing across
clusters of commonly available computers (in
thousands).
• Efficient: By distributing the data, it can process it in
parallel on the nodes where the data is located.
• Reliable: It automatically maintains multiple copies of
data and automatically redeploys computing tasks based
on failures.
Conclusion
 So major companies like facebook
amazon,yahoo,etc. are adapting Hadoop and in
future there can be many names in the list.
 This technology has bright future scope because day
by day need of data would increase and security
issues also the major point.
 Hence Hadoop Technology is the best appropriate
approach for handling the large data in smart way
and its future is bright…
Questions???
‘Thank You’
1 de 22

Recomendados

Introduction to Hadoop TechnologyIntroduction to Hadoop Technology
Introduction to Hadoop TechnologyManish Borkar
2.4K visualizações25 slides
Hadoop Seminar ReportHadoop Seminar Report
Hadoop Seminar ReportBhushan Kulkarni
3.8K visualizações27 slides
Hadoop technologyHadoop technology
Hadoop technologytipanagiriharika
2.9K visualizações85 slides

Mais conteúdo relacionado

Mais procurados

 Hadoop And Their Ecosystem ppt Hadoop And Their Ecosystem ppt
Hadoop And Their Ecosystem pptsunera pathan
9.9K visualizações24 slides
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoopjoelcrabb
2.5K visualizações30 slides
Hadoop Hadoop
Hadoop ABHIJEET RAJ
525 visualizações31 slides
HADOOP  TECHNOLOGY pptHADOOP  TECHNOLOGY ppt
HADOOP TECHNOLOGY pptsravya raju
38.5K visualizações22 slides
Seminar Presentation HadoopSeminar Presentation Hadoop
Seminar Presentation HadoopVarun Narang
81.5K visualizações51 slides

Mais procurados(20)

 Hadoop And Their Ecosystem ppt Hadoop And Their Ecosystem ppt
Hadoop And Their Ecosystem ppt
sunera pathan9.9K visualizações
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoop
joelcrabb2.5K visualizações
Hadoop Hadoop
Hadoop
ABHIJEET RAJ525 visualizações
HADOOP  TECHNOLOGY pptHADOOP  TECHNOLOGY ppt
HADOOP TECHNOLOGY ppt
sravya raju38.5K visualizações
Seminar Presentation HadoopSeminar Presentation Hadoop
Seminar Presentation Hadoop
Varun Narang81.5K visualizações
Introduction to Hadoop and Hadoop component Introduction to Hadoop and Hadoop component
Introduction to Hadoop and Hadoop component
rebeccatho2.7K visualizações
Hadoop Map ReduceHadoop Map Reduce
Hadoop Map Reduce
VNIT-ACM Student Chapter11.1K visualizações
Hadoop Tutorial For BeginnersHadoop Tutorial For Beginners
Hadoop Tutorial For Beginners
Dataflair Web Services Pvt Ltd1.3K visualizações
An Introduction to HadoopAn Introduction to Hadoop
An Introduction to Hadoop
DerrekYoungDotCom5.4K visualizações
Introduction to HiveQLIntroduction to HiveQL
Introduction to HiveQL
kristinferrier3.8K visualizações
Map ReduceMap Reduce
Map Reduce
Prashant Gupta14.8K visualizações
Hadoop ecosystemHadoop ecosystem
Hadoop ecosystem
Stanley Wang2.9K visualizações
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoop
Dr. C.V. Suresh Babu2K visualizações
HadoopHadoop
Hadoop
Nishant Gandhi19.5K visualizações
Hadoop seminarHadoop seminar
Hadoop seminar
KrishnenduKrishh477 visualizações
Introduction to Map ReduceIntroduction to Map Reduce
Introduction to Map Reduce
Apache Apex7.7K visualizações
Introduction to HDFSIntroduction to HDFS
Introduction to HDFS
Bhavesh Padharia2.4K visualizações
Big Data and HadoopBig Data and Hadoop
Big Data and Hadoop
Flavio Vit2.6K visualizações
Hadoop EcosystemHadoop Ecosystem
Hadoop Ecosystem
Sandip Darwade2.9K visualizações
Big data  Hadoop  presentation Big data  Hadoop  presentation
Big data Hadoop presentation
Shivanee garg558 visualizações

Destaque

Hadoop TechnologiesHadoop Technologies
Hadoop TechnologiesKannappan Sirchabesan
5.8K visualizações10 slides
Big data and HadoopBig data and Hadoop
Big data and HadoopRahul Agarwal
82.3K visualizações18 slides
Computer network pptComputer network ppt
Computer network pptSantosh Delwar
211K visualizações24 slides
Hadoop Overview & Architecture  Hadoop Overview & Architecture
Hadoop Overview & Architecture EMC
64.6K visualizações91 slides
Basic concepts of computer NetworkingBasic concepts of computer Networking
Basic concepts of computer NetworkingHj Habib
175.3K visualizações21 slides

Destaque(6)

Hadoop TechnologiesHadoop Technologies
Hadoop Technologies
Kannappan Sirchabesan5.8K visualizações
Practical Problem Solving with Apache Hadoop & PigPractical Problem Solving with Apache Hadoop & Pig
Practical Problem Solving with Apache Hadoop & Pig
Milind Bhandarkar237.2K visualizações
Big data and HadoopBig data and Hadoop
Big data and Hadoop
Rahul Agarwal82.3K visualizações
Computer network pptComputer network ppt
Computer network ppt
Santosh Delwar211K visualizações
Hadoop Overview & Architecture  Hadoop Overview & Architecture
Hadoop Overview & Architecture
EMC64.6K visualizações
Basic concepts of computer NetworkingBasic concepts of computer Networking
Basic concepts of computer Networking
Hj Habib175.3K visualizações

Similar a Hadoop Technology

Seminar pptSeminar ppt
Seminar pptRajatTripathi34
49 visualizações20 slides
hadoophadoop
hadoopswatic018
96 visualizações29 slides
hadoophadoop
hadoopswatic018
169 visualizações29 slides
getFamiliarWithHadoopgetFamiliarWithHadoop
getFamiliarWithHadoopAmirReza Mohammadi
54 visualizações54 slides
Managing Big data with HadoopManaging Big data with Hadoop
Managing Big data with HadoopNalini Mehta
786 visualizações24 slides
Big data and hadoopBig data and hadoop
Big data and hadoopRoushan Sinha
30 visualizações21 slides

Similar a Hadoop Technology(20)

Seminar pptSeminar ppt
Seminar ppt
RajatTripathi3449 visualizações
hadoophadoop
hadoop
swatic01896 visualizações
hadoophadoop
hadoop
swatic018169 visualizações
getFamiliarWithHadoopgetFamiliarWithHadoop
getFamiliarWithHadoop
AmirReza Mohammadi54 visualizações
Managing Big data with HadoopManaging Big data with Hadoop
Managing Big data with Hadoop
Nalini Mehta786 visualizações
Big data and hadoopBig data and hadoop
Big data and hadoop
Roushan Sinha30 visualizações
Big Data and HadoopBig Data and Hadoop
Big Data and Hadoop
Mr. Ankit253 visualizações
Apache Hadoop Big Data TechnologyApache Hadoop Big Data Technology
Apache Hadoop Big Data Technology
Jay Nagar493 visualizações
Apache hadoop basicsApache hadoop basics
Apache hadoop basics
saili mane1.1K visualizações
HadoopHadoop
Hadoop
Ankit Prasad168 visualizações
Topic 9a-Hadoop Storage- HDFS.pptxTopic 9a-Hadoop Storage- HDFS.pptx
Topic 9a-Hadoop Storage- HDFS.pptx
DanishMahmood2314 visualizações
OPERATING SYSTEM .pptxOPERATING SYSTEM .pptx
OPERATING SYSTEM .pptx
AltafKhadim6 visualizações
Bigdata and Hadoop IntroductionBigdata and Hadoop Introduction
Bigdata and Hadoop Introduction
umapavankumar kethavarapu408 visualizações
Hadoop  architecture-tutorialHadoop  architecture-tutorial
Hadoop architecture-tutorial
vinayiqbusiness38 visualizações
Cppt HadoopCppt Hadoop
Cppt Hadoop
chunkypandey1227 visualizações
CpptCppt
Cppt
chunkypandey12126 visualizações
CpptCppt
Cppt
chunkypandey12147 visualizações
Unit-1 Introduction to Big Data.pptxUnit-1 Introduction to Big Data.pptx
Unit-1 Introduction to Big Data.pptx
AnkitChauhan81782620 visualizações
2. hadoop fundamentals2. hadoop fundamentals
2. hadoop fundamentals
Lokesh Ramaswamy727 visualizações
AnjuAnju
Anju
Anju Shekhawat417 visualizações

Último

IWISS Catalog 2022IWISS Catalog 2022
IWISS Catalog 2022Iwiss Tools Co.,Ltd
22 visualizações66 slides
Informed search algorithms.pptxInformed search algorithms.pptx
Informed search algorithms.pptxDr.Shweta
12 visualizações19 slides
SNMPxSNMPx
SNMPxAmatullahbutt
10 visualizações12 slides

Último(20)

IWISS Catalog 2022IWISS Catalog 2022
IWISS Catalog 2022
Iwiss Tools Co.,Ltd22 visualizações
Activated sludge process .pdfActivated sludge process .pdf
Activated sludge process .pdf
8832RafiyaAltaf6 visualizações
Extensions of Time - Contract ManagementExtensions of Time - Contract Management
Extensions of Time - Contract Management
brainquisitive12 visualizações
Informed search algorithms.pptxInformed search algorithms.pptx
Informed search algorithms.pptx
Dr.Shweta12 visualizações
SNMPxSNMPx
SNMPx
Amatullahbutt10 visualizações
Saikat Chakraborty Java Oracle Certificate.pdfSaikat Chakraborty Java Oracle Certificate.pdf
Saikat Chakraborty Java Oracle Certificate.pdf
SaikatChakraborty7871489 visualizações
Investor PresentationInvestor Presentation
Investor Presentation
eser sevinç10 visualizações
Codes and Conventions.pptxCodes and Conventions.pptx
Codes and Conventions.pptx
IsabellaGraceAnkers5 visualizações
Object Oriented Programming with JAVAObject Oriented Programming with JAVA
Object Oriented Programming with JAVA
Demian Antony D'Mello49 visualizações
9_DVD_Dynamic_logic_circuits.pdf9_DVD_Dynamic_logic_circuits.pdf
9_DVD_Dynamic_logic_circuits.pdf
Usha Mehta5 visualizações
Solar PVSolar PV
Solar PV
Iwiss Tools Co.,Ltd10 visualizações
String.pptxString.pptx
String.pptx
Ananthi Palanisamy45 visualizações
Stone Masonry and Brick Masonry.pdfStone Masonry and Brick Masonry.pdf
Stone Masonry and Brick Masonry.pdf
Mohammed Abdullah Laskar17 visualizações
IEC 61850 Technical Overview.pdfIEC 61850 Technical Overview.pdf
IEC 61850 Technical Overview.pdf
ssusereeea975 visualizações
cloud computing-virtualization.pptxcloud computing-virtualization.pptx
cloud computing-virtualization.pptx
RajaulKarim2072 visualizações

Hadoop Technology

  • 1. Presented by, Atul Kushwaha B.Tech, 3rd Year(IT) 1109113033
  • 2. Introduction Big Data: •Big data is a term used to describe the voluminous amount of unstructured and semi-structured data a company creates. •Data that would take too much time and cost too much money to load into a relational database for analysis. • Big data doesn't refer to any specific quantity, the term is often used when speaking about petabytes and exabytes of data.
  • 3. • The New York Stock Exchange generates about one terabyte of new trade data per day. • Facebook hosts approximately 10 billion photos, taking up one petabyte of storage. • The Internet Archive stores around 2 petabytes of data, and is growing at a rate of 20 terabytes per month. • The Large Hadron Collider near Geneva, Switzerland, produces about 15 petabytes of data per year. Facts!!
  • 4. So What Is The Problem?  The transfer speed is around 100 MB/s  A standard disk is 1 Terabyte  Time to read entire disk= 10000 seconds or 3 Hours!  Increase in processing time may not be as helpful because • Network bandwidth is now more of a limiting factor • Physical limits of processor chips have been reached
  • 5. So What do We Do? •The obvious solution is that we use multiple processors to solve the same problem by fragmenting it into pieces. •Imagine if we had 100 drives, each holding one hundredth of the data. Working in parallel, we could read the data in under two minutes.
  • 6. Distributed Computing Vs Parallelization  Parallelization- Multiple processors or CPU’s in a single machine  Distributed Computing- Multiple computers connected via a network
  • 7. Problems In Distributed Computing • Hardware Failure: As soon as we start using many pieces of hardware, the chance that one will fail is fairly high. • Combine the data after analysis: Most analysis tasks need to be able to combine the data in some way; data read from one disk may need to be combined with the data from any of the other 99 disks.
  • 8. To The Rescue! Apache Hadoop is a framework for running applications on large cluster built of commodity hardware. A common way of avoiding data loss is through replication: redundant copies of the data are kept by the system so that in the event of failure, there is another copy available. The Hadoop Distributed Filesystem (HDFS), takes care of this problem. The second problem is solved by a simple programming model- Mapreduce. Hadoop is the popular open source implementation of MapReduce, a powerful tool designed for deep analysis and transformation of very large data sets.
  • 9. What is ? ? • It is an open source project by the Apache Foundation to handle large data processing • It was inspired by Google’s MapReduce and Google File System (GFS) papers in 2003 and 2004 • It was originally conceived by Doug Cutting in 2005 and first used by Yahoo! in 2006 • It is named after his son’s pet elephant incidentally • It is basically a distributed file system which is written in Java.
  • 10. Hadoop Approach to Distributed Computing  The theoretical 1000-CPU machine would cost a very large amount of money, far more than 1,000 single-CPU.  Hadoop will tie these smaller and more reasonably priced machines together into a single cost-effective computer cluster.  Hadoop provides a simplified programming model which allows the user to quickly write and test distributed systems, and its’ efficient, automatic distribution of data and work across machines and in turn utilizing the underlying parallelism of the CPU cores.
  • 12. Hadoop MapReduce  MapReduce is a programming model  Programs written in this functional style are automatically parallelized and executed on a large cluster of commodity machines  MapReduce is an associated implementation for processing and generating large data sets.
  • 13. The Programming Model Of MapReduce  Map, written by the user, takes an input pair and produces a set of intermediate key/value pairs. The MapReduce library groups together all intermediate values associated with the same intermediate key I and passes them to the Reduce function.
  • 14.  The Reduce function, also written by the user, accepts an intermediate key I and a set of values for that key. It merges together these values to form a possibly smaller set of values
  • 15.  Filesystems that manage the storage across a network of machines are called distributed filesystems.  Hadoop comes with a distributed filesystem called HDFS, which stands for Hadoop Distributed Filesystem.  HDFS, the Hadoop Distributed File System, is a distributed file system designed to hold very large amounts of data (terabytes or even petabytes), and provide high-throughput access to this information.  The Hadoop distributed file system (HDFS) is a distributed, scalable, and portable file-system written in Java for the Hadoop framework. HADOOP DISTRIBUTED FILESYSTEM (HDFS)
  • 16. HDFS  It manages storage on the cluster by breaking incoming files into pieces, called blocks  Stores each of the blocks redundantly across the pool of servers  It stores three complete copies of each file by copying each piece to three different servers
  • 17. Namenodes and Datanodes  A HDFS cluster has two types of node operating in a master-slave pattern: a namenode (the master) and a number of datanodes (slave).  The namenode manages the filesystem namespace. It maintains the filesystem tree and the metadata for all the files and directories in the tree.  Datanodes are the work horses of the filesystem.It manages storage attached to the nodes that they run on.  HDFS exposes a file system namespace and allows user data to be stored in files.  Internally, a file is split into one or more blocks and these blocks are stored in a set of DataNodes.
  • 20. Advantages Over RDBMS • Scalable: It can reliably store and process petabytes. • Economical: It distributes the data and processing across clusters of commonly available computers (in thousands). • Efficient: By distributing the data, it can process it in parallel on the nodes where the data is located. • Reliable: It automatically maintains multiple copies of data and automatically redeploys computing tasks based on failures.
  • 21. Conclusion  So major companies like facebook amazon,yahoo,etc. are adapting Hadoop and in future there can be many names in the list.  This technology has bright future scope because day by day need of data would increase and security issues also the major point.  Hence Hadoop Technology is the best appropriate approach for handling the large data in smart way and its future is bright…

Notas do Editor

  1. Gm…I think everybody heard abt hadoop bcz before fews days an institutional lecture was held in our clg but nobody attended it…so I m give u a brief introduction abt hadoop what it is,what is the need of it,its feature and applications.basically hadoop deals with big data.
  2. Apache Hadoop is an open-source software framework for storage and large scale processing of data-sets on clusters of commodity hardware. Hadoop is an Apache top-level project being built and used by a global community of contributors and users.[2] It is licensed under the Apache License 2.0.
  3. Hadoop is designed to efficiently process large volumes of information by connecting many commodity computers together to work in parallel. A 1000 CPU single machine (i.e a supercomputer with a vast memory storage) would cost a lot. Thus Hadoop parallelizes the computation by tying smaller and more reasonably priced machines together into a single costeffective compute cluster.
  4. Hadoop is designed to efficiently process large volumes of information by connecting many commodity computers together to work in parallel. A 1000 CPU single machine (i.e a supercomputer with a vast memory storage) would cost a lot. Thus Hadoop parallelizes the computation by tying smaller and more reasonably priced machines together into a single costeffective compute cluster.
  5. HDFS cluster/healing. MapReduce
  6. MapReduce is a programming model for processing large data sets with a parallel, distributed algorithm on a cluster.[1]
  7. First explain distributed file system?-managing storage across n/w of machines.
  8. Replication in computing involves sharing information so as to ensure consistency between redundant resources, such as software or hardware components, to improve reliability, fault-tolerance, or accessibility.
  9. Rackspace for log processing. Netflix for recommendations. LinkedIn for social graph. SU for page recommendations.