Word of the day: Hadoop Cluster!
Hadoop cluster
A Hadoop cluster
 is a special type of computational cluster designed specifically for 
storing and analyzing huge amounts of unstructured data in a distributed
 computing environment. 
Such clusters
 run Hadoop's open source distributed processing software on low-cost 
commodity computers. Typically one machine in the cluster is designated 
as the NameNode and another machine the as JobTracker; these are the 
masters. The rest of the machines in the cluster act as both DataNode 
and TaskTracker; these are the slaves. Hadoop clusters are often 
referred to as "shared nothing" systems because the only thing that is 
shared between nodes is the network that
 connects them. 
Hadoop 
clusters are known for boosting the speed of data analysis applications.
 They also are highly scalable: If a cluster's processing power is 
overwhelmed by growing volumes of data, additional cluster nodes can be 
added to increase throughput. Hadoop clusters also are highly resistant 
to failure because each piece of data is copied onto other cluster 
nodes, which ensures that the data is not lost if one node fails.  WhatIs.com
 
Comments
Post a Comment