Home > Software > BIGDATA > HADOOP
Interview Questions   Tutorials   Discussions   Programs   Videos   Discussion   

HADOOP - What is Hadoop framework?




772
views
asked SRVMTrainings July 25, 2014 11:05 AM  

What is Hadoop framework?


           

8 Answers



 
answered By   0  
Hadoop framework is a software library that allows distributed processing of huge data sets across group of computers using simple programming models. It is designed in such a way that a group of thousands machines or a single machine can be used as a local computer and can used for storage.It can also detect and handle the failure by it self and is not dependent on other machines.

For further details of Hadoop framework and architecture you may go through this concise video and blogs:-


flag   
   add comment

 
answered By Janaki   0  

Hadoop framework is a set of opensource projects that works in coordination for solutioning bigdata problems. It uses Mapreduce principle where it breaks up the input data into small dat chrunches and process them in parallell and independent. If we want anyfile to be process byy Hadoop that has to be placed in HDFS which is the storage area for Hadoop.
flag   
   add comment

 
answered By   0  
hadoop is a open source framework.
hadoop is a batchprocessing,fault-tolerant system
hadoop runs on commodity hardware.
hadoop is a parallel programmimg model for storage and processing for massive amount of data.
hadoop supports high avialability(no data loss).
hadoop is a distributed filesystem data is stored on HDFS(does not have any predefined containers),data processing on MapReduce.

flag   
   add comment

 
answered By   0  

Hadoop framework provides a facility to store large and large amounts of data with almost no breakdown while querying. 

It breaks the file into pieces, copies it multiple times (3 default) and stores it on different machines. Accessibility is ensured even if any machine breaks down or is thrown out from network. 

One can use Map Reduce programs to access and manipulate the data. The developer need not worry where the data is stored, he/she can reference the data from a single view provided from the Master Node which stores all metadata of all the files stored across the cluster. 



flag   
   add comment

 
answered By   0  
• Solution for big data
       – Deals with complexities of high
       volume, velocity & variety of data
• Set of open source projects
• Transforms commodity hardware
into a service that:
    – Stores petabytes of data reliably
    – Allows huge distributed computations
• Key attributes:
    – Redundant and reliable (no data loss)
    – Extremely powerful
    – Batch processing centric
    – Easy to program distributed apps
    – Runs on commodity hardware


flag   
   add comment

 
answered By   0  
Hadoop is a apache framework developed completely in java. Hadoop analyze and process large amount of data i.e peta bytes of data in parallel with less time located in distributed environment. In hadoop system, the data is distributed in thousands of nodes and processes parallely Please find the very simple example in below PPT about Hadoop framework


http://www.cs.sun.ac.za/rw334/2012/Slides/RW334_32a.ppt
flag   
   add comment

 
answered By hadooptrainings   0  



flag   
   add comment

 
answered By hadooptrainings   0  




flag   
   add comment

Your answer

Join with account you already have

FF

Preview


Ready to start your tutorial with us? That's great! Send us an email and we will get back to you as soon as possible!

Alert