If there are trillions of files coming to the system, what techniques will we use to push data in Hadoop. Will we use a staging environment?

Asked by Last Modified  

Follow 1
Answer

Please enter your answer

Beatrice-be innovative and knowledgeable

U know first of all we can directly store the data in memory rather than in disks for which my RAM management will be volatile where I can easily loose data but that will cause a problem with my metadata so immediately I have to create an interval offtime within which I have to create a disk where I...
read more
U know first of all we can directly store the data in memory rather than in disks for which my RAM management will be volatile where I can easily loose data but that will cause a problem with my metadata so immediately I have to create an interval offtime within which I have to create a disk where I can store the backup of my metadata which is basically called as FS IMAGE but again it will become big right ,so suppose today FSIMAGE is subdivided in two versions FS1 and FS2 there the backup of your data is done within every 24 hrs but as it's regularly not possible to backup my data daily we have to create a small file where we can easily store the backup of all our activities and that file is called as an editlog then we will surely not loose our metadata and some of the huge stored files will also get pushed up in Hadoop. Well I don't have a much knowledge about all these things whatever I have said just from my basic knowledge,I just wanted to try, so please don't mind or comment anything if I have done a mistake by answering this question. read less
Comments

Related Questions

Can anyone suggest about Hadoop?
Hadoop is good but it depends on your experience. If you don't know basic java, linux, shell scripting. Hadoop is not beneficial for you.
Ajay
What are the biggest pain points with Hadoop?
The biggest pain points with Hadoop are its complexity in setup and maintenance, slow processing due to disk I/O, high resource consumption, and difficulty in handling real-time data.
Anish
0 0
6
I want a lady Hadoop Trainer.
Yes. Career bridge it services, one of the best training insitute in hyderabad. we provide lady trainer for ofline / online batches. Please call and contact @970-532-3377. So that you can get all the details about trainings and career guiidance.
Chandrika

What is difference between data science and SAP. Which is best in compare for getting jobs as fast as possible

Hi Both have different uniquness with importance value. you will get a good prospectives on SAP for career growth.
Ravindra

Now ask question in any of the 1000+ Categories, and get Answers from Tutors and Trainers on UrbanPro.com

Ask a Question

Related Lessons

Hadoop v/s Spark
1. Introduction to Apache Spark: It is a framework for performing general data analytics on distributed computing cluster like Hadoop.It provides in memory computations for increase speed and data process...

Big Data
Bigdata Large amount of data and data may be various types such as structured, unstructured, and semi-structured, the data which cannot processed by our traditional database applications are not enough....

Why is the Hadoop essential?
Capacity to store and process large measures of any information, rapidly. With information volumes and assortments always expanding, particularly from web-based life and the Internet of Things (IoT), that...

HDFS And Mapreduce
1. HDFS (Hadoop Distributed File System): Makes distributed filesystem look like a regular filesystem. Breaks files down into blocks. Distributes blocks to different nodes in the cluster based on...

Loading Hive tables as a parquet File
Hive tables are very important when it comes to Hadoop and Spark as both can integrate and process the tables in Hive. Let's see how we can create a hive table that internally stores the records in it...

Recommended Articles

In the domain of Information Technology, there is always a lot to learn and implement. However, some technologies have a relatively higher demand than the rest of the others. So here are some popular IT courses for the present and upcoming future: Cloud Computing Cloud Computing is a computing technique which is used...

Read full article >

Big data is a phrase which is used to describe a very large amount of structured (or unstructured) data. This data is so “big” that it gets problematic to be handled using conventional database techniques and software.  A Big Data Scientist is a business employee who is responsible for handling and statistically evaluating...

Read full article >

We have already discussed why and how “Big Data” is all set to revolutionize our lives, professions and the way we communicate. Data is growing by leaps and bounds. The Walmart database handles over 2.6 petabytes of massive data from several million customer transactions every hour. Facebook database, similarly handles...

Read full article >

Hadoop is a framework which has been developed for organizing and analysing big chunks of data for a business. Suppose you have a file larger than your system’s storage capacity and you can’t store it. Hadoop helps in storing bigger files than what could be stored on one particular server. You can therefore store very,...

Read full article >

Find Hadoop near you

Looking for Hadoop ?

Learn from the Best Tutors on UrbanPro

Are you a Tutor or Training Institute?

Join UrbanPro Today to find students near you