Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Where does Big Data go and how is it stored?

I'm trying to get to grips with Big Data, and mainly with how Big Data is managed.

I'm familiar with the traditional form of data management and data life cycle; e.g.:

  1. Structured data collected (e.g. web form)
  2. Data stored in tables in an RDBMS on a database server
  3. Data cleaned and then ETL'd into a Data Warehouse
  4. Data is analysed using OLAP cubes and various other BI tools/techniques

However, in the case of Big Data, I'm confused about the equivalent version of points 2 and 3, mainly because I'm unsure about whether or not every Big Data "solution" always involves the use of a NoSQL database to handle and store unstructured data, and also what the Big Data equivalent is of a Data Warehouse.

From what I've seen, in some cases NoSQL isn't always used and can be totally omitted - is this true?

To me, the Big Data life cycle goes something on the lines of this:

  1. Data collected (structured/unstructured/semi)
  2. Data stored in NoSQL database on a Big Data platform; e.g. HBase on MapR Hadoop distribution of servers.
  3. Big Data analytic/data mining tools clean and analyse data

But I have a feeling that this isn't always the case, and point 3 may be totally wrong altogether. Can anyone shed some light on this?

like image 472
RoyalSwish Avatar asked Apr 20 '17 16:04

RoyalSwish


People also ask

What is big data and how it works?

With Big Data you store schemaless as first (often referred as unstructured data) on a distributed file system. This file system splits the huge data into blocks (typically around 128 MB) and distributes them in the cluster nodes. As the blocks get replicated, nodes can also go down.

How is big data stored and stored?

Big data can go into all sorts of systems and be stored in numerous ways, but it's often stored without structure first, and then it's turned into structured data clusters during the extract, transform, load (ETL) stage.

What is the future of big data management and storage?

Of course, it is undeniable that big data management and storage is moving promptly away from physical machinery and swiftly into the digital arena. Alongside all the developments in technology, big data grows faster, at such a rate that all the machines and warehouses in the world can’t quite contain it all.

What is big data and how does it affect cloud computing?

Alongside all the developments in technology, big data grows faster, at such a rate that all the machines and warehouses in the world can’t quite contain it all. Consequently, cloud computing has experienced a boom thanks to cloud storage services propelling the digital transition.


1 Answers

When we talk about Big Data, we talk in most cases about huge amount of data that is many cases constantly written. Data can have a lot of variety as well. Think of a typical data source for Big Data as a machine in a production line that produces all the time sensor data on temperature, humidity, etc. Not the typical kind of data you would find in your DWH.

What would happen if you transform all this data to fit into a relational database? If you have worked with ETL a lot, you know that extracting from the source, transforming the data to fit into a schema and then to store it takes time and it is a bottle neck. Creating a schema is too slow. Also mostly this solution is to costly as you need expensive appliances to run your DWH. You would not want to fill it with sensor data.

You need fast writes on cheap hardware. With Big Data you store schemaless as first (often referred as unstructured data) on a distributed file system. This file system splits the huge data into blocks (typically around 128 MB) and distributes them in the cluster nodes. As the blocks get replicated, nodes can also go down.

If you are coming from the traditional DWH world, you are used to technologies that can work well with data that is well prepared and structured. Hadoop and co are good for looking for insights like the search for the needle in the hay stack. You gain the power to generate insights by parallelising data processing and you process huge amount of data.

Imagine you collected Terabytes of data and you want to run some analytical analysis on it (e.g. a clustering). If you had to run it on a single machine it would take hours. The key of big data systems is to parallelise execution in a shared nothing architecture. If you want to increase performance, you can add hardware to scale out horizontally. With that you speed up your search with a huge amount of data.

Looking at a modern Big Data stack, you have data storage. This can be Hadoop with a distributed file system such as HDFS or a similar file system. Then you have on top of it a resource manager that manages the access on the file system. Then again on top of it, you have a data processing engine such as Apache Spark that orchestrates the execution on the storage layer.

Again on the core engine for data processing, you have applications and frameworks such as machine learning APIs that allow you to find patterns within your data. You can run either unsupervised learning algorithms to detect structure (such as a clustering algorithm) or supervised machine learning algorithms to give some meaning to patterns in the data and to be able to predict outcomes (e.g. linear regression or random forests).

This is my Big Data in a nutshell for people who are experienced with traditional database systems.

like image 175
Stefan Papp Avatar answered Oct 15 '22 00:10

Stefan Papp