Learn
Big Data Hadoop Overview
What is Big Data?
Big data can be described as a huge volume of structured, unstructured, text, images, audio, video, log files, emails, simulations, 3D models, military surveillance, e-commerce and so on that is so massive that it's difficult to process using traditional database and software techniques. In most enterprise scenarios the data is too big or it moves too fast or it exceeds current processing capacity.
We can formally define Big Data as:
Big Data usually includes data sets with sizes beyond the ability of commonly used software tools to capture, curate, manage, and process the data within a tolerable elapsed time.
What is Hadoop?
It is an Apache open source software project that enables the distributed processing of large data sets across clusters of commodity servers.
The following session explains in detail about:
- Big Data
- Problems related to Big Data
- Impact of Big Data in today’s world.
- Introduction to Hadoop
- Hadoop Architecture
- Main Hadoop components
If you would like to know how to install various Big Data Technologies, please visit our Installations page.