What is Hadoop ?🤔
If You Interested In Big Data analytic Field You Definitely Know This Framework !
Ok Let's See What is Hadoop & How It Works With Big Data analytic
Hadoop is an open-source framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models.
It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.
- Hadoop Distributed File System (HDFS):
A distributed file system that provides high-throughput access to application data.
HDFS splits large data files into smaller blocks and distributes them across a cluster of machines. This ensures fault tolerance and high availability.
- MapReduce:
A programming model and processing engine for large-scale data processing.
The MapReduce model consists of two main functions: Map (processes and filters data) and Reduce (aggregates and summarizes the results).
Use Cases :
Data Warehousing: Storing and processing large volumes of data from various sources.
Analytics: Running complex queries and data analysis on large datasets.
Big Data Applications: Supporting applications in fields such as finance, healthcare, retail, and telecommunications.
Subscribe to my newsletter
Read articles from Sandeepa Bandara Thennakoon directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
Sandeepa Bandara Thennakoon
Sandeepa Bandara Thennakoon
I am a passionate Fullstack Developer and a Computer Science undergraduate at the University of Jaffna. I have gained valuable experience as a Front End Developer at CodSoft Company in India. Additionally, I have studied project management and software QA concepts from universities abroad. My skill set includes software testing and project management, complementing my fullstack development expertise.