- Practical Big Data Analytics
- Nataraj Dasgupta
- 141字
- 2021-07-02 19:26:26
The core modules of Hadoop
The core modules of Hadoop consist of:
- Hadoop Common: Libraries and other common helper utilities required by Hadoop
- HDFS: A distributed, highly-available, fault-tolerant filesystem that stores data
- Hadoop MapReduce: A programming paradigm involving distributed computing across commodity servers (or nodes)
- Hadoop YARN: A framework for job scheduling and resource management
Of these core components, YARN was introduced in 2012 to address some of the shortcomings of the first release of Hadoop. The first version of Hadoop (or equivalently, the first model of Hadoop) used HDFS and MapReduce as its main components. As Hadoop gained in popularity, the need to use facilities beyond those provided by MapReduce became more and more important. This, along with some other technical considerations, led to the development of YARN.
Let's now look at the salient characteristics of Hadoop as itemized previously.
推薦閱讀
- 數(shù)據(jù)展現(xiàn)的藝術(shù)
- 亮劍.NET:.NET深入體驗與實戰(zhàn)精要
- Visualforce Development Cookbook(Second Edition)
- 我的J2EE成功之路
- 工業(yè)機器人工程應(yīng)用虛擬仿真教程:MotoSim EG-VRC
- Python Algorithmic Trading Cookbook
- 大型數(shù)據(jù)庫管理系統(tǒng)技術(shù)、應(yīng)用與實例分析:SQL Server 2005
- Visual Basic.NET程序設(shè)計
- 網(wǎng)絡(luò)布線與小型局域網(wǎng)搭建
- 深度學(xué)習(xí)與目標(biāo)檢測
- Working with Linux:Quick Hacks for the Command Line
- 筆記本電腦電路分析與故障診斷
- MPC5554/5553微處理器揭秘
- Mastering Predictive Analytics with scikit:learn and TensorFlow
- Serverless Design Patterns and Best Practices