- Hadoop 2.x Administration Cookbook
- Gurmukh Singh
- 270字
- 2021-07-09 20:10:27
Configuring HDFS block size
Getting ready
To step through the recipes in this chapter, make sure you have completed the recipes in Chapter 1, Hadoop Architecture and Deployment or at least understand the basic Hadoop cluster setup.
How to do it...
ssh
to the master node, which is Namenode, and navigate to the directory where Hadoop is installed. In the previous chapter, Hadoop was installed at/opt/cluster/hadoop
:$ ssh root@10.0.0.4
- Change to the
Hadoop
user, or any other user that is running Hadoop, by using the following:$ sudo su - hadoop
- Edit the
hdfs-site.xml
file and modify the parameter to reflect the changes, as shown in the following screenshot: dfs.blocksize
is the parameter that decides on the value of the HDFS block size. The unit is bytes and the default value is 64 MB in Hadoop 1 and 128 MB in Hadoop 2. The block size can be configured according to the need.- Once the changes are made to
hdfs-site.xml
, copy the file across all nodes in the cluster. - Then restart the Namenode and
datanode
daemons on all nodes. - The block size can be configured per file by specifying it during the copy process, as shown in the following screenshot:
How it works...
The best practice is to keep the configurations the same across all nodes in the cluster, but it is not mandatory. For example, the block size of Namenode can be different from that of the edge node. In that case, the parameters on the source node will be effective. It means that the parameter on the node from which the copying is done will be in effect.
推薦閱讀
- Big Data Analytics with Hadoop 3
- Mastering Spark for Data Science
- Verilog HDL數字系統設計入門與應用實例
- 協作機器人技術及應用
- R Machine Learning By Example
- 工業機器人入門實用教程(KUKA機器人)
- PHP開發手冊
- Python Algorithmic Trading Cookbook
- 永磁同步電動機變頻調速系統及其控制(第2版)
- 中國戰略性新興產業研究與發展·智能制造
- Windows 7寶典
- 工業機器人安裝與調試
- 分析力!專業Excel的制作與分析實用法則
- Mastering GitLab 12
- PLC與變頻技術應用