官术网_书友最值得收藏!

Configuring HDFS block size

Getting ready

To step through the recipes in this chapter, make sure you have completed the recipes in Chapter 1, Hadoop Architecture and Deployment or at least understand the basic Hadoop cluster setup.

How to do it...

  1. ssh to the master node, which is Namenode, and navigate to the directory where Hadoop is installed. In the previous chapter, Hadoop was installed at /opt/cluster/hadoop:
    $ ssh root@10.0.0.4
    
  2. Change to the Hadoop user, or any other user that is running Hadoop, by using the following:
    $ sudo su - hadoop
    
  3. Edit the hdfs-site.xml file and modify the parameter to reflect the changes, as shown in the following screenshot:
    How to do it...
  4. dfs.blocksize is the parameter that decides on the value of the HDFS block size. The unit is bytes and the default value is 64 MB in Hadoop 1 and 128 MB in Hadoop 2. The block size can be configured according to the need.
  5. Once the changes are made to hdfs-site.xml, copy the file across all nodes in the cluster.
  6. Then restart the Namenode and datanode daemons on all nodes.
  7. The block size can be configured per file by specifying it during the copy process, as shown in the following screenshot:
    How to do it...

How it works...

The best practice is to keep the configurations the same across all nodes in the cluster, but it is not mandatory. For example, the block size of Namenode can be different from that of the edge node. In that case, the parameters on the source node will be effective. It means that the parameter on the node from which the copying is done will be in effect.

主站蜘蛛池模板: 沂南县| 鄂托克旗| 化州市| 东兴市| 怀远县| 长武县| 金华市| 射洪县| 北京市| 澜沧| 沈阳市| 高淳县| 明光市| 墨脱县| 大理市| 孟村| 大兴区| 临潭县| 云林县| 化德县| 贡山| 益阳市| 齐齐哈尔市| 嘉祥县| 政和县| 克山县| 宁河县| 武穴市| 色达县| 海宁市| 达尔| 瑞金市| 青铜峡市| 景洪市| 读书| 英德市| 巍山| 安新县| 宁阳县| 九龙坡区| 长宁区|