一.项目经验:
1.hdfs设置多目录
why:当hdfs存储空间紧张,需要对datanode节点进行磁盘扩展
what:(1)在 hdfs-site.xml 文件中配置多目录,注意新挂载磁盘的访问权限问题
<property> <name>dfs.datanode.data.dir</name> <value> file:///${hadoop.tmp.dir}/dfs/data1, file:///hd2/dfs/ data2, file:///hd3/dfs/data3, file:///hd4/dfs/data4 </value> </property>
(2)增加磁盘后,保证每个目录数据均衡
<property> <name>io.compression.codecs</name> <value> org.apache.hadoop.io.compress.GzipCodec, org.apache.hadoop.io.compress.DefaultCodec,
org.apache.hadoop.io.compress.BZip2Codec, org.apache.hadoop.io.compress.SnappyCodec,
com.hadoop.compression.lzo.LzoCodec, com.hadoop.compression.lzo.LzopCodec </value>
</property> <property>
<name>io.compression.codec.lzo.class</name>
<value>com.hadoop.compression.lzo.LzoCodec</value>
</property>
hadoop jar hadoop-2.7.2/share/hadoop/mapreduce/hadoop-mapreduce-client-j obclient-2.7.2-tests.jar TestDFSIO -write -nrFiles 10 -fileSize 128MB
(2)测试读性能 测试内容:读取 HDFS 集群 10 个 128M 的文件
hadoop jar hadoop-2.7.2/share/hadoop/mapreduce/hadoop-mapreduce-client-j obclient-2.7.2-tests.jar TestDFSIO -read -nrFiles 10 -fileSize 128MB
hadoop jar hadoop-2.7.2/share/hadoop/mapreduce/hadoop-mapredu ce-client-jobclient-2.7.2-tests.jar TestDFSIO -clean
原文:https://www.cnblogs.com/ldy233/p/13377416.html