本文介绍了如何使用hadoop fsck命令来检查HDFS文件的副本分布情况,通过示例展示了如何找到文件的BLK块以及它们在不同DataNode上的存储位置。
开始部署hdfs的时候,文件冗余3份。那么1个文件分拆成那些BLK,分别存储在那里呢?
hadoop fsck <需要找的文件名> -files -blocks -locations
#######################
[root@master ~]# hadoop fsck --help
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.
Usage: DFSck <path> [-list-corruptfileblocks | [-move | -delete | -openforwrite] [-files [-blocks [-locations | -racks]]]]
<path> start checking from this path
-move move corrupted files to /lost+found
-delete delete corrupted files
-files print out files being checked
-openforwrite print out files opened for write
-includeSnapshots include snapshot data if the given path indicates a snapshottable directory or there are snapshottable directories under it
-list-corruptfileblocks print out list of missing blocks and files they belong to
-blocks print out block report
-locations print out locations for every block
-racks print out network topology for data-node locations