#!/bin/bash
##################
#@author: chichuduxing
#@date: 20161011
##################
##加載Hadoop環境變量
##略過
function showhelp()
{
echo "###################################!!! you must input at least two params !!!###################################"
echo "#Usage: [sh `basename $0` origin|-o hdfspath] show info with filename's ascending order ,eg: sh `basename $0` origin /home"
echo "#Usage: [sh `basename $0` asc|-a hdfspath] show info with size's ascending order,eg: sh `basename $0` asc /home"
echo "#Usage: [sh `basename $0` desc|-d hdfspath] show info with size's descending order,eg: sh `basename $0` desc /home"
echo "###############################################################################################################"
}
#首先判斷輸入參數的個數是否符合要求
if [ $# -ne 2 ]; then
#如果傳遞進來的參數個數不等於2,則記錄錯誤並退出
showhelp
#退出
exit 1
fi
function origin()
{
hadoop fs -du $1 | awk '{if(1==NR){print $0}else{size=$1/1024;if(size<1024){printf("%10.3f KB\t%s\n",size,$2);}else{size=size/1024;if(size<1024){printf("\033[36m%10.3f MB\t%s\n\033[0m",size,$2);}else{size=size/1024;if(size<1024){printf("\033[35m%10.3f GB\t%s\n\033[0m",size,$2);}else{size=size/1024;printf("\033[31m%10.3f TB\t%s\n\033[0m",size,$2);}}}}}'
}
function order()
{
hadoop fs -du $1 | tail -n +2 | sort $2 -n | awk '{size=$1/1024;if(size<1024){printf("%10.3f KB\t%s\n",size,$2);}else{size=size/1024;if(size<1024){printf("\033[36m%10.3f MB\t%s\n\033[0m",size,$2);}else{size=size/1024;if(size<1024){printf("\033[35m%10.3f GB\t%s\n\033[0m",size,$2);}else{size=size/1024;printf("\033[31m%10.3f TB\t%s\n\033[0m",size,$2);}}}}'
}
function asc()
{
order $1
}
function desc()
{
order $1 -r
}
case "$1" in
origin|-o)
origin $2
;;
asc|-a)
asc $2
;;
desc|-d)
desc $2
;;
*)
showhelp
exit 1
esac
exit 0
腳本名稱:hdfsdu.sh
查看hdfs上根目錄下所有文件大小並降序排列
sh hdfsdu.sh / desc