#/bin/bash
today=`date +%Y%m%d`
std_today="${today:0:4}-${today:4:2}-${today:6:2} 00:00:00"
today_timestamp=`date -d "${std_today}" +%s`
echo $today_timestamp
hadoop fs -ls $1 >file
cat file|while read quanxian temp user group size day hour filepath
do
current_file_timestamp=`date -d "${day} 00:00:00" +%s`
if [ $(($today_timestamp-$current_file_timestamp)) -ge $((3600*24*1)) ] ; then
logsize=`hadoop fs -dus -h $filepath 2>/dev/null`
echo "$logsize ${day} ${hour}"
hadoop fs -rmr $filepath >/dev/null 2>&1 &
fi
done
Hadoop运维-删除目录下过期的数据
最新推荐文章于 2024-12-10 09:23:56 发布