Hi, We are also facing the same issue. /user/hive/warehouse always reaches hard quota and jobs fail. Often we reachout to users to delete old tables/db’s. Is there a good way to handle this at enterprise level ( 100’s of users and 1000’s of databases)?
On Sun, Sep 16, 2018 at 00:31 Mahender Sarangam < mahender.bigd...@outlook.com> wrote: > Hi, > > Our storage holding TB of \User folder data. it has users and their logs. > is there a way to set limit or quota and automatically clean up folder if > it exceeds beyond certain limit. > > *$ sudo -u hdfs hdfs dfsadmin -setSpaceQuota 10g /user* > > I know above command sets the limit. But is there better way to do cleanup. > > > > >