Hi,

This seems pretty quick here on a jewel cluster here, But I guess the key
questions is how large is large? Is it perhaps a large number of smaller
files that is slowing this down? Is the bucket index shared / on SSD?

====

[root@korn ~]# time s3cmd du s3://seanbackup
1656225129419 29 objects s3://seanbackup/

real    0m0.314s
user    0m0.088s
sys     0m0.019s
[root@korn ~]#


On Thu, Jul 28, 2016 at 4:49 PM, Dan van der Ster <d...@vanderster.com>
wrote:

> On Thu, Jul 28, 2016 at 5:33 PM, Abhishek Lekshmanan <abhis...@suse.com>
> wrote:
> >
> > Dan van der Ster writes:
> >
> >> Hi,
> >>
> >> Does anyone know a fast way for S3 users to query their total bucket
> >> usage? 's3cmd du' takes a long time on large buckets (is it iterating
> >> over all the objects?). 'radosgw-admin bucket stats' seems to know the
> >> bucket usage immediately, but I didn't find a way to expose that to
> >> end users.
> >>
> >> Hoping this is an easy one for someone...
> >
> > If swift api is enabled swift stat on the user account might
> > probably a quicker way.
>
> This user wants to be S3-only, due to their app being compatible with
> the big commercial cloud provider.
>
> Maybe s3cmd du is slow because the cluster is running hammer -- can
> any jewel users confirm it's still slow for large buckets on jewel?
>
> Cheers, Dan
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to