Re: [ceph-users] Ceph and NFS

2016-01-19 Thread Arthur Liu
On Tue, Jan 19, 2016 at 12:36 PM, Gregory Farnum  wrote:

> > I've found that using knfsd does not preserve cephfs directory and file
> > layouts, but using nfs-ganesha does. I'm currently using nfs-ganesha
> 2.4dev5
> > and seems stable so far.
>
> Can you expand on that? In what manner is it not preserving directory
> and file layouts?
> -Greg
>
> My mistake - I wrongly assumed that the directories did not preserve
layout - it was only subdirectories created under a directory that had a
ceph.dir.layout xattr did not have ceph.dir.layout xattr. Files created
under a subdir that does not have ceph.dir.layout still inherits the file
layout. The same applies to kernel mounted as well as via knfsd, so it does
preserve directory and file layouts. On a separate note, there's no way of
knowing what the default ceph.dir.layout is for a subdirectory unless you
traverse up the tree.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CentOS 7 iscsi gateway using lrbd

2016-01-19 Thread Nick Fisk
But interestingly enough, if you look down to where they run the targetcli ls, 
it shows a RBD backing store.

Maybe it's using the krbd driver to actually do the Ceph side of the 
communication, but lio plugs into this rather than just talking to a dumb block 
device???

This needs further investigation. Could be very interesting.

> -Original Message-
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> ??? ???
> Sent: 18 January 2016 17:15
> To: Tyler Bishop 
> Cc: Dominik Zalewski ; ceph-users  us...@lists.ceph.com>
> Subject: Re: [ceph-users] CentOS 7 iscsi gateway using lrbd
> 
> https://github.com/swiftgist/lrbd/wiki
> According to lrbd wiki it still uses KRBD (see those /dev/rbd/...
> devices in targetcli config).
> I was thinking that Mike Christie developed a librbd module for LIO.
> So what is it - KRBD or librbd?
> 
> 2016-01-18 20:23 GMT+08:00 Tyler Bishop
> :
> >
> > Well that's interesting.
> >
> > I've mounted block devices to the kernel and exported them to iscsi
> > but the performance was horrible.. I wonder if this is any different?
> >
> >
> > 
> > From: "Dominik Zalewski" 
> > To: ceph-users@lists.ceph.com
> > Sent: Monday, January 18, 2016 6:35:20 AM
> > Subject: [ceph-users] CentOS 7 iscsi gateway using lrbd
> >
> > Hi,
> > I'm looking into implementing iscsi gateway with MPIO using lrbd -
> > https://github.com/swiftgist/lrb
> >
> >
> >
> https://www.suse.com/docrep/documents/kgu61iyowz/suse_enterprise_st
> ora
> > ge_2_and_iscsi.pdf
> >
> > https://www.susecon.com/doc/2015/sessions/TUT16512.pdf
> >
> > From above examples:
> >
> > For iSCSI failover and load-balancing,
> >
> > these servers must run a kernel supporting the target_core_
> >
> > rbd module. This also requires that the target servers run at
> >
> > least the version 3.12.48-52.27.1 of the kernel-default ­package.
> >
> > Updates packages are available from the SUSE Linux
> >
> > Enterprise Server maintenance channel.
> >
> >
> > I understand that lrbd is basically a nice way to configure LIO and
> > rbd across ceph osd nodes/iscsi gatways. Does CentOS 7 have same
> > target_core_rbd module in the kernel or this is something Suse
> > Enterprise Storage specific only?
> >
> >
> > Basically will LIO+rbd work the same way on CentOS 7? Has anyone using
> > it with CentOS?
> >
> >
> > Thanks
> >
> >
> > Dominik
> >
> >
> >
> >
> >
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > ___
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CentOS 7 iscsi gateway using lrbd

2016-01-19 Thread Ilya Dryomov
On Tue, Jan 19, 2016 at 10:34 AM, Nick Fisk  wrote:
> But interestingly enough, if you look down to where they run the targetcli 
> ls, it shows a RBD backing store.
>
> Maybe it's using the krbd driver to actually do the Ceph side of the 
> communication, but lio plugs into this rather than just talking to a dumb 
> block device???

It does use krbd driver.

Thanks,

Ilya
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] CephFS

2016-01-19 Thread willi.feh...@t-online.de
 

Hi, 

 

I would like to know if CephFS will be marked as stable as soon as Ceph-1.0 
will be released.

 

Regards - Willi___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CentOS 7 iscsi gateway using lrbd

2016-01-19 Thread Василий Ангапов
So is it a different approach that was used here by Mike Christie:
http://www.spinics.net/lists/target-devel/msg10330.html ?
It seems to be a confusion because it also implements target_core_rbd
module. Or not?

2016-01-19 18:01 GMT+08:00 Ilya Dryomov :
> On Tue, Jan 19, 2016 at 10:34 AM, Nick Fisk  wrote:
>> But interestingly enough, if you look down to where they run the targetcli 
>> ls, it shows a RBD backing store.
>>
>> Maybe it's using the krbd driver to actually do the Ceph side of the 
>> communication, but lio plugs into this rather than just talking to a dumb 
>> block device???
>
> It does use krbd driver.
>
> Thanks,
>
> Ilya
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Again - state of Ceph NVMe and SSDs

2016-01-19 Thread Tyler Bishop
It sounds like your just assuming these drives don't perform good...

- Original Message -
From: "Mark Nelson" 
To: ceph-users@lists.ceph.com
Sent: Monday, January 18, 2016 2:17:19 PM
Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs

Take Greg's comments to heart, because he's absolutely correct here. 
Distributed storage systems almost as a rule love parallelism and if you 
have enough you can often hide other issues.  Latency is probably the 
more interesting question, and frankly that's where you'll often start 
seeing the kernel, ceph code, drivers, random acts of god, etc, get in 
the way.  It's very easy for any one of these things to destroy your 
performance, so you have to be *very* *very* careful to understand 
exactly what you are seeing.  As such, don't trust any one benchmark. 
Wait until it's independently verified, possibly by multiple sources, 
before putting too much weight into it.

Mark

On 01/18/2016 01:02 PM, Tyler Bishop wrote:
> One of the other guys on the list here benchmarked them.  They spanked every 
> other ssd on the *recommended* tree..
>
> - Original Message -
> From: "Gregory Farnum" 
> To: "Tyler Bishop" 
> Cc: "David" , "Ceph Users" 
> Sent: Monday, January 18, 2016 2:01:44 PM
> Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs
>
> On Sun, Jan 17, 2016 at 12:34 PM, Tyler Bishop
>  wrote:
>> The changes you are looking for are coming from Sandisk in the ceph "Jewel" 
>> release coming up.
>>
>> Based on benchmarks and testing, sandisk has really contributed heavily on 
>> the tuning aspects and are promising 90%+ native iop of a drive in the 
>> cluster.
>
> Mmmm, they've gotten some very impressive numbers but most people
> shouldn't be expecting 90% of an SSD's throughput out of their
> workloads. These tests are *very* parallel and tend to run multiple
> OSD processes on a single SSD, IIRC.
> -Greg
>
>>
>> The biggest changes will come from the memory allocation with writes.  
>> Latency is going to be a lot lower.
>>
>>
>> - Original Message -
>> From: "David" 
>> To: "Wido den Hollander" 
>> Cc: ceph-users@lists.ceph.com
>> Sent: Sunday, January 17, 2016 6:49:25 AM
>> Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs
>>
>> Thanks Wido, those are good pointers indeed :)
>> So we just have to make sure the backend storage (SSD/NVMe journals) won’t 
>> be saturated (or the controllers) and then go with as many RBD per VM as 
>> possible.
>>
>> Kind Regards,
>> David Majchrzak
>>
>> 16 jan 2016 kl. 22:26 skrev Wido den Hollander :
>>
>>> On 01/16/2016 07:06 PM, David wrote:
 Hi!

 We’re planning our third ceph cluster and been trying to find how to
 maximize IOPS on this one.

 Our needs:
 * Pool for MySQL, rbd (mounted as /var/lib/mysql or equivalent on KVM
 servers)
 * Pool for storage of many small files, rbd (probably dovecot maildir
 and dovecot index etc)

>>>
>>> Not completely NVMe related, but in this case, make sure you use
>>> multiple disks.
>>>
>>> For MySQL for example:
>>>
>>> - Root disk for OS
>>> - Disk for /var/lib/mysql (data)
>>> - Disk for /var/log/mysql (binary log)
>>> - Maybe even a InnoDB logfile disk
>>>
>>> With RBD you gain more performance by sending I/O into the cluster in
>>> parallel. So when ever you can, do so!
>>>
>>> Regarding small files, it might be interesting to play with the stripe
>>> count and stripe size there. By default this is 1 and 4MB. But maybe 16
>>> and 256k work better here.
>>>
>>> With Dovecot as well, use a different RBD disk for the indexes and a
>>> different one for the Maildir itself.
>>>
>>> Ceph excels at parallel performance. That is what you want to aim for.
>>>
 So I’ve been reading up on:

 https://communities.intel.com/community/itpeernetwork/blog/2015/11/20/the-future-ssd-is-here-pcienvme-boosts-ceph-performance

 and ceph-users from october 2015:

 http://www.spinics.net/lists/ceph-users/msg22494.html

 We’re planning something like 5 OSD servers, with:

 * 4x 1.2TB Intel S3510
 * 8st 4TB HDD
 * 2x Intel P3700 Series HHHL PCIe 400GB (one for SSD Pool Journal and
 one for HDD pool journal)
 * 2x 80GB Intel S3510 raid1 for system
 * 256GB RAM
 * 2x 8 core CPU Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz or better

 This cluster will probably run Hammer LTS unless there are huge
 improvements in Infernalis when dealing 4k IOPS.

 The first link above hints at awesome performance. The second one from
 the list not so much yet..

 Is anyone running Hammer or Infernalis with a setup like this?
 Is it a sane setup?
 Will we become CPU constrained or can we just throw more RAM on it? :D

 Kind Regards,
 David Majchrzak


 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-cep

Re: [ceph-users] Again - state of Ceph NVMe and SSDs

2016-01-19 Thread Mark Nelson
Not at all!  For all we know, the drives may be the fastest ones on the 
planet.  My comment still stands though.  Be skeptical of any one 
benchmark that shows something unexpected.  90% of native SSD/NVMe IOPS 
performance in a distributed storage system is just such a number.  Look 
for test repeatability.  Look for independent verification.  The memory 
allocator tests that Sandisk initially performed and that were later 
independently verified by Intel, Red Hat, and others is a great example 
of this kind of process.


Mark

On 01/19/2016 07:01 AM, Tyler Bishop wrote:

It sounds like your just assuming these drives don't perform good...

- Original Message -
From: "Mark Nelson" 
To: ceph-users@lists.ceph.com
Sent: Monday, January 18, 2016 2:17:19 PM
Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs

Take Greg's comments to heart, because he's absolutely correct here.
Distributed storage systems almost as a rule love parallelism and if you
have enough you can often hide other issues.  Latency is probably the
more interesting question, and frankly that's where you'll often start
seeing the kernel, ceph code, drivers, random acts of god, etc, get in
the way.  It's very easy for any one of these things to destroy your
performance, so you have to be *very* *very* careful to understand
exactly what you are seeing.  As such, don't trust any one benchmark.
Wait until it's independently verified, possibly by multiple sources,
before putting too much weight into it.

Mark

On 01/18/2016 01:02 PM, Tyler Bishop wrote:

One of the other guys on the list here benchmarked them.  They spanked every 
other ssd on the *recommended* tree..

- Original Message -
From: "Gregory Farnum" 
To: "Tyler Bishop" 
Cc: "David" , "Ceph Users" 
Sent: Monday, January 18, 2016 2:01:44 PM
Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs

On Sun, Jan 17, 2016 at 12:34 PM, Tyler Bishop
 wrote:

The changes you are looking for are coming from Sandisk in the ceph "Jewel" 
release coming up.

Based on benchmarks and testing, sandisk has really contributed heavily on the 
tuning aspects and are promising 90%+ native iop of a drive in the cluster.


Mmmm, they've gotten some very impressive numbers but most people
shouldn't be expecting 90% of an SSD's throughput out of their
workloads. These tests are *very* parallel and tend to run multiple
OSD processes on a single SSD, IIRC.
-Greg



The biggest changes will come from the memory allocation with writes.  Latency 
is going to be a lot lower.


- Original Message -
From: "David" 
To: "Wido den Hollander" 
Cc: ceph-users@lists.ceph.com
Sent: Sunday, January 17, 2016 6:49:25 AM
Subject: Re: [ceph-users] Again - state of Ceph NVMe and SSDs

Thanks Wido, those are good pointers indeed :)
So we just have to make sure the backend storage (SSD/NVMe journals) won’t be 
saturated (or the controllers) and then go with as many RBD per VM as possible.

Kind Regards,
David Majchrzak

16 jan 2016 kl. 22:26 skrev Wido den Hollander :


On 01/16/2016 07:06 PM, David wrote:

Hi!

We’re planning our third ceph cluster and been trying to find how to
maximize IOPS on this one.

Our needs:
* Pool for MySQL, rbd (mounted as /var/lib/mysql or equivalent on KVM
servers)
* Pool for storage of many small files, rbd (probably dovecot maildir
and dovecot index etc)



Not completely NVMe related, but in this case, make sure you use
multiple disks.

For MySQL for example:

- Root disk for OS
- Disk for /var/lib/mysql (data)
- Disk for /var/log/mysql (binary log)
- Maybe even a InnoDB logfile disk

With RBD you gain more performance by sending I/O into the cluster in
parallel. So when ever you can, do so!

Regarding small files, it might be interesting to play with the stripe
count and stripe size there. By default this is 1 and 4MB. But maybe 16
and 256k work better here.

With Dovecot as well, use a different RBD disk for the indexes and a
different one for the Maildir itself.

Ceph excels at parallel performance. That is what you want to aim for.


So I’ve been reading up on:

https://communities.intel.com/community/itpeernetwork/blog/2015/11/20/the-future-ssd-is-here-pcienvme-boosts-ceph-performance

and ceph-users from october 2015:

http://www.spinics.net/lists/ceph-users/msg22494.html

We’re planning something like 5 OSD servers, with:

* 4x 1.2TB Intel S3510
* 8st 4TB HDD
* 2x Intel P3700 Series HHHL PCIe 400GB (one for SSD Pool Journal and
one for HDD pool journal)
* 2x 80GB Intel S3510 raid1 for system
* 256GB RAM
* 2x 8 core CPU Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz or better

This cluster will probably run Hammer LTS unless there are huge
improvements in Infernalis when dealing 4k IOPS.

The first link above hints at awesome performance. The second one from
the list not so much yet..

Is anyone running Hammer or Infernalis with a setup like this?
Is it a sane setup?
Will we become CPU constrained or can we just throw more RAM on it? :D

Kind Regards,
David 

Re: [ceph-users] RGW -- 404 on keys in bucket.list() thousands of multipart ids listed as well.

2016-01-19 Thread seapasu...@uchicago.edu
I was wondering if there is anything else I could provide outside of the 
radosgw logs during the file upload? At this point I am uploading 7mb 
files repeatedly to try to reproduce this error but so far I do not have 
any missing keys in my test bucket. I can't see this happening if we 
were to just use multipart uploads for files larger than say 1Gib but 
then who is to say that we will not have corrupt multipart uploads.


On 1/15/16 4:36 PM, Yehuda Sadeh-Weinraub wrote:

The head object of a multipart object has 0 size, so it's expected.
What's missing is the tail of the object. I don't assume you have any
logs from when the object was uploaded?

Yehuda

On Fri, Jan 15, 2016 at 2:12 PM, seapasu...@uchicago.edu
 wrote:

Sorry for the confusion::

When I grepped for the prefix of the missing object::
"2015\/01\/01\/PAKC\/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar.2~pcu5Hz6foFXjlSxBat22D8YMcHlQOBD"

I am not able to find any chunks of the object::

lacadmin@kh28-10:~$ rados -p .rgw.buckets ls | grep 'pcu5Hz6'
lacadmin@kh28-10:~$

The only piece of the object that I can seem to find is the original one I
posted::
lacadmin@kh28-10:~$ rados -p .rgw.buckets ls | grep
'NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959'
default.384153.1_2015/01/01/PAKC/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar

And when we stat this object is is 0 bytes as shown earlier::
lacadmin@kh28-10:~$ rados -p .rgw.buckets stat
'default.384153.1_2015/01/01/PAKC/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar'
.rgw.buckets/default.384153.1_2015/01/01/PAKC/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar
mtime 2015-11-04 15:29:30.00, size 0

Sorry again for the confusion.



On 1/15/16 3:58 PM, Yehuda Sadeh-Weinraub wrote:

Ah, I see. Misread that and the object names were very similar. No,
don't copy it. You can try to grep for the specific object name and
see if there are pieces of it lying around under a different upload
id.

Yehuda

On Fri, Jan 15, 2016 at 1:44 PM, seapasu...@uchicago.edu
 wrote:

Sorry I am a bit confused. The successful list that I provided is from a
different object of the same size to show that I could indeed get a list.
Are you saying to copy the working object to the missing object? Sorry
for
the confusion.


On 1/15/16 3:20 PM, Yehuda Sadeh-Weinraub wrote:

That's interesting, and might point at the underlying issue that
caused it. Could be a racing upload that somehow ended up with the
wrong object head. The 'multipart' object should be 4M in size, and
the 'shadow' one should have the remainder of the data. You can run
'rados stat -p .rgw.buckets ' to validate that. If that's the
case, you can copy these to the expected object names:

$ src_uploadid=wksHvto9gRgHUJbhm_TZPXJTZUPXLT2
$ dest_uploadid=pcu5Hz6foFXjlSxBat22D8YMcHlQOBD

$ rados -p .rgw.buckets cp


default.384153.1__multipart_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~${src_uploadid}.1


default.384153.1__multipart_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~${dest_uploadid}.1

$ rados -p .rgw.buckets cp


default.384153.1__shadow_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~${src_upload_id}.1_1


default.384153.1__shadow_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~${dest_upload_id}.1_1

Yehuda


On Fri, Jan 15, 2016 at 1:02 PM, seapasu...@uchicago.edu
 wrote:

lacadmin@kh28-10:~$ rados -p .rgw.buckets ls | grep 'pcu5Hz6'
lacadmin@kh28-10:~$

Nothing was found. That said when I run the command with another prefix
snippet::
lacadmin@kh28-10:~$ rados -p .rgw.buckets ls | grep 'wksHvto'


default.384153.1__shadow_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~wksHvto9gRgHUJbhm_TZPXJTZUPXLT2.1_1


default.384153.1__multipart_2015/01/01/KABR/NWS_NEXRAD_NXL2DP_KABR_2015010113_20150101135959.tar.2~wksHvto9gRgHUJbhm_TZPXJTZUPXLT2.1




On 1/15/16 12:05 PM, Yehuda Sadeh-Weinraub wrote:

On Fri, Jan 15, 2016 at 9:36 AM, seapasu...@uchicago.edu
 wrote:

Hello Yehuda,

Here it is::

radosgw-admin object stat --bucket="noaa-nexrad-l2"



--object="2015/01/01/PAKC/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar"
{
"name":



"2015\/01\/01\/PAKC\/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar",
"size": 7147520,
"policy": {
"acl": {
"acl_user_map": [
{
"user": "b05f707271774dbd89674a0736c9406e",
"acl": 15
}
],
"acl_group_map": [
{
"group": 1,
"acl": 1
}
],
"grant_map": [
{
"id": "",
"grant": {
"type": {
"type": 2
   

Re: [ceph-users] CentOS 7 iscsi gateway using lrbd

2016-01-19 Thread Mike Christie
Everyone is right - sort of :)

It is that target_core_rbd module that I made that was rejected
upstream, along with modifications from SUSE which added persistent
reservations support. I also made some modifications to rbd so
target_core_rbd and krbd could share code. target_core_rbd uses rbd like
a lib. And it is also modifications to the targetcli related tool and
libs, so you can use them to control the new rbd backend. SUSE's lrbd
then handles setup/management of across multiple targets/gatways.

I was going to modify targetcli more and have the user just pass in the
rbd info there, but did not get finished. That is why in that suse stuff
you still make the krbd device like normal. You then pass that to the
target_core_rbd module with targetcli and that is how that module knows
about the rbd device.

The target_core_rbd module was rejected upstream, so I stopped
development and am working on the approach suggested by those reviewers
which instead of going from lio->target_core_rbd->krbd goes
lio->target_core_iblock->linux block layer->krbd. With this approach you
just use the normal old iblock driver and krbd and then I am modifying
them to just work and do the right thing.


On 01/19/2016 05:45 AM, Василий Ангапов wrote:
> So is it a different approach that was used here by Mike Christie:
> http://www.spinics.net/lists/target-devel/msg10330.html ?
> It seems to be a confusion because it also implements target_core_rbd
> module. Or not?
> 
> 2016-01-19 18:01 GMT+08:00 Ilya Dryomov :
>> On Tue, Jan 19, 2016 at 10:34 AM, Nick Fisk  wrote:
>>> But interestingly enough, if you look down to where they run the targetcli 
>>> ls, it shows a RBD backing store.
>>>
>>> Maybe it's using the krbd driver to actually do the Ceph side of the 
>>> communication, but lio plugs into this rather than just talking to a dumb 
>>> block device???
>>
>> It does use krbd driver.
>>
>> Thanks,
>>
>> Ilya

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] RGW -- 404 on keys in bucket.list() thousands of multipart ids listed as well.

2016-01-19 Thread Yehuda Sadeh-Weinraub
On Fri, Jan 15, 2016 at 5:04 PM, seapasu...@uchicago.edu
 wrote:
> I have looked all over and I do not see any explicit mention of
> "NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959" in the logs nor do I
> see a timestamp from November 4th although I do see log rotations dating
> back to october 15th. I don't think it's possible it wasn't logged so I am
> going through the bucket logs from the 'radosgw-admin log show --object'
> side and I found the following::
>
> 4604932 {
> 4604933 "bucket": "noaa-nexrad-l2",
> 4604934 "time": "2015-11-04 21:29:27.346509Z",
> 4604935 "time_local": "2015-11-04 15:29:27.346509",
> 4604936 "remote_addr": "",
> 4604937 "object_owner": "b05f707271774dbd89674a0736c9406e",
> 4604938 "user": "b05f707271774dbd89674a0736c9406e",
> 4604939 "operation": "PUT",

I'd expect a multipart upload completion to be done with a POST, not a PUT.

> 4604940 "uri":
> "\/noaa-nexrad-l2\/2015\/01\/01\/PAKC\/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar",
> 4604941 "http_status": "200",
> 4604942 "error_code": "",
> 4604943 "bytes_sent": 19,
> 4604944 "bytes_received": 0,
> 4604945 "object_size": 0,

Do you see a zero object_size for other multipart uploads?

Yehuda

> 4604946 "total_time": 142640400,
> 4604947 "user_agent": "Boto\/2.38.0 Python\/2.7.7
> Linux\/2.6.32-573.7.1.el6.x86_64",
> 4604948 "referrer": ""
> 4604949 }
>
> Does this help at all. The total time seems exceptionally high. Would it be
> possible that there is a timeout issue where the put request started a
> multipart upload with the correct header and then timed out but the radosgw
> took the data anyway?
>
> I am surprised the radosgw returned a 200 let alone placed the key in the
> bucket listing.
>
>
> That said here is another object (different object) that 404s:
> 1650873 {
> 1650874 "bucket": "noaa-nexrad-l2",
> 1650875 "time": "2015-11-05 04:50:42.606838Z",
> 1650876 "time_local": "2015-11-04 22:50:42.606838",
> 1650877 "remote_addr": "",
> 1650878 "object_owner": "b05f707271774dbd89674a0736c9406e",
> 1650879 "user": "b05f707271774dbd89674a0736c9406e",
> 1650880 "operation": "PUT",
> 1650881 "uri":
> "\/noaa-nexrad-l2\/2015\/02\/25\/KVBX\/NWS_NEXRAD_NXL2DP_KVBX_2015022516_20150225165959.tar",
> 1650882 "http_status": "200",
> 1650883 "error_code": "",
> 1650884 "bytes_sent": 19,
> 1650885 "bytes_received": 0,
> 1650886 "object_size": 0,
> 1650887 "total_time": 0,
> 1650888 "user_agent": "Boto\/2.38.0 Python\/2.7.7
> Linux\/2.6.32-573.7.1.el6.x86_64",
> 1650889 "referrer": ""
> 1650890 }
>
> And this one fails with a 404 as well. Does this help at all? Here is a
> successful object (different object) log entry as well just in case::
>
> 17462367 {
> 17462368 "bucket": "noaa-nexrad-l2",
> 17462369 "time": "2015-11-04 21:16:44.148603Z",
> 17462370 "time_local": "2015-11-04 15:16:44.148603",
> 17462371 "remote_addr": "",
> 17462372 "object_owner": "b05f707271774dbd89674a0736c9406e",
> 17462373 "user": "b05f707271774dbd89674a0736c9406e",
> 17462374 "operation": "PUT",
> 17462375 "uri":
> "\/noaa-nexrad-l2\/2015\/01\/01\/KAKQ\/NWS_NEXRAD_NXL2DP_KAKQ_2015010108_20150101085959.tar",
> 17462376 "http_status": "200",
> 17462377 "error_code": "",
> 17462378 "bytes_sent": 19,
> 17462379 "bytes_received": 0,
> 17462380 "object_size": 0,
> 17462381 "total_time": 0,
> 17462382 "user_agent": "Boto\/2.38.0 Python\/2.7.7
> Linux\/2.6.32-573.7.1.el6.x86_64",
> 17462383 "referrer": ""
> 17462384 }
>
> So I am guessing these are not pertinent as they look nearly identical.
> Unfortunately I do not have any client.radosgw.logs to show for the failed
> files for some reason. Is there anything else I can do to troubleshoot this
> issue. In the end the radosgw should have never list these files as they
> never completed successfully, right?
>
>
>
>
>
>
> On 1/15/16 4:36 PM, Yehuda Sadeh-Weinraub wrote:
>>
>> The head object of a multipart object has 0 size, so it's expected.
>> What's missing is the tail of the object. I don't assume you have any
>> logs from when the object was uploaded?
>>
>> Yehuda
>>
>> On Fri, Jan 15, 2016 at 2:12 PM, seapasu...@uchicago.edu
>>  wrote:
>>>
>>> Sorry for the confusion::
>>>
>>> When I grepped for the prefix of the missing object::
>>>
>>> "2015\/01\/01\/PAKC\/NWS_NEXRAD_NXL2DP_PAKC_2015010111_20150101115959.tar.2~pcu5Hz6foFXjlSxBat22D8YMcHlQOBD"
>>>
>>> I am not a

[ceph-users] Repository with some internal utils

2016-01-19 Thread Lionel Bouton
Hi,

someone asked me if he could get access to the BTRFS defragmenter we
used for our Ceph OSDs. I took a few minutes to put together a small
github repository with :
- the defragmenter I've been asked about (tested on 7200 rpm drives and
designed to put low IO load on them),
- the scrub scheduler we use to avoid load spikes on Firefly,
- some basic documentation (this is still rough around the edges so you
better like to read Ruby code if you want to peak at most of the logic,
tune or hack these).

Here it is: https://github.com/jtek/ceph-utils

This is running in production for several months now and I didn't touch
the code or the numerous internal tunables these scripts have for
several weeks so it probably won't destroy your clusters. These scripts
come without warranties though.

Best regards,

Lionel
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] s3 upload to ceph slow after few chunks

2016-01-19 Thread Rishiraj Rana
Hey guys, I am having an s3 upload to ceph issue where in the upload seems to 
crawl after the first few chunks for a multipart upload. The test file is 38M 
in size and the upload was tried with s3 default chunk size at 15M and then 
tried again with chunk size set to 5M and then was tested again with multipart 
disabled. Each time the upload halted at seeming random amount and I cannot 
seem to figure out why.


[10.0.144.2] $ /opt/utilities/s3cmd/s3cmd put ACCOUNT\ 
EXTRACT-GENERATED.PDF.test.2 's3://'
 -> s3:///  [part 1 of 3, 15MB]
 15728640 of 15728640   100% in2s 7.43 MB/s  done
 -> s3:///  [part 2 of 3, 15MB]
 4096 of 15728640 0% in0s31.68 kB/s
ERROR: Upload of 'ACCOUNT EXTRACT-GENERATED.PDF.test.2' part 2 failed.
Use
  / 2~T2OI6bU-TYE_dOtm3tPIndtYTek-V0r
to abort the upload, or
 ./s3cmd --upload-id 2~T2OI6bU-TYE_dOtm3tPIndtYTek-V0r put ...
to continue the upload.
See ya!


Rishiraj Rana

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] How to observed civetweb.

2016-01-19 Thread Ben Hines
Hey Kobi,

You stated:

> >> You can add:
> >> *access_log_file=/var/log/civetweb/access.log
> >> error_log_file=/var/log/civetweb/error.log*
> >>
> >> to *rgw frontends* in ceph.conf though these logs are thin on info
> >> (Source IP, date, and request)



How is this done exactly in the config file? I've tried various ways,
nothing works.


ie, neither of these work:


[client.radosgw.]
host = 
rgw socket path = /tmp/radosgw.sock
keyring = /etc/ceph/ceph.client.radosgw.keyring
log file = /var/log/radosgw/client.radosgw..log   <--  this
log gets output only
access_log_file=/var/log/civetweb/access.log

rgw frontends error log file = /var/log/radosgw/civetweb.error.log  <-- nada
rgw frontends access log file = /var/log/radosgw/civetweb.access.log <-- nada

rgw print continue = True
rgw enable ops log = False


[rgw frontends]

access_log_file=/var/log/civetweb/access.log  <-- nada



-Ben


On Tue, Sep 8, 2015 at 11:21 AM, Kobi Laredo 
wrote:

> Vickie,
>
> You can add:
> *access_log_file=/var/log/civetweb/access.log
> error_log_file=/var/log/civetweb/error.log*
>
> to *rgw frontends* in ceph.conf though these logs are thin on info
> (Source IP, date, and request)
>
> Check out
> https://github.com/civetweb/civetweb/blob/master/docs/UserManual.md for
> more civetweb configs you can inject through  *rgw frontends* config
> attribute in ceph.conf
>
> We are currently testing tuning civetweb's num_threads
> and request_timeout_ms to improve radosgw performance
>
> *Kobi Laredo*
> *Cloud Systems Engineer* | (*408) 409-KOBI*
>
> On Tue, Sep 8, 2015 at 8:20 AM, Yehuda Sadeh-Weinraub 
> wrote:
>
>> You can increase the civetweb logs by adding 'debug civetweb = 10' in
>> your ceph.conf. The output will go into the rgw logs.
>>
>> Yehuda
>>
>> On Tue, Sep 8, 2015 at 2:24 AM, Vickie ch  wrote:
>> > Dear cephers,
>> >Just upgrade radosgw from apache to civetweb.
>> > It's really simple to installed and used. But I can't find any
>> parameters or
>> > logs to adjust(or observe) civetweb. (Like apache log).  I'm really
>> confuse.
>> > Any ideas?
>> >
>> >
>> > Best wishes,
>> > Mika
>> >
>> >
>> > ___
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] How to observed civetweb.

2016-01-19 Thread Ben Hines
Of course, i figured this out. You meant just append it to the frontends
setting. Very confusing as it's unlike every other ceph setting.

rgw frontends = civetweb num_threads=150
error_log_file=/var/log/radosgw/civetweb.error.log
access_log_file=/var/log/radosgw/civetweb.access.log

Any documentation, at all, on civetweb + radosgw on the Ceph site would be
awesome.. Currently it all only references Apache+FastCGi.



On Tue, Jan 19, 2016 at 8:42 PM, Ben Hines  wrote:

> Hey Kobi,
>
> You stated:
>
> > >> You can add:
> > >> *access_log_file=/var/log/civetweb/access.log
> > >> error_log_file=/var/log/civetweb/error.log*
> > >>
> > >> to *rgw frontends* in ceph.conf though these logs are thin on info
> > >> (Source IP, date, and request)
>
>
>
> How is this done exactly in the config file? I've tried various ways, nothing 
> works.
>
>
> ie, neither of these work:
>
>
> [client.radosgw.]
> host = 
> rgw socket path = /tmp/radosgw.sock
> keyring = /etc/ceph/ceph.client.radosgw.keyring
> log file = /var/log/radosgw/client.radosgw..log   <--  this log gets 
> output only
> access_log_file=/var/log/civetweb/access.log
>
> rgw frontends error log file = /var/log/radosgw/civetweb.error.log  <-- nada
> rgw frontends access log file = /var/log/radosgw/civetweb.access.log <-- nada
>
> rgw print continue = True
> rgw enable ops log = False
>
>
> [rgw frontends]
>
> access_log_file=/var/log/civetweb/access.log  <-- nada
>
>
>
> -Ben
>
>
> On Tue, Sep 8, 2015 at 11:21 AM, Kobi Laredo 
> wrote:
>
>> Vickie,
>>
>> You can add:
>> *access_log_file=/var/log/civetweb/access.log
>> error_log_file=/var/log/civetweb/error.log*
>>
>> to *rgw frontends* in ceph.conf though these logs are thin on info
>> (Source IP, date, and request)
>>
>> Check out
>> https://github.com/civetweb/civetweb/blob/master/docs/UserManual.md for
>> more civetweb configs you can inject through  *rgw frontends* config
>> attribute in ceph.conf
>>
>> We are currently testing tuning civetweb's num_threads
>> and request_timeout_ms to improve radosgw performance
>>
>> *Kobi Laredo*
>> *Cloud Systems Engineer* | (*408) 409-KOBI*
>>
>> On Tue, Sep 8, 2015 at 8:20 AM, Yehuda Sadeh-Weinraub 
>> wrote:
>>
>>> You can increase the civetweb logs by adding 'debug civetweb = 10' in
>>> your ceph.conf. The output will go into the rgw logs.
>>>
>>> Yehuda
>>>
>>> On Tue, Sep 8, 2015 at 2:24 AM, Vickie ch 
>>> wrote:
>>> > Dear cephers,
>>> >Just upgrade radosgw from apache to civetweb.
>>> > It's really simple to installed and used. But I can't find any
>>> parameters or
>>> > logs to adjust(or observe) civetweb. (Like apache log).  I'm really
>>> confuse.
>>> > Any ideas?
>>> >
>>> >
>>> > Best wishes,
>>> > Mika
>>> >
>>> >
>>> > ___
>>> > ceph-users mailing list
>>> > ceph-users@lists.ceph.com
>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >
>>> ___
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com