[ceph-users] Erasure coded calculation

2021-02-25 Thread Simon Sutter
Hello everyone!

I'm trying to calculate the theoretical usable storage of a ceph cluster with 
erasure coded pools.

I have 8 nodes and the profile for all data pools will be k=6 m=2.
If every node has 6 x 1TB wouldn't the calculation be like this:
RAW capacity: 8Nodes x 6Disks x 1TB = 48TB
Loss to m=2: 48TB / 8Nodes x 2m = 12TB
EC capacity: 48TB - 12TB = 36TB

At the moment I have one cluster with 8 nodes and different disks than the 
sample (but every node has the same amount of disks and the same sized disks).
The output of ceph df detail is:
--- RAW STORAGE ---
CLASS  SIZE AVAILUSED RAW USED  %RAW USED
hdd109 TiB  103 TiB  5.8 TiB   5.9 TiB   5.41
TOTAL  109 TiB  103 TiB  5.8 TiB   5.9 TiB   5.41

--- POOLS ---
POOL   ID  PGS  STORED   OBJECTS  %USED  MAX AVAIL
device_health_metrics   11   51 MiB   48  0 30 TiB
rep_data_fs 2   32   14 KiB3.41k  0 30 TiB
rep_meta_fs 3   32  227 MiB1.72k  0 30 TiB
ec_bkp14   32  4.2 TiB1.10M   6.11 67 TiB

So ec_bkp1 uses 4.2TiB an there are 67TiB free usable Storage.
This means total EC usable storage would be 71.2TiB.
But calculating with the 109TiB RAW storage, shouldn't it be  81.75?
Are the 10TiB just some overhead (that would be much overhead) or is the 
calculation not correct?

And what If I want to expand the cluster in the first sample above by three 
nodes with 6 x 2TB, which means not the same sized disks as the others.
Will the calculation with the same EC profile still be the same?
RAW capacity: 8Nodes x 6Disks x 1TB + 3Nodes x 6Disks x 2TB = 84TB
Loss to m=2: 84TB / 11Nodes x 2m = 15.27TB
EC capacity: 84TB - 15.27TB = 68.72TB


Thanks in advance,
Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Cephfs mount not recovering after icmp-not-reachable

2021-06-14 Thread Simon Sutter
Hello everyone!

We had a switch outage and the ceph kernel mount did not work anymore.
This is the fstab entry:
10.99.10.1:/somefolder   /cephfs  ceph 
_netdev,nofail,name=cephcluster,secret=IsSecret  0 0

I reproduced it with disabling the vlan on the switch on which the ceph is 
reachable, which gives a icmp-not-reachable.
I did this for five minutes, after that, "ls /cephfs" just gives a "permission 
denied"

in dmesg i can see this:

[ 1412.994921] libceph: mon1 10.99.10.4:6789 session lost, hunting for new mon
[ 1413.009325] libceph: mon0 10.99.10.1:6789 session established
[ 1452.998646] libceph: mon2 10.99.15.3:6789 session lost, hunting for new mon
[ 1452.998679] libceph: mon0 10.99.10.1:6789 session lost, hunting for new mon
[ 1461.989549] libceph: mon4 10.99.15.5:6789 socket closed (con state 
CONNECTING)
---
[ 1787.045148] libceph: mon3 10.99.15.4:6789 socket closed (con state 
CONNECTING)
[ 1787.062587] libceph: mon0 10.99.10.1:6789 session established
[ 1787.086103] libceph: mon4 10.99.15.5:6789 session established
[ 1814.028761] libceph: mds0 10.99.10.4:6801 socket closed (con state OPEN)
[ 1815.029811] libceph: mds0 10.99.10.4:6801 connection reset
[ 1815.029829] libceph: reset on mds0
[ 1815.029831] ceph: mds0 closed our session
[ 1815.029833] ceph: mds0 reconnect start
[ 1815.052219] ceph: mds0 reconnect denied
[ 1815.052229] ceph:  dropping dirty Fw state for 9d9085da1340 1099512175611
[ 1815.052231] ceph:  dropping dirty+flushing Fw state for 9d9085da1340 
1099512175611
[ 1815.273008] libceph: mds0 10.99.10.4:6801 socket closed (con state 
NEGOTIATING)
[ 1816.033241] ceph: mds0 rejected session
[ 1829.018643] ceph: mds0 hung
[ 1880.088504] ceph: mds0 came back
[ 1880.088662] ceph: mds0 caps renewed
[ 1880.094018] ceph: get_quota_realm: ino (1000afe.fffe) null 
i_snap_realm
[ 1881.100367] ceph: get_quota_realm: ino (1000afe.fffe) null 
i_snap_realm
[ 2046.768969] conntrack: generic helper won't handle protocol 47. Please 
consider loading the specific helper module.
[ 2061.731126] ceph: get_quota_realm: ino (1000afe.fffe) null 
i_snap_realm

Is this a bug to report or wrong configuration?
Did someone else had this before?

To solve the problem, a simple remount does the trick.


Thanks in advance
Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] configure fuse in fstab

2021-07-01 Thread Simon Sutter
Hello Everyone!

I'm trying to mount the ceph, with the fuse client under debian 9 (ceph-fuse 
10.2.11-2).

Ceph is on the latest Octopus release.
The direct command is working, but writing it in fstab does not.


Command I use:
ceph-fuse --id dev.wsc -k /etc/ceph/ceph.clinet.dev.wsc.keyring -r 
/testing/dev.wsc/userdata /mnt/ceph


/etc/ceph.conf:

[global]
fsid = [fsid]
mon_host = 10.99.15.1 10.99.15.2 10.99.15.3 10.99.15.4 10.99.15.5


/etc/ceph/ceph.clinet.dev.wsc.keyring:

[client.dev.wsc]
key = [key]


fstab:

none /mnt/ceph fuse.ceph 
ceph.id=dev.wsc,ceph.conf=/etc/ceph/ceph.conf,ceph.client_mountpoint=/testing/dev.wsc/userdata,_netdev,defaults
 0 0


With mount -a it tells me "ceph mount failed with (1) Operation not permitted".


Where can I find authentication logs on the cluster?


Thanks in advance and kind regards,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: configure fuse in fstab

2021-07-01 Thread Simon Sutter
Hello Stefan!

Thanks for the input.


Yes that was a typo.

I have created the file /etc/ceph/ceph.client.dev.wsc.key and entered just the 
key without anything else.


fstab line now looks like this:

none/mnt/ceph   fuse.ceph   
ceph.id=dev.wsc,ceph.client_mountpoint=/testing/dev.wsc/userdata,_netdev,defaults
   0   0


I still get the same "Permission denied" message.


But this fuse command also works:
ceph-fuse --id dev.wsc -r /testing/dev.wsc/userdata /mnt/ceph


Does anyone have any other ideas?


Thanks,

Simon


Von: Stefan Kooman 
Gesendet: Donnerstag, 1. Juli 2021 16:42:56
An: Simon Sutter; ceph-users@ceph.io
Betreff: Re: [ceph-users] configure fuse in fstab

On 7/1/21 4:14 PM, Simon Sutter wrote:
> Hello Everyone!
>
> I'm trying to mount the ceph, with the fuse client under debian 9 (ceph-fuse 
> 10.2.11-2).
>
> Ceph is on the latest Octopus release.
> The direct command is working, but writing it in fstab does not.
>
>
> Command I use:
> ceph-fuse --id dev.wsc -k /etc/ceph/ceph.clinet.dev.wsc.keyring -r 
> /testing/dev.wsc/userdata /mnt/ceph
>
>
> /etc/ceph.conf:
>
> [global]
>  fsid = [fsid]
>  mon_host = 10.99.15.1 10.99.15.2 10.99.15.3 10.99.15.4 10.99.15.5
>
>
> /etc/ceph/ceph.clinet.dev.wsc.keyring:

^^ I guess that is a typo

>
> [client.dev.wsc]
>  key = [key]

Can you try to put the file "ceph.client.dev.wsc.key" in /etc/ceph and
only put the key in it (nothing else) ... and try again

Gr. Stefan

P.s.

There should be no need for ceph.conf=/etc/ceph/ceph.conf as its the
default location.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Performance optimization

2021-09-06 Thread Simon Sutter
Hello everyone!

I have built two clusters with old hardware, which is lying around, the 
possibility to upgrade is there.
The clusters main usecase is hot backup. This means it's getting written 24/7 
where 99% is writing and 1% is reading.


It should be based on harddisks.



At the moment, the nodes look like this:
8 Nodes
Worst CPU: i7-3930K (up to i7-6850K)

Worst ammount of RAM: 24GB (up to 64GB)
HDD Layout:
1x 1TB
4x 2TB
1x 6TB
all sata, some just 5400rpm

I had to put the OS on the 6TB HDDs, because there are no more sata connections 
on the motherboard.

The servers, which have to be backed up, have mounted the ceph with cephfs.
99% of the files, that have to be backed up, are harddisk images, so sizes from 
5GB to 1TB.

All files are written to an erasure-coded pool with k=6 m=2, compression is on 
passive snappy, default settings.

I'm getting really bad performace with this setup.
This is a bench, run with: "rados -p ec_test bench -b 524288 60 write" while 
normal operations:

Total time run: 63.4957
Total writes made:  459
Write size: 524288
Object size:524288
Bandwidth (MB/sec): 3.61442
Stddev Bandwidth:   3.30073
Max bandwidth (MB/sec): 16
Min bandwidth (MB/sec): 0
Average IOPS:   7
Stddev IOPS:6.6061
Max IOPS:   32
Min IOPS:   0
Average Latency(s): 2.151
Stddev Latency(s):  2.3661
Max latency(s): 14.0916
Min latency(s): 0.0420954
Cleaning up (deleting benchmark objects)
Removed 459 objects
Clean up completed and total clean up time :35.6908

[root@testnode01 ~]# ceph osd perf
osd  commit_latency(ms)  apply_latency(ms)
  6 655655
  9  13 13
 11  15 15
  7  17 17
 10  19 19
  8  12 12
 24 153153
 25  22 22
 47  20 20
 46  23 23
 45  43 43
 44   8  8
 16  26 26
 15  18 18
 14  14 14
 13  23 23
 12  47 47
 18 595595
  1  20 20
 38  25 25
 17  17 17
  0 317317
 37  19 19
 19  14 14
  2  16 16
 39   9  9
 20  16 16
  3  18 18
 40  10 10
 21  23 23
  4  17 17
 41  29 29
  5  18 18
 42  16 16
 22  16 16
 23  13 13
 26  20 20
 27  10 10
 28  28 28
 29  13 13
 30  34 34
 31  10 10
 32  31 31
 33  44 44
 34  21 21
 35  22 22
 36 295295
 43   9  9



What do you think is the most obvious Problem?

- The one 6TB disk, per node?
- The OS on the 6TB disk?

What would you suggest?

What I hope to replace with this setup:
6 servers, each with 4x3TB disks, with lvm, no redundancy. (two times, that's 
why I have set up two clusters)

Thanks in advance

Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Performance optimization

2021-09-06 Thread Simon Sutter


Hello

Thanks for this first input, I already found, at least one of those 6TB Disks 
is a WD Blue WD60EZAZ which is according to WD with SMR.
I will replace everything with SMR in it, but in the process of replacing 
hardware, should I replace all disks with for example all 3TB disks?
And what do you think about having the os on one of the disks, used by ceph?

Thanks in advance,
Simon



Von: Kai Börnert 
Gesendet: Montag, 6. September 2021 10:54:24
An: ceph-users@ceph.io
Betreff: [ceph-users] Re: Performance optimization

Hi,

are any of those old disks SMR ones? Because they will absolutely
destroy any kind of performance (ceph does not use writecaches due to
powerloss concerns, so they kinda do their whole magic for each
writerequest).

Greetings

On 9/6/21 10:47 AM, Simon Sutter wrote:
> Hello everyone!
>
> I have built two clusters with old hardware, which is lying around, the 
> possibility to upgrade is there.
> The clusters main usecase is hot backup. This means it's getting written 24/7 
> where 99% is writing and 1% is reading.
>
>
> It should be based on harddisks.
>
>
>
> At the moment, the nodes look like this:
> 8 Nodes
> Worst CPU: i7-3930K (up to i7-6850K)
>
> Worst ammount of RAM: 24GB (up to 64GB)
> HDD Layout:
> 1x 1TB
> 4x 2TB
> 1x 6TB
> all sata, some just 5400rpm
>
> I had to put the OS on the 6TB HDDs, because there are no more sata 
> connections on the motherboard.
>
> The servers, which have to be backed up, have mounted the ceph with cephfs.
> 99% of the files, that have to be backed up, are harddisk images, so sizes 
> from 5GB to 1TB.
>
> All files are written to an erasure-coded pool with k=6 m=2, compression is 
> on passive snappy, default settings.
>
> I'm getting really bad performace with this setup.
> This is a bench, run with: "rados -p ec_test bench -b 524288 60 write" while 
> normal operations:
>
> Total time run: 63.4957
> Total writes made:  459
> Write size: 524288
> Object size:524288
> Bandwidth (MB/sec): 3.61442
> Stddev Bandwidth:   3.30073
> Max bandwidth (MB/sec): 16
> Min bandwidth (MB/sec): 0
> Average IOPS:   7
> Stddev IOPS:6.6061
> Max IOPS:   32
> Min IOPS:   0
> Average Latency(s): 2.151
> Stddev Latency(s):  2.3661
> Max latency(s): 14.0916
> Min latency(s): 0.0420954
> Cleaning up (deleting benchmark objects)
> Removed 459 objects
> Clean up completed and total clean up time :35.6908
>
> [root@testnode01 ~]# ceph osd perf
> osd  commit_latency(ms)  apply_latency(ms)
>6 655655
>9  13 13
>   11  15 15
>7  17 17
>   10  19 19
>8  12 12
>   24 153153
>   25  22 22
>   47  20 20
>   46  23 23
>   45  43 43
>   44   8  8
>   16  26 26
>   15  18 18
>   14  14 14
>   13  23 23
>   12  47 47
>   18 595595
>1  20 20
>   38  25 25
>   17  17 17
>0 317317
>   37  19 19
>   19  14 14
>2  16 16
>   39   9  9
>   20  16 16
>3  18 18
>   40  10 10
>   21  23 23
>4  17 17
>   41  29 29
>5  18 18
>   42  16 16
>   22  16 16
>   23  13 13
>   26  20 20
>   27  10 10
>   28  28 28
>   29  13 13
>   30  34 34
>   31  10 10
>   32  31 31
>   33  44 44
>   34 

[ceph-users] Re: Performance optimization

2021-09-06 Thread Simon Sutter
Hello


> >

> > >> - The one 6TB disk, per node?
> > >
> > > You get bad distribution of data, why not move drives around between
> > these to clusters, so you have more the same in each.
> > >
> >
> > I would assume that this behaves exactly the other way around. As long
> > as you have the same number of block devices with the same size
> > distribution in each node you will get an even data distribution.
> >
> > If you have a node with 4 3TB drives and one with 4 6TB drives Ceph
> > cannot use the 6TB drives efficiently.
> >
> He has 2 clusters thus 3TB -> cluster 1, 6TB -> cluster eg.



Sorry for the bad Information.

I have two clusters, but my question was about just one of them.


Yes Robert is right, instead of this configuration:

| node1 | node2 | node3 | node4 | node5 | node6 | node7 | node8 |
| 1x1TB | 1x1TB | 1x1TB | 1x1TB | 1x1TB | 1x1TB | 1x1TB | 1x1TB |
| 4x2TB | 4x2TB | 4x2TB | 4x2TB | 4x2TB | 4x2TB | 4x2TB | 4x2TB |
| 1x6TB | 1x6TB | 1x6TB | 1x6TB | 1x6TB | 1x6TB | 1x6TB | 1x6TB |

This:
| node1 | node2 | node3 | node4 | node5 | node6 | node7 | node8 |
| 6x3TB | 6x3TB | 6x3TB | 6x3TB | 6x3TB | 6x3TB | 6x3TB | 6x3TB |


Would this even be a noticeable performance difference? Because if I'm not 
mistaken, ceph will try to fill every disk on one node to the same percentage.


And about Erasure Coded: what would be the recommended specification?
Because replicated uses so much more storage, it wasn't really an option until 
now.
We didn't have any problems with CPU utilization and I can go to 32GB for every 
node, and 64 for MDS nodes.


Thanks




Von: Marc 
Gesendet: Montag, 6. September 2021 13:53:06
An: Robert Sander; ceph-users@ceph.io
Betreff: [ceph-users] Re: Performance optimization

>
> >> - The one 6TB disk, per node?
> >
> > You get bad distribution of data, why not move drives around between
> these to clusters, so you have more the same in each.
> >
>
> I would assume that this behaves exactly the other way around. As long
> as you have the same number of block devices with the same size
> distribution in each node you will get an even data distribution.
>
> If you have a node with 4 3TB drives and one with 4 6TB drives Ceph
> cannot use the 6TB drives efficiently.
>
He has 2 clusters thus 3TB -> cluster 1, 6TB -> cluster eg.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Performance optimization

2021-09-07 Thread Simon Sutter


I never looked into RocksDB, because I thought writing data 24/7 does not 
benefit from caching.
But this is metadata storage, so I might profit from it.

Due to lack of sata ports, is it possible to save all RocksDB's on one ssd?
It should still be faster, to write it to just one ssd, instead of writing it 
to the disk directly.

Tahnks,
Simon

Von: Robert Sander 
Gesendet: Montag, 6. September 2021 16:48:52
An: Simon Sutter; Marc; ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Performance optimization

Am 06.09.21 um 16:44 schrieb Simon Sutter:

> |node1|node2|node3|node4|node5|node6|node7|node8|
> |1x1TB|1x1TB|1x1TB|1x1TB|1x1TB|1x1TB|1x1TB|1x1TB|
> |4x2TB|4x2TB|4x2TB|4x2TB|4x2TB|4x2TB|4x2TB|4x2TB|
> |1x6TB|1x6TB|1x6TB|1x6TB|1x6TB|1x6TB|1x6TB|1x6TB|

"ceph osd df tree" should show the data distribution among the OSDs.

Are all of these HDDs? Are these HDDs equipped with RocksDB on SSD?
HDD only will have abysmal performance.

Regards
--
Robert Sander
Heinlein Consulting GmbH
Schwedter Str. 8/9b, 10119 Berlin

http://www.heinlein-support.de

Tel: 030 / 405051-43
Fax: 030 / 405051-19

Zwangsangaben lt. §35a GmbHG:
HRB 93818 B / Amtsgericht Berlin-Charlottenburg,
Geschäftsführer: Peer Heinlein -- Sitz: Berlin
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Upgrading to Octopus

2020-04-22 Thread Simon Sutter
Hello everybody


In octopus there are some interesting looking features, so I tried to upgrading 
my Centos 7 test nodes, according to:
https://docs.ceph.com/docs/master/releases/octopus/

Everything went fine and the cluster is healthy.


To test out the new dashboard functions, I tried to install it, but there are 
missing dependencies:

yum install ceph-mgr-dashboard.noarch

.

--> Finished Dependency Resolution
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-routes
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-jwt
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-cherrypy


Installing them with pip3 does of course make no difference, because those are 
yum dependencies.

Does anyone know a workaround?

Do I have to upgrade to Centos 8 for this to work?


Thanks in advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Upgrading to Octopus

2020-04-23 Thread Simon Sutter
Hello Khodayar


Of cours I tried installing them with yum...

They are not available in the centos base or epel repos, here are the ones, 
which are available:


[root@node1 ~]# yum list | egrep "cherrypy|jwt|routes"
python-cherrypy.noarch 3.2.2-4.el7@base
python-cherrypy2.noarch2.3.0-19.el7   @epel
python-jwt.noarch  1.5.3-1.el7@base
python-routes.noarch   1.13-2.el7 @epel
nodejs-jwt-simple.noarch   0.2.0-1.el7epel
python36-jwt.noarch1.6.4-2.el7epel


How do I get either: The right packages or a workaround because i can install 
the dependencies with pip?


Regards,

Simon



Von: Khodayar Doustar 
Gesendet: Mittwoch, 22. April 2020 20:02:04
An: Simon Sutter
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Upgrading to Octopus

Hi Simon,

Have you tried installing them with yum?




On Wed, Apr 22, 2020 at 6:16 PM Simon Sutter 
mailto:ssut...@hosttech.ch>> wrote:
Hello everybody


In octopus there are some interesting looking features, so I tried to upgrading 
my Centos 7 test nodes, according to:
https://docs.ceph.com/docs/master/releases/octopus/

Everything went fine and the cluster is healthy.


To test out the new dashboard functions, I tried to install it, but there are 
missing dependencies:

yum install ceph-mgr-dashboard.noarch

.

--> Finished Dependency Resolution
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-routes
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-jwt
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-cherrypy


Installing them with pip3 does of course make no difference, because those are 
yum dependencies.

Does anyone know a workaround?

Do I have to upgrade to Centos 8 for this to work?


Thanks in advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Upgrading to Octopus

2020-04-23 Thread Simon Sutter
Khodayar,


I added all those repos, but sitll, those packages are missing.

I can of course search for the exact package name like this:


[root@node1 ~]# yum search python3-cherrypy
Loaded plugins: fastestmirror, langpacks, priorities
Loading mirror speeds from cached hostfile
 * base: pkg.adfinis-sygroup.ch
 * centos-ceph-nautilus: pkg.adfinis-sygroup.ch
 * centos-nfs-ganesha28: pkg.adfinis-sygroup.ch
 * centos-openstack-stein: pkg.adfinis-sygroup.ch
 * centos-qemu-ev: pkg.adfinis-sygroup.ch
 * centos-sclo-rh: pkg.adfinis-sygroup.ch
 * centos-sclo-sclo: pkg.adfinis-sygroup.ch
 * epel: pkg.adfinis-sygroup.ch
 * extras: pkg.adfinis-sygroup.ch
 * updates: pkg.adfinis-sygroup.ch
Warning: No matches found for: python3-cherrypy
No matches found


But as you can see, it cannot find it.

Anything else I can try?


Regards,

Simon


Von: Khodayar Doustar 
Gesendet: Donnerstag, 23. April 2020 14:41:38
An: Simon Sutter
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Upgrading to Octopus

Simon,

You can try to search for the exact package name, you can try these repos as 
well:

yum<https://www.server-world.info/en/command/html/yum.html> -y install 
epel-release centos-release-ceph-nautilus centos-release-openstack-stein


On Thu, Apr 23, 2020 at 11:57 AM Simon Sutter 
mailto:ssut...@hosttech.ch>> wrote:
Hello Khodayar


Of cours I tried installing them with yum...

They are not available in the centos base or epel repos, here are the ones, 
which are available:


[root@node1 ~]# yum list | egrep "cherrypy|jwt|routes"
python-cherrypy.noarch 3.2.2-4.el7@base
python-cherrypy2.noarch2.3.0-19.el7   @epel
python-jwt.noarch  1.5.3-1.el7@base
python-routes.noarch   1.13-2.el7 @epel
nodejs-jwt-simple.noarch   0.2.0-1.el7epel
python36-jwt.noarch1.6.4-2.el7epel


How do I get either: The right packages or a workaround because i can install 
the dependencies with pip?


Regards,

Simon



Von: Khodayar Doustar mailto:dous...@rayanexon.ir>>
Gesendet: Mittwoch, 22. April 2020 20:02:04
An: Simon Sutter
Cc: ceph-users@ceph.io<mailto:ceph-users@ceph.io>
Betreff: Re: [ceph-users] Upgrading to Octopus

Hi Simon,

Have you tried installing them with yum?




On Wed, Apr 22, 2020 at 6:16 PM Simon Sutter 
mailto:ssut...@hosttech.ch><mailto:ssut...@hosttech.ch<mailto:ssut...@hosttech.ch>>>
 wrote:
Hello everybody


In octopus there are some interesting looking features, so I tried to upgrading 
my Centos 7 test nodes, according to:
https://docs.ceph.com/docs/master/releases/octopus/

Everything went fine and the cluster is healthy.


To test out the new dashboard functions, I tried to install it, but there are 
missing dependencies:

yum install ceph-mgr-dashboard.noarch

.

--> Finished Dependency Resolution
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-routes
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-jwt
Error: Package: 2:ceph-mgr-dashboard-15.2.1-0.el7.noarch (Ceph-noarch)
   Requires: python3-cherrypy


Installing them with pip3 does of course make no difference, because those are 
yum dependencies.

Does anyone know a workaround?

Do I have to upgrade to Centos 8 for this to work?


Thanks in advance,

Simon
___
ceph-users mailing list -- 
ceph-users@ceph.io<mailto:ceph-users@ceph.io><mailto:ceph-users@ceph.io<mailto:ceph-users@ceph.io>>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io><mailto:ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>>
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Upgrading to Octopus

2020-04-28 Thread Simon Sutter
Hello,


Yes I upgraded the system to Centos8 and now I can install the dashboard module.

But the problem now is, I cannot log in to the dashboard.

I deleted every cached file on my end and reinstalled the mgr and dashboard 
several times.

If I try to log in with a wrong password, it tells me that it's wrong, but if i 
use the right password, it just gives me a "500 Internal Server Error".
I enabled the debug mode for the mgr: ceph config set mgr 
mgr/dashboard/log_level debug
But in the /var/log/ceph/ceph-mgr.node1.log it just tells me this generic 
message (ips replaced with 0.0.0.0):


2020-04-28T14:11:15.191+0200 7f0baba8c700  0 [dashboard DEBUG request] 
[:::0.0.0.0:61383] [POST] [None] /api/auth
2020-04-28T14:11:15.282+0200 7f0bcf164700  0 log_channel(cluster) log [DBG] : 
pgmap v316: 273 pgs: 273 active+clean; 2.4 TiB data, 7.1 TiB used, 18 TiB / 25 
TiB avail
2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard DEBUG controllers.auth] 
Login successful
2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard ERROR request] 
[:::0.0.0.0:61383] [POST] [500] [0.264s] [513.0B] 
[100ecd9a-5d09-419f-8b9f-31bc3d4042b4] /api/auth
2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard ERROR request] 
[b'{"status": "500 Internal Server Error", "detail": "The server encountered an 
unexpected condition which prevented it from fulfilling the request.", 
"request_id": "100ecd9a-5d09-419f-8b9f-31bc3d4042b4"}   



 ']
2020-04-28T14:11:15.454+0200 7f0baba8c700  0 [dashboard INFO request] 
[:::0.0.0.0:61383] [POST] [500] [0.264s] [513.0B] 
[100ecd9a-5d09-419f-8b9f-31bc3d4042b4] /api/auth


How can I find out, where the problem is?


Thanks in advance,

Simon


Von: gert.wieberd...@ziggo.nl 
Gesendet: Donnerstag, 23. April 2020 20:34:58
An: ceph-users@ceph.io
Betreff: [ceph-users] Re: Upgrading to Octopus

Hello Simon,
I think that Khodayar is right. I managed to install a new Ceph cluster on 
CentOS 8.1. Therefore you will need the ceph-el8.repo for the time being. For 
some reason, "they" left the py3 packages you mentioned out of EPEL (as with 
leveldb, but this package appeared luckily last week in EPEL).
Please find below the ceph-el8.repo file, which you have to create in 
/etc/yum.repos.d/

[copr:copr.fedorainfracloud.org:ktdreyer:ceph-el8]
name=Copr repo for ceph-el8 owned by ktdreyer
baseurl=https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/epel-8-$basearch/
type=rpm-md
skip_if_unavailable=True
gpgcheck=1
gpgkey=https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/pubkey.gpg
repo_gpgcheck=0
enabled=1
enabled_metadata=1

This repository - and CentOS 8.x - should have been sufficient to bring up a 
fresh Ceph cluster.
Please let me know if you still have problems in configuring your Ceph cluster.
rgds,
-gw
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Upgrading to Octopus

2020-04-28 Thread Simon Sutter
Hello Gert,

I recreated the self signed certificate.

SELinux was disabled and I temporarely disabled the firewall.

It still doesn't work and there is no entry in journalctl -f.


Somewhere there is still something from the previous nautilus or centos7 
installation, causing this problem.
I think I'll have to reinstall the node.


I'll update you.

Thanks and kind regards,

Simon


Von: Gert Wieberdink 
Gesendet: Dienstag, 28. April 2020 21:16:10
An: Simon Sutter; ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Upgrading to Octopus

Sorry for the typo: must be journalctl -f instead of syslogctl -f.
-gw


On Tue, 2020-04-28 at 19:12 +, Gert Wieberdink wrote:

Hello Simon,ceph-mgr and dashboard installation should be

straightforward.

These are tough ones (internal server error 500). Did you create a self

signed cert for dashboard?Did you check firewalld (port 8443) and/or

SELinux? Does syslogctl -f show anything?

rgds,-gw

On Tue, 2020-04-28 at 12:17 +0000, Simon Sutter wrote:








Hello,




Yes I upgraded the system to Centos8 and now I can install the

dashboard module.

But the problem now is, I cannot log in to the dashboard.

I deleted every cached file on my end and reinstalled the mgr and

dashboard several times.

If I try to log in with a wrong password, it tells me that it's

wrong, but if i use the right password, it just gives me a "500

Internal Server Error".


I enabled the debug mode for the mgr: ceph config set mgr

mgr/dashboard/log_level debug


But in the /var/log/ceph/ceph-mgr.node1.log it just tells me this

generic message (ips replaced with 0.0.0.0):






2020-04-28T14:11:15.191+0200 7f0baba8c700  0 [dashboard DEBUG

request] [:::0.0.0.0:61383] [POST] [None] /api/auth

2020-04-28T14:11:15.282+0200 7f0bcf164700  0 log_channel(cluster) log

[DBG] : pgmap v316: 273 pgs: 273 active+clean; 2.4 TiB data, 7.1 TiB

used, 18 TiB / 25 TiB avail

2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard DEBUG

controllers.auth] Login successful

2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard ERROR

request] [:::0.0.0.0:61383] [POST] [500] [0.264s] [513.0B]

[100ecd9a-5d09-419f-8b9f-31bc3d4042b4] /api/auth

2020-04-28T14:11:15.453+0200 7f0baba8c700  0 [dashboard ERROR

request] [b'{"status": "500 Internal Server Error", "detail": "The

server encountered an unexpected condition which prevented it from

fulfilling the request.", "request_id": "100ecd9a-5d09-419f-8b9f-

31bc3d4042b4"}









   ']

2020-04-28T14:11:15.454+0200 7f0baba8c700  0 [dashboard INFO request]

[:::0.0.0.0:61383] [POST] [500] [0.264s] [513.0B] [100ecd9a-5d09-

419f-8b9f-31bc3d4042b4] /api/auth




How can I find out, where the problem is?




Thanks in advance,

Simon



Von:

<mailto:gert.wieberd...@ziggo.nl>

gert.wieberd...@ziggo.nl

 <

<mailto:gert.wieberd...@ziggo.nl>

gert.wieberd...@ziggo.nl

>


Gesendet: Donnerstag, 23. April 2020 20:34:58


An:

<mailto:ceph-users@ceph.io>

ceph-users@ceph.io



Betreff: [ceph-users] Re: Upgrading to Octopus






Hello Simon,


I think that Khodayar is right. I managed to install a new Ceph

cluster on CentOS 8.1. Therefore you will need the ceph-el8.repo for

the time being. For some reason, "they" left the py3 packages you

mentioned out of EPEL (as with leveldb, but this package appeared

 luckily last week in EPEL).


Please find below the ceph-el8.repo file, which you have to create in

/etc/yum.repos.d/




[copr:copr.fedorainfracloud.org:ktdreyer:ceph-el8]


name=Copr repo for ceph-el8 owned by ktdreyer


baseurl=

<https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/epel-8-$basearch/>

https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/epel-8-$basearch/



type=rpm-md


skip_if_unavailable=True


gpgcheck=1


gpgkey=

<https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/pubkey.gpg>

https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/pubkey.gpg



repo_gpgcheck=0


enabled=1


enabled_metadata=1




This repository - and CentOS 8.x - should have been sufficient to

bring up a fresh Ceph cluster.


Please let me know if you still have problems in configuring your

Ceph cluster.


rgds,


-gw


___


ceph-users mailing list --

<mailto:ceph-users@ceph.io>

ceph-users@ceph.io



To unsubscribe send an email to

<mailto:ceph-users-le...@ceph.io>

ceph-users-le...@ceph.io








___

ceph-users mailing list --

<mailto:ceph-users@ceph.io>

ceph-users@ceph.io


To unsubscribe send an email to

<mailto:ceph-users-le...@ceph.io>

ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Add lvm in cephadm

2020-05-05 Thread Simon Sutter
Hello Everyone,



The new cephadm is giving me a headache.

I'm setting up a new testenvironment, where I have to use lvm partitions, 
because I don't have more Hardware.

I could't find any information about the compatibility of existing lvm 
partitions and cephadm/octopus.

I tried the old method, where I had to use ceph-volume to use an existing lvm:


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Add lvm in cephadm

2020-05-05 Thread Simon Sutter
Sorry I missclicked, here the second part:


ceph-volume --cluster ceph lvm prepare --data /dev/centos_node1/ceph
But that gives me just:

Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd 
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 
f3b442b1-68f7-456a-9991-92254e7c9c30
 stderr: [errno 13] RADOS permission denied (error connecting to the cluster)
-->  RuntimeError: Unable to create a new OSD id



So, how do I create an osd out of an existing lvm partition with cephadm?


Thanks in advance
Simon


Von: Simon Sutter 
Gesendet: Dienstag, 5. Mai 2020 10:43:10
An: ceph-users@ceph.io
Betreff: [ceph-users] Add lvm in cephadm

Hello Everyone,



The new cephadm is giving me a headache.

I'm setting up a new testenvironment, where I have to use lvm partitions, 
because I don't have more Hardware.

I could't find any information about the compatibility of existing lvm 
partitions and cephadm/octopus.

I tried the old method, where I had to use ceph-volume to use an existing lvm:


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
hosttech GmbH | Simon Sutter
hosttech.ch<https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Add lvm in cephadm

2020-05-12 Thread Simon Sutter
Hello,


Thank you very much Joshua, it worked.

I have set up three nodes with the cephadm tool, which was very easy.

But I asked myself, what if node 1 goes down?

Before cephadm I just could manage everything from the other nodes with the 
ceph commands.

Now I'm a bit stuck, because this cephadm container is just running on one node.

I've installed it on the second one, but i'm getting a "[errno 13] RADOS 
permission denied (error connecting to the cluster)".

Do I need some special "cephadm" keyring from the first node? Which one? And 
where to put it?

Caphadm might be an easy to handle solution, but for me as a beginner, the 
added layer is very complicated to get in.

We are trying to build a new Ceph cluster (never got in touch with it before) 
but I might not go with octopus, but instead  use nautilus with ceph-deploy.

That's a bit easyer to understand, and the documentation out there is way 
better.


Thanks in advance,

Simon


Von: Joshua Schmid 
Gesendet: Dienstag, 5. Mai 2020 16:39:29
An: Simon Sutter
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Add lvm in cephadm

On 20/05/05 08:46, Simon Sutter wrote:
> Sorry I missclicked, here the second part:
>
>
> ceph-volume --cluster ceph lvm prepare --data /dev/centos_node1/ceph
> But that gives me just:
>
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd 
> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 
> f3b442b1-68f7-456a-9991-92254e7c9c30
>  stderr: [errno 13] RADOS permission denied (error connecting to the cluster)
> -->  RuntimeError: Unable to create a new OSD id

Hey Simon,

This still works but is now encapsulated in a cephadm
command.

ceph orch daemon add osd :

so in your case:

ceph orch daemon add osd $host:centos_node1/ceph


hth

--
Joshua Schmid
Software Engineer
SUSE Enterprise Storage
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: how to restart daemons on 15.2 on Debian 10

2020-05-15 Thread Simon Sutter
Hello Michael,


I had the same problems. It's very unfamiliar, if you never worked with the 
cephadm tool.

The Way I'm doing it is to go into the cephadm container:
# cephadm shell

Here you can list all containers (for each service, one container) with the 
orchestration tool:

# ceph orch ps

and then restart it with the orchestration tool:

# ceph orch restart {name from ceph orch ps}


Hope it helps.


Ceers,

Simon


Von: Ml Ml 
Gesendet: Freitag, 15. Mai 2020 12:27:09
An: ceph-users
Betreff: [ceph-users] how to restart daemons on 15.2 on Debian 10

Hello List,

how do you restart daemons (mgr, mon, osd) on 15.2.1?

It used to be something like:
  systemctl stop ceph-osd@10

Or:
  systemctl start ceph-mon@ceph03

however, those command do nothing on my setup.

Is this because i use cephadm and that docker stuff?

The Logs also seem to be missing.
/var/log/ceph/5436dd5d-83d4-4dc8-a93b-60ab5db145df is pretty empty.

I feel like i am missing a lot of documentation here? Can anyone point
me to my missing parts?

Thanks a lot.

Cheers,
Michael
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] mds container dies during deployment

2020-05-25 Thread Simon Sutter
Hello everyone


I've got a fresh ceph octopus installation and I'm trying to set up a cephfs 
with erasure code configuration.
The metadata pool was set up as default.
The erasure code pool was set up with this command:
-> ceph osd pool create ec-data_fs 128 erasure default
Enabled overwrites:
-> ceph osd pool set ec-data_fs allow_ec_overwrites true
And create fs:
-> ceph fs new ec-data_fs meta_fs ec-data_fs --force


Then I tried deploying the mds, but this fails:
-> ceph orch daemon add mds ec-data_fs magma01
returns:
-> Deployed mds.ec-data_fs.magma01.ujpcly on host 'magma01'

The mds daemon is not there.

Aparently the container dies without any information, as seen in the journal:

May 25 16:11:56 magma01 podman[9348]: 2020-05-25 16:11:56.670510456 +0200 CEST 
m=+0.186462913 container create 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:56 magma01 systemd[1]: Started 
libpod-conmon-0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.scope.
May 25 16:11:56 magma01 systemd[1]: Started libcontainer container 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.112182262 +0200 CEST 
m=+0.628134873 container init 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.137011897 +0200 CEST 
m=+0.652964354 container start 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.137110412 +0200 CEST 
m=+0.653062853 container attach 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 systemd[1]: 
libpod-0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.scope: 
Consumed 327ms CPU time
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.182968802 +0200 CEST 
m=+0.698921275 container died 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.413743787 +0200 CEST 
m=+0.929696266 container remove 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)

Can someone help me debugging this?

Cheers
Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: mds container dies during deployment

2020-05-26 Thread Simon Sutter
Hello,


Didn't read the right one:
https://docs.ceph.com/docs/master/cephadm/install/#deploy-mdss
There it says, how to do it right.

The command I was using, was just to add a mds daemon if you have already one.


Hopes it helps others.

Cheers, Simon


Von: Simon Sutter 
Gesendet: Montag, 25. Mai 2020 16:44:54
An: ceph-users@ceph.io
Betreff: [ceph-users] mds container dies during deployment

Hello everyone


I've got a fresh ceph octopus installation and I'm trying to set up a cephfs 
with erasure code configuration.
The metadata pool was set up as default.
The erasure code pool was set up with this command:
-> ceph osd pool create ec-data_fs 128 erasure default
Enabled overwrites:
-> ceph osd pool set ec-data_fs allow_ec_overwrites true
And create fs:
-> ceph fs new ec-data_fs meta_fs ec-data_fs --force


Then I tried deploying the mds, but this fails:
-> ceph orch daemon add mds ec-data_fs magma01
returns:
-> Deployed mds.ec-data_fs.magma01.ujpcly on host 'magma01'

The mds daemon is not there.

Aparently the container dies without any information, as seen in the journal:

May 25 16:11:56 magma01 podman[9348]: 2020-05-25 16:11:56.670510456 +0200 CEST 
m=+0.186462913 container create 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:56 magma01 systemd[1]: Started 
libpod-conmon-0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.scope.
May 25 16:11:56 magma01 systemd[1]: Started libcontainer container 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.112182262 +0200 CEST 
m=+0.628134873 container init 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.137011897 +0200 CEST 
m=+0.652964354 container start 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.137110412 +0200 CEST 
m=+0.653062853 container attach 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 systemd[1]: 
libpod-0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90.scope: 
Consumed 327ms CPU time
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.182968802 +0200 CEST 
m=+0.698921275 container died 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)
May 25 16:11:57 magma01 podman[9348]: 2020-05-25 16:11:57.413743787 +0200 CEST 
m=+0.929696266 container remove 
0fdf8c508b330adac713ffb04c72b5df770277ad191d844888f7387f28e3cc90 
(image=docker.io/ceph/ceph:v15, name=competent_cori)

Can someone help me debugging this?

Cheers
Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
hosttech GmbH | Simon Sutter
hosttech.ch<https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Ceph client on rhel6?

2020-05-26 Thread Simon Sutter
Hello again,


I have a new question:
We want to upgrade a server, with an os based on rhel6.

The ceph cluster is atm on octopus.

How can I install the client packages to mount cephfs and do a backup of the 
server?
Is it even possible?

Are the client packages from hammer compatible with the octopus release?


Thanks in advance,

Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Deploy nfs from cephadm

2020-06-02 Thread Simon Sutter
Hello Ceph users,


I'm trying to deploy nfs-ganesha with cephadm on octopus.

According to the docs, it's as easy as running the command in the docs: 
https://docs.ceph.com/docs/master/cephadm/install/#deploying-nfs-ganesha
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Deploy nfs from cephadm

2020-06-02 Thread Simon Sutter
:CRIT :dbus_bus_get 
failed (Failed to connect to socket /run/dbus/system_bus_socket: No such file 
or d>
Jun 02 15:17:48 testnode1 bash[63432]: 02/06/2020 13:17:48 : epoch 5ed6517a : 
testnode1 : ganesha.nfsd-1[main] gsh_dbus_register_path :DBUS :CRIT 
:dbus_connection_register_object_path called with no DBUS connection
Jun 02 15:17:48 testnode1 bash[63432]: 02/06/2020 13:17:48 : epoch 5ed6517a : 
testnode1 : ganesha.nfsd-1[main] nfs_Init_svc :DISP :CRIT :Cannot acquire 
credentials for principal nfs
Jun 02 15:17:48 testnode1 bash[63432]: 02/06/2020 13:17:48 : epoch 5ed6517a : 
testnode1 : ganesha.nfsd-1[main] __Register_program :DISP :MAJ :Cannot register 
NFS V3 on UDP
Jun 02 15:17:48 testnode1 podman[63531]: 2020-06-02 15:17:48.976740123 +0200 
CEST m=+2.565960305 container died 
aab4e893b162a643181cea3a9f5f687aae236eb2f4a7f6fad27d503d1fdee893 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11e>
Jun 02 15:17:49 testnode1 podman[63531]: 2020-06-02 15:17:49.051611802 +0200 
CEST m=+2.640831963 container remove 
aab4e893b162a643181cea3a9f5f687aae236eb2f4a7f6fad27d503d1fdee893 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
Jun 02 15:17:49 testnode1 systemd[1]: 
ceph-915cdf28-8f66-11ea-bb83-ac1f6b4cd516@nfs.cephnfs.testnode1.service: Main 
process exited, code=exited, status=2/INVALIDARGUMENT
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.158593029 +0200 
CEST m=+0.082356027 container create 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.481433978 +0200 
CEST m=+0.405196986 container init 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11e>
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.495210618 +0200 
CEST m=+0.418973669 container start 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11>
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.495293786 +0200 
CEST m=+0.419056865 container attach 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
Jun 02 15:17:49 testnode1 bash[64057]: rados_connect: -13
Jun 02 15:17:49 testnode1 bash[64057]: Can't connect to cluster: -13
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.55130226 +0200 
CEST m=+0.475065324 container died 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11ea>
Jun 02 15:17:49 testnode1 podman[64059]: 2020-06-02 15:17:49.633569478 +0200 
CEST m=+0.557332529 container remove 
d55e6b6797ad129c6de503c0f6fac6b127fcf978ae44b636a54b0b93ad5010b2 
(image=docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
Jun 02 15:17:49 testnode1 systemd[1]: 
ceph-915cdf28-8f66-11ea-bb83-ac1f6b4cd516@nfs.cephnfs.testnode1.service: Failed 
with result 'exit-code'.


I can see, one container fails to connect to the cluster, but wher can I find 
out why?

Thanks in Advance and sorry for the split mail,
Simon


Von: Simon Sutter 
Gesendet: Dienstag, 2. Juni 2020 16:26:15
An: ceph-users@ceph.io
Betreff: [ceph-users] Deploy nfs from cephadm

Hello Ceph users,


I'm trying to deploy nfs-ganesha with cephadm on octopus.

According to the docs, it's as easy as running the command in the docs: 
https://docs.ceph.com/docs/master/cephadm/install/#deploying-nfs-ganesha
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
hosttech GmbH | Simon Sutter
hosttech.ch<https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Deploy nfs from cephadm

2020-06-03 Thread Simon Sutter
Hello,


Thank you very much.

I was a bit worried about all the other messages, especially those two from a 
different container (started before the right one?):


Jun 03 08:22:23 testnode1 bash[3169]: rados_connect: -13
Jun 03 08:22:23 testnode1 bash[3169]: Can't connect to cluster: -13


Nevertheless, installing and starting rpcbind did it for me.

After I configured the ganesha service in 
"/var/lib/ceph/{UUID}/nfs.cephnfs.testnode1/etc/ganesha/ganesha.conf" I was 
able to start and accesss cephfs from a nfs client.

Maybe a hint in the docs, to install rpcbind, or to enable just nfsv4 would 
also help others.



Best Regards,

Simon



Von: Michael Fritch 
Gesendet: Dienstag, 2. Juni 2020 22:33:46
An: ceph-users@ceph.io
Betreff: [ceph-users] Re: Deploy nfs from cephadm

Hi,

Do you have a running rpcbind service?
$ systemctl status rpcbind

NFSv3 requires rpcbind, but this dependency will be removed in a later release 
of Octopus. I've updated the tracker with more detail.

Hope this helps,
-Mike


John Zachary Dover wrote:
> I've created a docs tracker ticket for this issue:
>
> https://tracker.ceph.com/issues/45819
Bug #45819: Possible error in deploying-nfs-ganesha docs - Orchestrator - 
Ceph<https://tracker.ceph.com/issues/45819>
tracker.ceph.com
Redmine



>
> Zac
> Ceph Docs
>
> On Wed, Jun 3, 2020 at 12:34 AM Simon Sutter 
> >  Sorry, allways the wrong button...
> >
> >  So I ran the command:
> >  ceph orch apply nfs cephnfs cephfs.backuptest.data
> >
> >  And there is now a not working container:
> >  ceph orch ps:
> >  nfs.cephnfs.testnode1testnode1  error  6m ago 71m
> >docker.io/ceph/ceph:v15  
> > 
> >
> >  journalctl tells me this:
> >
> >  Jun 02 15:17:45 testnode1 systemd[1]: Starting Ceph nfs.cephnfs.testnode1
> >  for 915cdf28-8f66-11ea-bb83-ac1f6b4cd516...
> >  Jun 02 15:17:45 testnode1 podman[63413]: Error: no container with name or
> >  ID ceph-915cdf28-8f66-11ea-bb83-ac1f6b4cd516-nfs.cephnfs.testnode1 found:
> >  no such container
> >  Jun 02 15:17:45 testnode1 systemd[1]: Started Ceph nfs.cephnfs.testnode1
> >  for 915cdf28-8f66-11ea-bb83-ac1f6b4cd516.
> >  Jun 02 15:17:45 testnode1 podman[63434]: 2020-06-02 15:17:45.867685349
> >  +0200 CEST m=+0.080338785 container create
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
> >  Jun 02 15:17:46 testnode1 podman[63434]: 2020-06-02 15:17:46.196760186
> >  +0200 CEST m=+0.409413617 container init
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11e>
> >  Jun 02 15:17:46 testnode1 podman[63434]: 2020-06-02 15:17:46.211149759
> >  +0200 CEST m=+0.423803191 container start
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11>
> >  Jun 02 15:17:46 testnode1 podman[63434]: 2020-06-02 15:17:46.21122888
> >  +0200 CEST m=+0.423882373 container attach
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11>
> >  Jun 02 15:17:46 testnode1 bash[63432]: rados_connect: -13
> >  Jun 02 15:17:46 testnode1 bash[63432]: Can't connect to cluster: -13
> >  Jun 02 15:17:46 testnode1 podman[63434]: 2020-06-02 15:17:46.300445833
> >  +0200 CEST m=+0.513099326 container died
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11e>
> >  Jun 02 15:17:46 testnode1 podman[63434]: 2020-06-02 15:17:46.391730251
> >  +0200 CEST m=+0.604383723 container remove
> >  7290cc21b0e2498876773f1ef2a2be24abf62e9ed058d60e79c8f3d3d3e9e0d3 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
> >  Jun 02 15:17:46 testnode1 podman[63531]: 2020-06-02 15:17:46.496154808
> >  +0200 CEST m=+0.085374929 container create
> >  aab4e893b162a643181cea3a9f5f687aae236eb2f4a7f6fad27d503d1fdee893 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-1>
> >  Jun 02 15:17:46 testnode1 podman[63531]: 2020-06-02 15:17:46.81399203
> >  +0200 CEST m=+0.403212198 container init
> >  aab4e893b162a643181cea3a9f5f687aae236eb2f4a7f6fad27d503d1fdee893 (image=
> >  docker.io/ceph/ceph:v15, name=ceph-915cdf28-8f66-11ea>
> >  Jun 02 15:17:46 testnode1 podman[63531]: 2020-06-02 15:17:46.828546918
> >  +0200 CEST m=+0.417767036 container start
> >  aab4e893b162a643181cea3a9f5f687aae2

[ceph-users] Change mon bind address / Change IPs with the orchestrator

2020-06-03 Thread Simon Sutter
Hello,


I think I missunderstood the internal / public network concepts in the docs 
https://docs.ceph.com/docs/master/rados/configuration/network-config-ref/.

Now there are two questions:

- Is it somehow possible to bind the MON daemon to 0.0.0.0?
I tried it with manually add the ip in  /var/lib/ceph/{UUID}/mon.node01/config


[mon.node01]
public bind addr = 0.0.0.0


But that does not work, in netstat I can see, the mon still binds to it's 
internal ip. Is this an expected behaviour?

If I set this value to the public ip, the other nodes cannot communicate with 
it, so this leads to the next question:

- What's the Right way to correct the problem with the orchestrator?
So the correct way to configure the ip's, would be to set every mon, mds and so 
on, to the public ip and just let the osd's stay on their internal ip. 
(described here 
https://docs.ceph.com/docs/master/rados/configuration/network-config-ref/)

Do I have to remove every daemon and redeploy them with "ceph orch daemon rm" / 
"ceph orch apply"?

Or do I have to go to every node and manually apply the settings in the daemon 
config file?


Thanks in advance,


Simon

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Change mon bind address / Change IPs with the orchestrator

2020-06-05 Thread Simon Sutter
Hello,


Ok, thanks Wido.

I have now sorted out the correct network configuration and deployed new mon's 
with the new IP.
Everything is now on the new IP and works so far.


Simon




Von: Wido den Hollander 
Gesendet: Donnerstag, 4. Juni 2020 08:47:29
An: Simon Sutter; ceph-users@ceph.io
Betreff: Re: [ceph-users] Change mon bind address / Change IPs with the 
orchestrator



On 6/3/20 4:49 PM, Simon Sutter wrote:
> Hello,
>
>
> I think I missunderstood the internal / public network concepts in the docs 
> https://docs.ceph.com/docs/master/rados/configuration/network-config-ref/.
>
> Now there are two questions:
>
> - Is it somehow possible to bind the MON daemon to 0.0.0.0?

No

> I tried it with manually add the ip in  /var/lib/ceph/{UUID}/mon.node01/config

Won't work :-)

>
>
> [mon.node01]
> public bind addr = 0.0.0.0
>
>
> But that does not work, in netstat I can see, the mon still binds to it's 
> internal ip. Is this an expected behaviour?
>

Yes. This is not an orchestrator thing, this is how the MONs work. They
need to bind to a specific IP and that can't be 0.0.0.0

You then need to make sure proper routing is in place so all clients and
OSDs can talk to the MONs.

So don't attempt anything like NAT. Make sure everything works with
proper IP-routing.

Wido

> If I set this value to the public ip, the other nodes cannot communicate with 
> it, so this leads to the next question:
>
> - What's the Right way to correct the problem with the orchestrator?
> So the correct way to configure the ip's, would be to set every mon, mds and 
> so on, to the public ip and just let the osd's stay on their internal ip. 
> (described here 
> https://docs.ceph.com/docs/master/rados/configuration/network-config-ref/)
>
> Do I have to remove every daemon and redeploy them with "ceph orch daemon rm" 
> / "ceph orch apply"?
>
> Or do I have to go to every node and manually apply the settings in the 
> daemon config file?
>
>
> Thanks in advance,
>
>
> Simon
>
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm and Ceph versions

2020-06-05 Thread Simon Sutter
Hello Andy,


I had mixed experiences with cephadm.

What I would do:

Check if all your daemons indeed are running in the according containers on 
every node.
You can check it with "ceph orch ps"


If that is the case, you can get rid of the old rpms and install the new 
ceph-common v15 rpm.

You do not need all the rpms you had before, because everything should now be 
inside those containers.

Technically you don't need any ceph rpm installed on any of the nodes except 
for cephadm, but with the ceph-common rpm the conversion is a bit easier.

Please don't do this directly on a production running system and test it first 
in a testenvironment.


Simon


Von: bioh...@yahoo.com 
Gesendet: Freitag, 5. Juni 2020 11:18:41
An: ceph-users@ceph.io
Betreff: [ceph-users] Cephadm and Ceph versions

Hi

I had a cluster on v13 (mimic) and have converted it to Octopus (15.2.3) and 
using Cephadm.  In  the dashboard is showing as all v15

What do I need to do with the Ceph rpms that are installed as they are all Ceph 
version 13.

Do I remove them and install Ceph rpms with version 15 ?

Regards
Andy
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Octopus: orchestrator not working correctly with nfs

2020-06-08 Thread Simon Sutter
Hello


I know that nfs on octopus is still a bit under development.

I'm trying to deploy nfs daemons and have some issues with the orchestartor.

For the other daemons, for example monitors, I can issue the command "ceph orch 
apply mon 3"

This will tell the orchestrator to deploy or remove monitor daemons until there 
are three of them.

The command does not work with nfs, and now the orchestrator is a bit 
missconfigured...

And with missconfigured I mean, that I have now a nfs daemon on node 1 and the 
orchestrator wants to create another one on node 1 but with wrong settings (it 
fails).
Also a "ceph orch apply nfs –unconfigured" does not work, so I can't manually 
manage the nfs containers.

Is there a manual way to tell ceph orch, to not create or remove nfs daemons? 
then I would be able to set them up manually.
Or a manual way of configuring the orchestrator so it does the right thing.


Thanks in advance

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Octopus: orchestrator not working correctly with nfs

2020-06-10 Thread Simon Sutter
Hello,


Did I not provide enough information, or simply nobody knows how to solve the 
problem?
Should I write to the ceph tracker or does this just produce unnecessary 
overhead?


Thanks in advance,

Simon


Von: Simon Sutter 
Gesendet: Montag, 8. Juni 2020 10:56:00
An: ceph-users@ceph.io
Betreff: [ceph-users] Octopus: orchestrator not working correctly with nfs

Hello


I know that nfs on octopus is still a bit under development.

I'm trying to deploy nfs daemons and have some issues with the orchestartor.

For the other daemons, for example monitors, I can issue the command "ceph orch 
apply mon 3"

This will tell the orchestrator to deploy or remove monitor daemons until there 
are three of them.

The command does not work with nfs, and now the orchestrator is a bit 
missconfigured...

And with missconfigured I mean, that I have now a nfs daemon on node 1 and the 
orchestrator wants to create another one on node 1 but with wrong settings (it 
fails).
Also a "ceph orch apply nfs –unconfigured" does not work, so I can't manually 
manage the nfs containers.

Is there a manual way to tell ceph orch, to not create or remove nfs daemons? 
then I would be able to set them up manually.
Or a manual way of configuring the orchestrator so it does the right thing.


Thanks in advance

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
hosttech GmbH | Simon Sutter
hosttech.ch<https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Octopus: orchestrator not working correctly with nfs

2020-06-11 Thread Simon Sutter
Hello,


Well, yes and no, In the stability section 
(https://docs.ceph.com/docs/octopus/cephadm/stability/) there is written, it's 
still under development.

But in the set-up docs it's written whitout any hint of development: 
https://docs.ceph.com/docs/octopus/cephadm/install/#deploying-nfs-ganesha.

If the orchestrator is not working, then this would not be a problem for me, 
because I could deploy them manually.
But at the moment, the orchestrator is trying to deploy daemons and I don't 
know how to stop it.


Regards

Simon


Von: Amudhan P 
Gesendet: Donnerstag, 11. Juni 2020 09:32:36
An: Simon Sutter
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Octopus: orchestrator not working correctly with 
nfs

Hi,

I have not worked with orchestrator but I remember I read somewhere that NFS 
implementation is not supported.

Refer Cephadm documentation and for NFS you have configure nfs Ganesha.

You can manage NFS thru dashboard but for that you have initial config in 
dashboard and in nfsganaesha you have refer it.

Regards
Amudhan

On Thu 11 Jun, 2020, 11:40 AM Simon Sutter, 
mailto:ssut...@hosttech.ch>> wrote:
Hello,


Did I not provide enough information, or simply nobody knows how to solve the 
problem?
Should I write to the ceph tracker or does this just produce unnecessary 
overhead?


Thanks in advance,

Simon

________
Von: Simon Sutter mailto:ssut...@hosttech.ch>>
Gesendet: Montag, 8. Juni 2020 10:56:00
An: ceph-users@ceph.io<mailto:ceph-users@ceph.io>
Betreff: [ceph-users] Octopus: orchestrator not working correctly with nfs

Hello


I know that nfs on octopus is still a bit under development.

I'm trying to deploy nfs daemons and have some issues with the orchestartor.

For the other daemons, for example monitors, I can issue the command "ceph orch 
apply mon 3"

This will tell the orchestrator to deploy or remove monitor daemons until there 
are three of them.

The command does not work with nfs, and now the orchestrator is a bit 
missconfigured...

And with missconfigured I mean, that I have now a nfs daemon on node 1 and the 
orchestrator wants to create another one on node 1 but with wrong settings (it 
fails).
Also a "ceph orch apply nfs –unconfigured" does not work, so I can't manually 
manage the nfs containers.

Is there a manual way to tell ceph orch, to not create or remove nfs daemons? 
then I would be able to set them up manually.
Or a manual way of configuring the orchestrator so it does the right thing.


Thanks in advance

Simon
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
hosttech GmbH | Simon Sutter
hosttech.ch<http://hosttech.ch><https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch><http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Octopus: orchestrator not working correctly with nfs

2020-06-11 Thread Simon Sutter
Hello,


You just copied the same message.

I'll make a ticket in the tracker.


Regards,

Simon


Von: Amudhan P 
Gesendet: Donnerstag, 11. Juni 2020 09:32:36
An: Simon Sutter
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Re: Octopus: orchestrator not working correctly with 
nfs

Hi,

I have not worked with orchestrator but I remember I read somewhere that NFS 
implementation is not supported.

Refer Cephadm documentation and for NFS you have configure nfs Ganesha.

You can manage NFS thru dashboard but for that you have initial config in 
dashboard and in nfsganaesha you have refer it.

Regards
Amudhan

On Thu 11 Jun, 2020, 11:40 AM Simon Sutter, 
mailto:ssut...@hosttech.ch>> wrote:
Hello,


Did I not provide enough information, or simply nobody knows how to solve the 
problem?
Should I write to the ceph tracker or does this just produce unnecessary 
overhead?


Thanks in advance,

Simon

____
Von: Simon Sutter mailto:ssut...@hosttech.ch>>
Gesendet: Montag, 8. Juni 2020 10:56:00
An: ceph-users@ceph.io<mailto:ceph-users@ceph.io>
Betreff: [ceph-users] Octopus: orchestrator not working correctly with nfs

Hello


I know that nfs on octopus is still a bit under development.

I'm trying to deploy nfs daemons and have some issues with the orchestartor.

For the other daemons, for example monitors, I can issue the command "ceph orch 
apply mon 3"

This will tell the orchestrator to deploy or remove monitor daemons until there 
are three of them.

The command does not work with nfs, and now the orchestrator is a bit 
missconfigured...

And with missconfigured I mean, that I have now a nfs daemon on node 1 and the 
orchestrator wants to create another one on node 1 but with wrong settings (it 
fails).
Also a "ceph orch apply nfs –unconfigured" does not work, so I can't manually 
manage the nfs containers.

Is there a manual way to tell ceph orch, to not create or remove nfs daemons? 
then I would be able to set them up manually.
Or a manual way of configuring the orchestrator so it does the right thing.


Thanks in advance

Simon
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
hosttech GmbH | Simon Sutter
hosttech.ch<http://hosttech.ch><https://www.hosttech.ch>

WE LOVE TO HOST YOU.

create your own website!
more information & online-demo: 
www.website-creator.ch<http://www.website-creator.ch><http://www.website-creator.ch<http://www.website-creator.ch<http://www.website-creator.ch>>
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Error initializing cluster client: ObjectNotFound('RADOS object not found (error calling conf_read_file)')

2020-06-15 Thread Simon Sutter
Hello,


When you deploy ceph to other nodes with the orchestrator, they "just" have the 
containers you deployed to them.
This means in your case, you started the monitor container on ceph101 and you 
must have installed at least the ceph-common package (else the ceph command 
would not work).

If you enter the command ceph -s there is no configuration file, nor is there a 
key file (ceph does not know, where to connect).
Ceph's configuration directory defaults to /etc/ceph/ (should be empty or not 
exist on ceph101).

So in your case, you can either create the configuration files manually (read 
throug "ceph auth" and how the config and the keyring file should look like), 
or just copy the ceph.conf and the admin keyring to /etc/ceph/ on ceph101.


Regards,

Simon


Von: cemzafer 
Gesendet: Montag, 15. Juni 2020 21:27:30
An: ceph-us...@ceph.com
Betreff: [ceph-users] Error initializing cluster client: ObjectNotFound('RADOS 
object not found (error calling conf_read_file)')

I have installed  simple ceph system with two nodes (ceph100, ceph101)
with cephadm and ceph orch host add command. I put the ssh-copy-id -f -i
/etc/ceph/ceph.pub key to the second host (ceph101). I can execute the
ceph -s command from the first host(ceph100) but when I execute the
command in the second host(ceph101), I get the following error.

Error initializing cluster client: ObjectNotFound('RADOS object not
found (error calling conf_read_file)')

Also, when I execute the 'ceph orch ps' command the output seems
suspicious to me.

NAME HOST   STATUSREFRESHED AGE
VERSIONIMAGE NAMEIMAGE ID CONTAINER ID

mon.ceph101ceph101  starting -  -


Has anyone any idea what could be the problem or anyone give me a fine
link for the octopus cephadm installation?

Regards.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Current status of multipe cephfs

2020-06-16 Thread Simon Sutter
Hello,


What is the current status, of using multiple cephfs?
In octopuss I get lots of warnings, that this feature is still not fully 
tested, but the latest entry regarding multiple cephfs in the mailinglist is 
from about 2018.

Is someone using multiple cephfs in production?


Thanks in Advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Orchestrator: Cannot add node after mistake

2020-06-18 Thread Simon Sutter
Hello,


I did a mistake, while deploying a new node on octopus.
The node is a fresh installed Centos8 machine.
Bevore I did a "ceph orch host add node08" I pasted the wrong command:

ceph orch daemon add osd node08:cl_node08/ceph

That did not return anything, so I tried to add the node first with the host 
add command, but now I get an error:
Error ENOENT: New host node08 (node08) failed check: ['Traceback (most recent 
call last):', '  File "", line 4580, in ', '  File "", 
line 3592, in command_check_host', "UnboundLocalError: local variable 
'container_path' referenced before assignment"]


I'm not a developer, so I don't know where to look, and how to fix this.

I tried to reboot every node, to see if it's just a cached problem, but no luck 
there.
Do any of you know how to fix this?


Thanks in advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Orchestrator: Cannot add node after mistake

2020-06-18 Thread Simon Sutter
After some hours of searching around in the docs and finaly taking a look at 
the sourcecode of cephadm, I figured out it has to do with the docker software 
(podman in my case).

So I figured out that podman is not installed on the new node.
I never installed podman on any node so I don't know if my installation was not 
good, or if cephadm just had a problem installing it.

Also the message in the error is not very meaningful.

For example, if I forget to install python3, the "python3 is not installed" 
message is very clear.


Simon

________
Von: Simon Sutter 
Gesendet: Donnerstag, 18. Juni 2020 14:09:18
An: ceph-users@ceph.io
Betreff: [ceph-users] Orchestrator: Cannot add node after mistake

Hello,


I did a mistake, while deploying a new node on octopus.
The node is a fresh installed Centos8 machine.
Bevore I did a "ceph orch host add node08" I pasted the wrong command:

ceph orch daemon add osd node08:cl_node08/ceph

That did not return anything, so I tried to add the node first with the host 
add command, but now I get an error:
Error ENOENT: New host node08 (node08) failed check: ['Traceback (most recent 
call last):', '  File "", line 4580, in ', '  File "", 
line 3592, in command_check_host', "UnboundLocalError: local variable 
'container_path' referenced before assignment"]


I'm not a developer, so I don't know where to look, and how to fix this.

I tried to reboot every node, to see if it's just a cached problem, but no luck 
there.
Do any of you know how to fix this?


Thanks in advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: How to ceph-volume on remote hosts?

2020-06-24 Thread Simon Sutter
Hello,

If you do it, like Sebastian told you, you would automatically deploy osd's.
For a beginner I would recommend to do it semi-automated, so you know a bit 
more what's going on.

So first do the "ceph orch device ls" which should print every disk, in all 
nodes.
Then I recommend to zap devices you want to use first, even if it says you can 
use them:
ceph orch device zap {node} /dev/sd? --force

And then you can add devices one by one:
ceph orch daemon add osd {node}:/dev/sd?

Just to top it off.

Regards,
Simon

-Ursprüngliche Nachricht-
Von: Sebastian Wagner [mailto:swag...@suse.com] 
Gesendet: Mittwoch, 24. Juni 2020 14:06
An: steven prothero ; ceph-users@ceph.io
Betreff: [ceph-users] Re: How to ceph-volume on remote hosts?



Am 24.06.20 um 05:15 schrieb steven prothero:
> Hello,
> 
> I am new to CEPH and on a few test servers attempting to setup and 
> learn a test ceph system.
> 
> I started off the install with the "Cephadm" option and it uses podman 
> containers.
> Followed steps here:
> https://docs.ceph.com/docs/master/cephadm/install/
> 
> I ran the bootstrap, added remote hosts, added monitors and everything 
> is looking good.
> 
> Now I would like to add OSDs...
> 
> On the bootstrapped server i did a :
> 
> ceph-volume lvm prepare   --data /dev/sda6
>and then the "activate" and "ceph orch daemon add osd (etc)" to add 
> it and it works...
> 
> But now I am ready to add OSDs on the remote nodes. I am not able to 
> find documentation or examples on how to do :
> 
>   ceph-volume lvm prepare & activate steps on the remote hosts.
> 
> How do we prepare & activate the remote hosts disks?

ceph orch apply osd

as described in
https://docs.ceph.com/docs/master/cephadm/install/#deploy-osds

should do the trick. In case it doesn't, what's the output of

ceph orch device ls

?

> 
> Thank you very much for your input,
> 
> Cheers
> Steve
> ___
> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an 
> email to ceph-users-le...@ceph.io
> 

--
SUSE Software Solutions Germany GmbH, Maxfeldstr. 5, 90409 Nürnberg, Germany 
(HRB 36809, AG Nürnberg). Geschäftsführer: Felix Imendörffer

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Feedback of the used configuration

2020-06-24 Thread Simon Sutter
Hello,

After two months of the "ceph try and error game", I finally managed to get an 
Octopuss cluster up and running.
The unconventional thing about it is, it's just for hot backups, no virtual 
machines on there.
All the  nodes are without any caching ssd's, just plain hdd's.
At the moment there are eight of them with a total of 50TB. We are planning to 
go up to 25 and bigger disks so we end on 300TB-400TB

I decided to go with cephfs, because I don't have any experience in things like 
S3 and I need to read the same file system from more than one client.

I made one cephfs with a replicated pool.
On there I added erasure-coded pools to save some Storage.
To add those pools, I did it with the setfattr command like this:
setfattr -n ceph.dir.layout.pool -v ec_data_server1 /cephfs/nfs/server1

Some of our servers cannot use cephfs (old kernels, special OS's) so I have to 
use nfs.
This is set up with the included ganesha-nfs.
Exported is the /cephfs/nfs folder and clients can mount folders below this.

There are two final questions:

-  Was it right to go with the way of "mounting" pools with setfattr, 
or should I have used multiple cephfs?

First I was thinking about using multiple cephfs but there are warnings 
everywhere. The deeper I got in, the more it seems I would have been fine with 
multiple cephfs.

-  Is there a way I don't know, but it would be easier?

I still don't know much about Rest, S3, RBD etc... so there may be a better way

Other remarks are desired.

Thanks in advance,
Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Feedback of the used configuration

2020-06-25 Thread Simon Sutter
Hello Paul,

Thanks for the Answer.
I took a look at the subvolumes, but they are a bit odd in my opinion.
If I create one with a subvolume-group, the folder structure will look like 
this:
/cephfs/volumes/group-name/subvolume-name/random-uuid/
And I have to issue two commands, first set the group and then set the volume 
name, but why so complicated?

Wouldn't it be easier to  just make subvolumes anywhere inside the cephfs?
I can see the intended use for groups, but if I want to publish a pool in some 
different directory that's not possible (except for setfattr).
Without first creating subvolume-groups, the orchestrator creates subvolumes in 
the /cephfs/volumes/_nogroup/subvolume-name/randmon-uuid/ folder.

And the more important question is, why is there a new folder with a random 
uuid inside the subvolume?
I try to understand the points the devs had, when they developed this, but for 
me, this is something I have to explain to some devs in our team and at the 
moment I can't.

It is indeed easier to deploy but comes with much less flexibility.
Maybe something to write in the tracker about?

Thanks in advance,
Simon

Von: Paul Emmerich [mailto:paul.emmer...@croit.io]
Gesendet: Mittwoch, 24. Juni 2020 17:35
An: Simon Sutter 
Cc: ceph-users@ceph.io
Betreff: Re: [ceph-users] Feedback of the used configuration

Have a look at cephfs subvolumes: 
https://docs.ceph.com/docs/master/cephfs/fs-volumes/#fs-subvolumes

They are internally just directories with quota/pool placement layout/namespace 
with some mgr magic to make it easier than doing that all by hand

Paul

--
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io<http://www.croit.io>
Tel: +49 89 1896585 90


On Wed, Jun 24, 2020 at 4:38 PM Simon Sutter 
mailto:ssut...@hosttech.ch>> wrote:
Hello,

After two months of the "ceph try and error game", I finally managed to get an 
Octopuss cluster up and running.
The unconventional thing about it is, it's just for hot backups, no virtual 
machines on there.
All the  nodes are without any caching ssd's, just plain hdd's.
At the moment there are eight of them with a total of 50TB. We are planning to 
go up to 25 and bigger disks so we end on 300TB-400TB

I decided to go with cephfs, because I don't have any experience in things like 
S3 and I need to read the same file system from more than one client.

I made one cephfs with a replicated pool.
On there I added erasure-coded pools to save some Storage.
To add those pools, I did it with the setfattr command like this:
setfattr -n ceph.dir.layout.pool -v ec_data_server1 /cephfs/nfs/server1

Some of our servers cannot use cephfs (old kernels, special OS's) so I have to 
use nfs.
This is set up with the included ganesha-nfs.
Exported is the /cephfs/nfs folder and clients can mount folders below this.

There are two final questions:

-  Was it right to go with the way of "mounting" pools with setfattr, 
or should I have used multiple cephfs?

First I was thinking about using multiple cephfs but there are warnings 
everywhere. The deeper I got in, the more it seems I would have been fine with 
multiple cephfs.

-  Is there a way I don't know, but it would be easier?

I still don't know much about Rest, S3, RBD etc... so there may be a better way

Other remarks are desired.

Thanks in advance,
Simon
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Octopus Grafana inside the dashboard

2020-06-29 Thread Simon Sutter
Hello

I'm trying to get Grafana working inside the Dashboard.
If I press on "Overall Performance" tab, I get an error, because the iframe 
tries to connect to the internal hostname, which cannot be resolved from my 
machine.
If I directly open grafana, everything works.
How can I tell the dashboard, to use the full domain name?
I have tried to set "ceph dashboard set-grafana-api-url  
https://node01.mycorp.local:3000"; but that does not work and always sets itself 
back to "https://node01:3000";.

Thanks in advance,
Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] cephfs needs access from two networks

2020-08-26 Thread Simon Sutter
Hello,

So I know, the mon services can only bind to just one ip.
But I have to make it accessible to two networks because internal and external 
servers have to mount the cephfs.
The internal ip is 10.99.10.1 and the external is some public-ip.
I tried nat'ing it  with this: "firewall-cmd --zone=public 
--add-forward-port=port=6789:proto=tcp:toport=6789:toaddr=10.99.10.1 -permanent"

So the nat is working, because I get a "ceph v027" (alongside with some 
gibberish) when I do a telnet "telnet *public-ip* 6789"
But when I try to mount it, I get just a timeout:
mount - -t ceph *public-ip*:6789:/testing /mnt -o 
name=test,secretfile=/root/ceph.client. test.key
mount error 110 = Connection timed out

The tcpdump also recognizes a "Ceph Connect" packet, coming from the mon.

How can I get around this problem?
Is there something I have missed?

Specs:
Latest Octopus 15.2.4
Centos 8
8 Nodes
No health warnings.

Thanks in advance,
Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: cephfs needs access from two networks

2020-08-31 Thread Simon Sutter
Hello again

So I have changed the network configuration.
Now my Ceph is reachable from outside, this also means all osd’s of all nodes 
are reachable.
I still have the same behaviour which is a timeout.

The client can resolve all nodes with their hostnames.
The mon’s are still listening on the internal network so the nat rule is still 
there.
I have set “public bind addr” to the external ip and restarted the mon but it’s 
still not working.

[root@testnode1 ~]# ceph config get mon.public_bind_addr
WHO MASK  LEVEL OPTIONVALUERO
mon   advanced  public_bind_addr  v2:[ext-addr]:0/0 *

Do I have to change them somewhere else too?

Thanks in advance,
Simon


Von: Janne Johansson [mailto:icepic...@gmail.com]
Gesendet: 27 August 2020 20:01
An: Simon Sutter 
Betreff: Re: [ceph-users] cephfs needs access from two networks

Den tors 27 aug. 2020 kl 12:05 skrev Simon Sutter 
mailto:ssut...@hosttech.ch>>:
Hello Janne

Oh I missed that point. No, the client cannot talk directly to the osds.
In this case it’s extremely difficult to set this up.

This is an absolute requirement to be a ceph client.

How is the mon telling the client, which host and port of the osd, it should 
connect to?

The same port and ip that the ODS called into the mon with when it started up 
and joined the clusster.

Can I have an influence on it?


Well, you set the ip on the OSD hosts, and the port ranges in use for OSDs are 
changeable/settable, but it would not really help the above-mentioned client.

Von: Janne Johansson [mailto:icepic...@gmail.com<mailto:icepic...@gmail.com>]
Gesendet: 26 August 2020 15:09
An: Simon Sutter mailto:ssut...@hosttech.ch>>
Cc: ceph-users@ceph.io<mailto:ceph-users@ceph.io>
Betreff: Re: [ceph-users] cephfs needs access from two networks

Den ons 26 aug. 2020 kl 14:16 skrev Simon Sutter 
mailto:ssut...@hosttech.ch>>:
Hello,
So I know, the mon services can only bind to just one ip.
But I have to make it accessible to two networks because internal and external 
servers have to mount the cephfs.
The internal ip is 10.99.10.1 and the external is some public-ip.
I tried nat'ing it  with this: "firewall-cmd --zone=public 
--add-forward-port=port=6789:proto=tcp:toport=6789:toaddr=10.99.10.1 -permanent"

So the nat is working, because I get a "ceph v027" (alongside with some 
gibberish) when I do a telnet "telnet *public-ip* 6789"
But when I try to mount it, I get just a timeout:
mount - -t ceph *public-ip*:6789:/testing /mnt -o 
name=test,secretfile=/root/ceph.client. test.key
mount error 110 = Connection timed out

The tcpdump also recognizes a "Ceph Connect" packet, coming from the mon.

How can I get around this problem?
Is there something I have missed?

Any ceph client will need direct access to all OSDs involved also. Your mail 
doesn't really say if the cephfs-mounting client can talk to OSDs?

In ceph, traffic is not shuffled via mons, mons only tell the client which OSDs 
it needs to talk to, then all IO goes directly from client to any involved OSD 
servers.

--
May the most significant bit of your life be positive.


--
May the most significant bit of your life be positive.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Syncing cephfs from Ceph to Ceph

2020-09-08 Thread Simon Sutter
Hello,


Is it possible to somehow sync a ceph from one site to a ceph form another site?
I'm just using the cephfs feature and no block devices.

Being able to sync cephfs pools between two sites would be great for a hot 
backup, in case one site fails.


Thanks in advance,

Simon
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Syncing cephfs from Ceph to Ceph

2020-09-08 Thread Simon Sutter
Thanks Stefan,

First of all, for a bit more context, we use this ceph cluster just for hot 
backups, so 99% write 1% read, no need for low latency.

Ok so the snapshot function would mean, we would have like a colder backup.
Just like a snapshot of a VM, without any incremental functionality, which also 
means scheduled but huge transfer rates.

What about the idea of creating the cluster over two data centers?
Would it be possible to modify the crush map, so one pool gets replicated over 
those two data centers and if one fails, the other one would still be 
functional?
Additionally, would it be possible to prioritize one data center over the other?
This would allow saving data from site1 to a pool on site2 in case of a 
disaster on site1, site2 would still have those Backups.

We have a 10G connection with around 0.5ms latency.


Thanks in advance,
Simon



Von: Stefan Kooman 
Gesendet: Dienstag, 8. September 2020 11:38:29
An: Simon Sutter; ceph-users@ceph.io
Betreff: Re: [ceph-users] Syncing cephfs from Ceph to Ceph

On 2020-09-08 11:22, Simon Sutter wrote:
> Hello,
>
>
> Is it possible to somehow sync a ceph from one site to a ceph form another 
> site?
> I'm just using the cephfs feature and no block devices.
>
> Being able to sync cephfs pools between two sites would be great for a hot 
> backup, in case one site fails.

It's a work in progress [1]. This might do what you want to right now:
[2]. Note: I haven't used [2] myself.

Gr. Stefan

[1]: https://docs.ceph.com/docs/master/dev/cephfs-mirroring/
[2]: https://github.com/oliveiradan/cephfs-sync
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] New to ceph / Very unbalanced cluster

2020-04-15 Thread Simon Sutter
Hello everybody,



I'm very new to ceph and installed a testenvironment (nautilus).

The current goal of this cluster is, to be a short period backup.

For this goal we want to use older, mixed hardware, so I was thinking, for 
testing I will set up very unbalanced nodes (you can learn the most, from 
exceptional circumstances, right?).

I created for my cephfs two pools, one for metadata and one for storage data.



I have three nodes and the ceph osd tree looks like this:
ID  CLASS WEIGHT   TYPE NAME  STATUS REWEIGHT PRI-AFF
 -1  21.54213root default
 -30.75679 host node1
  0   hdd  0.75679 osd.0  up  0.00636 1.0
 -5   5.39328 host node2
  1   hdd  2.66429 osd.1  up  0.65007 1.0
  3   hdd  2.72899 osd.3  up  0.65007 1.0
-1015.39206 host node3
  5   hdd  7.27739 osd.5  up  1.0 1.0
  6   hdd  7.27739 osd.6  up  1.0 1.0
  2   ssd  0.38249 osd.2  up  1.0 1.0
  4   ssd  0.45479 osd.4  up  1.0 1.0


The PGs and thus the data is extremely unbalanced, you can see it in the ceph 
osd df overview:
ID CLASS WEIGHT  REWEIGHT SIZERAW USE DATAOMAPMETA AVAIL   %USE 
 VAR  PGS STATUS
 0   hdd 0.75679  0.00636 775 GiB 651 GiB 650 GiB  88 KiB 1.5 GiB 124 GiB   
84.02   7.26 112 up
 1   hdd 2.66429  0.65007   2.7 TiB 497 GiB 496 GiB  88 KiB 1.2 GiB   2.2 
TiB   18.22   1.57  81 up
 3   hdd 2.72899  0.65007   2.7 TiB 505 GiB 504 GiB8 KiB 1.3 GiB   2.2 
TiB   18.07   1.56  88 up
 5   hdd 7.27739  1.0   7.3 TiB 390 GiB 389 GiB8 KiB 1.2 GiB   6.9 
TiB 5.24   0.45  67 up
 6   hdd 7.27739  1.0   7.3 TiB 467 GiB 465 GiB  64 KiB 1.3 GiB   6.8 
TiB 6.26   0.54  78 up
 2   ssd 0.38249   1.0  392 GiB   14 GiB   13 GiB  11 KiB 1024 MiB 377 GiB  
   3.68   0.32   2 up
 4   ssd 0.45479   1.0  466 GiB   28 GiB   27 GiB4 KiB 1024 MiB 438 GiB 
6.030.52   4 up
TOTAL  22 TiB 2.5 TiB 2.5 TiB 273 KiB  8.4 GiB  19 TiB 11.57
MIN/MAX VAR: 0.32/7.26  STDDEV: 6.87

To counteract this, I tried to turn on the balancer module.

The module is decreasing the reweight of the osd0 more and more, while ceph pg 
stat is telling me, there are more misplaced objects:

144 pgs: 144 active+clean+remapped; 853 GiB data, 2.5 TiB used, 19 TiB / 22 TiB 
avail; 30 MiB/s wr, 7 op/s; 242259/655140 objects misplaced (36.978%)



So my question is: is ceph supposed to do that?
Why are all those objects misplaced? Because of those 112 PGs on osd0?
Why are there 112 PGs on osd0? I did not set any pg settings except the number: 
512



Thank you very much
Simon Sutter
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: New to ceph / Very unbalanced cluster

2020-04-16 Thread Simon Sutter
Thank you very much, I couldn't see the forest for the trees.

Now I moved a disk and added another one, now the problem is gone, I have 8TB 
to use.


Thanks again.

Simon Sutter


Von: Reed Dier 
Gesendet: Mittwoch, 15. April 2020 22:59:12
An: Simon Sutter
Cc: ceph-users
Betreff: Re: [ceph-users] New to ceph / Very unbalanced cluster

Copying the ML, because I forgot to reply-all.

Reed

On Apr 15, 2020, at 3:58 PM, Reed Dier 
mailto:reed.d...@focusvq.com>> wrote:

The problem is almost certainly stemming from unbalanced OSD distribution among 
your hosts, and assuming you are using a default 3x replication across hosts 
crush rule set.

You are limited by your smallest bin size.

In this case you have a 750GB HDD as the only OSD on node1, so when it wants 3 
copies across 3 hosts, there are only ~750GB of space that can fulfill this 
requirement.

Having lots of different size OSDs and differing OSDs in your topology is going 
to lead to issues of under/over utilization.

ID  CLASS WEIGHT   TYPE NAME  STATUS REWEIGHT PRI-AFF
-1  21.54213root default
-3  0.75679 host node1
-5  5.39328 host node2
-1015.39206 host node3

You either need to redistribute your OSDs across your hosts, or possibly 
rethink your disk strategy.
You could move osd.5 to node1, and osd.0 to node2, which would give you roughly 
6TiB of usable hdd space across your three nodes.

Reed

On Apr 15, 2020, at 10:50 AM, Simon Sutter 
mailto:ssut...@hosttech.ch>> wrote:

Hello everybody,



I'm very new to ceph and installed a testenvironment (nautilus).

The current goal of this cluster is, to be a short period backup.

For this goal we want to use older, mixed hardware, so I was thinking, for 
testing I will set up very unbalanced nodes (you can learn the most, from 
exceptional circumstances, right?).

I created for my cephfs two pools, one for metadata and one for storage data.



I have three nodes and the ceph osd tree looks like this:
ID  CLASS WEIGHT   TYPE NAME  STATUS REWEIGHT PRI-AFF
-1  21.54213root default
-30.75679 host node1
 0   hdd  0.75679 osd.0  up  0.00636 1.0
-5   5.39328 host node2
 1   hdd  2.66429 osd.1  up  0.65007 1.0
 3   hdd  2.72899 osd.3  up  0.65007 1.0
-1015.39206 host node3
 5   hdd  7.27739 osd.5  up  1.0 1.0
 6   hdd  7.27739 osd.6  up  1.0 1.0
 2   ssd  0.38249 osd.2  up  1.0 1.0
 4   ssd  0.45479 osd.4  up  1.0 1.0


The PGs and thus the data is extremely unbalanced, you can see it in the ceph 
osd df overview:
ID CLASS WEIGHT  REWEIGHT SIZERAW USE DATAOMAPMETA AVAIL   %USE 
 VAR  PGS STATUS
0   hdd 0.75679  0.00636 775 GiB 651 GiB 650 GiB  88 KiB 1.5 GiB 124 GiB   
84.02   7.26 112 up
1   hdd 2.66429  0.65007   2.7 TiB 497 GiB 496 GiB  88 KiB 1.2 GiB   2.2 
TiB   18.22   1.57  81 up
3   hdd 2.72899  0.65007   2.7 TiB 505 GiB 504 GiB8 KiB 1.3 GiB   2.2 
TiB   18.07   1.56  88 up
5   hdd 7.27739  1.0   7.3 TiB 390 GiB 389 GiB8 KiB 1.2 GiB   6.9 
TiB 5.24   0.45  67 up
6   hdd 7.27739  1.0   7.3 TiB 467 GiB 465 GiB  64 KiB 1.3 GiB   6.8 
TiB 6.26   0.54  78 up
2   ssd 0.38249   1.0  392 GiB   14 GiB   13 GiB  11 KiB 1024 MiB 377 GiB   
  3.68   0.32   2 up
4   ssd 0.45479   1.0  466 GiB   28 GiB   27 GiB4 KiB 1024 MiB 438 GiB  
   6.030.52   4 up
   TOTAL  22 TiB 2.5 TiB 2.5 TiB 273 KiB  8.4 GiB  19 TiB 11.57
MIN/MAX VAR: 0.32/7.26  STDDEV: 6.87

To counteract this, I tried to turn on the balancer module.

The module is decreasing the reweight of the osd0 more and more, while ceph pg 
stat is telling me, there are more misplaced objects:

144 pgs: 144 active+clean+remapped; 853 GiB data, 2.5 TiB used, 19 TiB / 22 TiB 
avail; 30 MiB/s wr, 7 op/s; 242259/655140 objects misplaced (36.978%)



So my question is: is ceph supposed to do that?
Why are all those objects misplaced? Because of those 112 PGs on osd0?
Why are there 112 PGs on osd0? I did not set any pg settings except the number: 
512



Thank you very much
Simon Sutter
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io