[ceph-users] Re: Diskprediction_local mgr module removal - Call for feedback

2025-04-08 Thread Michal Strnad

Hi.

From our point of view, it's important to keep disk failure prediction 
tool as part of Ceph, ideally as an MGR module. In environments with 
hundreds or thousands of disks, it's crucial to know whether, for 
example, a significant number of them are likely to fail within a month 
- which, in the best-case scenario, would mean performance degradation, 
and in the worst-case, data loss.


Some have already responded to the deprecation of diskprediction by 
starting to develop their own solutions. For instance, just yesterday, 
Daniel Persson published a solution [1] on his website that addresses 
the same problem.


Would it be possible to join forces and try to revive that module?

[1] https://www.youtube.com/watch?v=Gr_GtC9dcMQ

Thanks,
Michal


On 4/8/25 01:18, Yaarit Hatuka wrote:

Hi everyone,

On today's Ceph Steering Committee call we discussed the idea of removing
the diskprediction_local mgr module, as the current prediction model is
obsolete and not maintained.

We would like to gather feedback from the community about the usage of this
module, and find out if anyone is interested in maintaining it.

Thanks,
Yaarit
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io





smime.p7s
Description: S/MIME Cryptographic Signature
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Eugen Block

These are your two Luminous clients:

---snip---
{
"name": "unknown.0",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
}
]
},
"socket_addr": {
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
...

{
"name": "client.104098",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
}
]
},
"socket_addr": {
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
---snip---

Zitat von quag...@bol.com.br:


Hi Eugen!  Thanks a lot!   I was able to find luminous connections,
but I still can't identify which client process. Here is the output:
Rafael.
──
 De: "Eugen Block"  Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
Assunto:  [ceph-users] Re: Ceph squid fresh install   Hi,  you can query
the MON sessions to identify your older clients with:  ceph tell mon.
sessions  It will show you the IP address, con_features_release (Luminous)
and a couple of other things.  Zitat von Laura Flores :  > Hi Rafael, >> I
would not force the min_compat_client to be reef when there are still >
luminous clients connected, as it is important for all clients to be >=Reef

to understand/encode the pg_upmap_primary feature in the osdmap. >> As

for checking which processes are still luminous, I am copying @Radoslaw >
Zarzynski  who may be able to help more with that. >> Thanks, > Laura
Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br  > wrote:

Hi, >> I just did a new Ceph installation and would like to enable the

"read >> balancer". >> However, the documentation requires that the minimum
client version >> be reef. I checked this information through "ceph
features" and came across >> the situation of having 2 luminous clients. >>
# ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",

"release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>

"features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }

], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":

"squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
"0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
"features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }

], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":

"squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
version to reef and received the >> following alert: >> # ceph osd
set-require-min-compat-client reef >> Error EPERM: cannot set
require_min_compat_client to reef: 2 connected >> client(s) look like
luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
anyway  Is it ok do confirm anyway? >> Which processes are still as
luminous?  Rafael. >> ___

ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an

email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> She/Her/Hers

Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |

lflo...@redhat.com  > M: +17087388804 >
___ > ceph-users mailing list
-- ceph-users@ceph.io > To unsubscribe send an email to
ceph-users-le...@ceph.io   ___
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email
to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Anthony D'Atri


> On Apr 8, 2025, at 9:13 AM, quag...@bol.com.br wrote:
> 
> These 2 IPs are from the storage servers.

What is a “storage server”?


> There are no user processes running on them. It only has the operating system 
> and ceph installed.

Nobody said anything about user processes.

> 
> 
> Rafael.
> 
> De: "Eugen Block" 
> Enviada: 2025/04/08 09:35:35
> Para: quag...@bol.com.br
> Cc: ceph-users@ceph.io
> Assunto: Re: [ceph-users] Ceph squid fresh install
>  
> These are your two Luminous clients:
> 
> ---snip---
> {
> "name": "unknown.0",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> }
> ]
> },
> "socket_addr": {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ...
> 
> {
> "name": "client.104098",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> }
> ]
> },
> "socket_addr": {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ---snip---
> 
> Zitat von quag...@bol.com.br:
> 
> > Hi Eugen! Thanks a lot! I was able to find luminous connections,
> > but I still can't identify which client process. Here is the output:
> > Rafael.
> > ──
> > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
> > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> > the MON sessions to identify your older clients with: ceph tell mon.
> > sessions It will show you the IP address, con_features_release (Luminous)
> > and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
> > would not force the min_compat_client to be reef when there are still >
> > luminous clients connected, as it is important for all clients to be >=Reef
> >> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
> > for checking which processes are still luminous, I am copying @Radoslaw >
> > Zarzynski who may be able to help more with that. >> Thanks, > Laura
> > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
>  Hi, >> I just did a new Ceph installation and would like to enable the
> > "read >> balancer". >> However, the documentation requires that the minimum
> > client version >> be reef. I checked this information through "ceph
> > features" and came across >> the situation of having 2 luminous clients. >>
> > # ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",
> >>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
> >>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
> > "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
> >>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
> > version to reef and received the >> following alert: >> # ceph osd
> > set-require-min-compat-client reef >> Error EPERM: cannot set
> > require_min_compat_client to reef: 2 connected >> client(s) look like
> > luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
> > anyway  Is it ok do confirm anyway? >> Which processes are still as
> > luminous?  Rafael. >> ___
> >>> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an
> > email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> She/Her/Hers
> >>> Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |
> > lflo...@redhat.com > M: +17087388804 >
> > ___ > ceph-users mailing list
> > -- ceph-users@ceph.io > To unsubscribe send an email to
> > ceph-users-le...@ceph.io ___
> > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email
> > to ceph-users-le...@ceph.io
> 
> 
>  ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Diskprediction_local mgr module removal - Call for feedback

2025-04-08 Thread Anthony D'Atri


> Hi.
> 
> From our point of view, it's important to keep disk failure prediction tool 
> as part of Ceph, ideally as an MGR module


> . In environments with hundreds or thousands of disks, it's crucial to know 
> whether, for example, a significant number of them are likely to fail within 
> a month - which, in the best-case scenario, would mean performance 
> degradation, and in the worst-case, data loss.

Agreed, but the current implementation does not achieve this.

> Some have already responded to the deprecation of diskprediction by starting 
> to develop their own solutions. For instance, just yesterday, Daniel Persson 
> published a solution [1]

Videos and costumes aren’t solutions.


> on his website that addresses the same problem.
> 
> Would it be possible to join forces

Are you volunteering to code?  Can you sign up to be involved going forward?  
Ceph has suffered time and again from something being partly implemented, only 
to have the sole proponent wander away.

> and try to revive that module?

It really needs to be reimplemented in a less opaque way so that admins can 
contribute and tweak.  No binary blobs.

> 
> [1] https://www.youtube.com/watch?v=Gr_GtC9dcMQ
> 
> Thanks,
> Michal
> 
> 
> On 4/8/25 01:18, Yaarit Hatuka wrote:
>> Hi everyone,
>> On today's Ceph Steering Committee call we discussed the idea of removing
>> the diskprediction_local mgr module, as the current prediction model is
>> obsolete and not maintained.
>> We would like to gather feedback from the community about the usage of this
>> module, and find out if anyone is interested in maintaining it.
>> Thanks,
>> Yaarit
>> ___
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
> 
> 
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
These 2 IPs are from the storage servers.
There are no user processes running on them. It only has the operating system and ceph installed.


Rafael.


De: "Eugen Block" 
Enviada: 2025/04/08 09:35:35
Para: quag...@bol.com.br
Cc:  ceph-users@ceph.io
Assunto:  Re: [ceph-users] Ceph squid fresh install
 
These are your two Luminous clients:

---snip---
{
"name": "unknown.0",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
}
]
},
"socket_addr": {
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
...

{
"name": "client.104098",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
}
]
},
"socket_addr": {
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
---snip---

Zitat von quag...@bol.com.br:

> Hi Eugen! Thanks a lot! I was able to find luminous connections,
> but I still can't identify which client process. Here is the output:
> Rafael.
> ──
> De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
> Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> the MON sessions to identify your older clients with: ceph tell mon.
> sessions It will show you the IP address, con_features_release (Luminous)
> and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
> would not force the min_compat_client to be reef when there are still >
> luminous clients connected, as it is important for all clients to be >=Reef
>> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
> for checking which processes are still luminous, I am copying @Radoslaw >
> Zarzynski who may be able to help more with that. >> Thanks, > Laura
> Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
 Hi, >> I just did a new Ceph installation and would like to enable the
> "read >> balancer". >> However, the documentation requires that the minimum
> client version >> be reef. I checked this information through "ceph
> features" and came across >> the situation of having 2 luminous clients. >>
> # ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",
>>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
> "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
>>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
> "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
> "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
> "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
>>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
> "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
> version to reef and received the >> following alert: >> # ceph osd
> set-require-min-compat-client reef >> Error EPERM: cannot set
> require_min_compat_client to reef: 2 connected >> client(s) look like
> luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
> anyway  Is it ok do confirm anyway? >> Which processes are still as
> luminous?  Rafael. >> ___
>>> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an
> email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> She/Her/Hers
>>> Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |
> lflo...@redhat.com > M: +17087388804 >
> ___ > ceph-users mailing list
> -- ceph-users@ceph.io > To unsubscribe send an email to
> ceph-users-le...@ceph.io ___
> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email
> to ceph-users-le...@ceph.io


 ___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: NIH Datasets

2025-04-08 Thread Anthony D'Atri


> Maybe one of the many existing projects could be adapted, re-formed,
> re-aimed. Or maybe they're all dead in the water because they failed
> one of more of the above five bullets.

Often the unstated sixth bullet:

— Has a supportable architecture

The ongoing viability of a system should not rely on someone constantly having 
to monitor and manage contracts with thousands of high-touch low-savvy micro 
providers.

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Eugen Block
What is the locally installed ceph version? Run 'ceph --version' to  
get the local rpm/deb package version.


Zitat von quag...@bol.com.br:


These 2 IPs are from the storage servers.
There are no user processes running on them. It only has the operating
system and ceph installed.

Rafael.

-

DE: "Eugen Block" 
ENVIADA: 2025/04/08 09:35:35
PARA: quag...@bol.com.br
CC: ceph-users@ceph.io
ASSUNTO: Re: [ceph-users] Ceph squid fresh install
 
These are your two Luminous clients:

---snip---
{
"name": "unknown.0",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
}
]
},
"socket_addr": {
"type": "none",
"addr": "172.27.254.7:0",
"nonce": 443842330
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
...

{
"name": "client.104098",
"entity_name": "client.admin",
"addrs": {
"addrvec": [
{
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
}
]
},
"socket_addr": {
"type": "v1",
"addr": "172.27.254.6:0",
"nonce": 2027668300
},
"con_type": "client",
"con_features": 3387146417253690110,
"con_features_hex": "2f018fb87aa4aafe",
"con_features_release": "luminous",
---snip---

Zitat von quag...@bol.com.br:


Hi Eugen! Thanks a lot! I was able to find luminous connections,
but I still can't identify which client process. Here is the output:
Rafael.


──

De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
the MON sessions to identify your older clients with: ceph tell mon.
sessions It will show you the IP address, con_features_release (Luminous)
and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
would not force the min_compat_client to be reef when there are still >
luminous clients connected, as it is important for all clients to be
=Reef

to understand/encode the pg_upmap_primary feature in the osdmap. >> As

for checking which processes are still luminous, I am copying @Radoslaw >
Zarzynski who may be able to help more with that. >> Thanks, > Laura
Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:

Hi, >> I just did a new Ceph installation and would like to enable the

"read >> balancer". >> However, the documentation requires that the

minimum

client version >> be reef. I checked this information through "ceph
features" and came across >> the situation of having 2 luminous clients.



# ceph features >> { >> "mon": [ >> { >> "features":

"0x3f03cffd",

"release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>

"features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }

], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":

"squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
"0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
"features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }

], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":

"squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
version to reef and received the >> following alert: >> # ceph osd
set-require-min-compat-client reef >> Error EPERM: cannot set
require_min_compat_client to reef: 2 connected >> client(s) look like
luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
anyway  Is it ok do confirm anyway? >> Which processes are still as
luminous?  Rafael. >> ___

ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an

email to ceph-users-le...@ceph.io >> -- >> Laura Flores >>

She/Her/Hers

Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |

lflo...@redhat.com > M: +17087388804 >
___ > ceph-users mailing list
-- ceph-users@ceph.io > To unsubscribe send an email to
ceph-users-le...@ceph.io ___
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an

email

to ceph-users-le...@ceph.io


 



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Laimis Juzeliūnas
Hey Rafael,

So these client mounts probably use the kernel mounts with an older kernel 
version that gets identified as the luminous client.
You can try upgrading the kernel version and remounting.

Laimis J.

> On 8 Apr 2025, at 16:13, quag...@bol.com.br wrote:
> 
> These 2 IPs are from the storage servers.
> There are no user processes running on them. It only has the operating system 
> and ceph installed.
> 
> 
> Rafael.
> 
> De: "Eugen Block" 
> Enviada: 2025/04/08 09:35:35
> Para: quag...@bol.com.br
> Cc: ceph-users@ceph.io
> Assunto: Re: [ceph-users] Ceph squid fresh install
>  
> These are your two Luminous clients:
> 
> ---snip---
> {
> "name": "unknown.0",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> }
> ]
> },
> "socket_addr": {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ...
> 
> {
> "name": "client.104098",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> }
> ]
> },
> "socket_addr": {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ---snip---
> 
> Zitat von quag...@bol.com.br:
> 
> > Hi Eugen! Thanks a lot! I was able to find luminous connections,
> > but I still can't identify which client process. Here is the output:
> > Rafael.
> > ──
> > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
> > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> > the MON sessions to identify your older clients with: ceph tell mon.
> > sessions It will show you the IP address, con_features_release (Luminous)
> > and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
> > would not force the min_compat_client to be reef when there are still >
> > luminous clients connected, as it is important for all clients to be >=Reef
> >> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
> > for checking which processes are still luminous, I am copying @Radoslaw >
> > Zarzynski who may be able to help more with that. >> Thanks, > Laura
> > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
>  Hi, >> I just did a new Ceph installation and would like to enable the
> > "read >> balancer". >> However, the documentation requires that the minimum
> > client version >> be reef. I checked this information through "ceph
> > features" and came across >> the situation of having 2 luminous clients. >>
> > # ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",
> >>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
> >>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
> > "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
> >>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
> > version to reef and received the >> following alert: >> # ceph osd
> > set-require-min-compat-client reef >> Error EPERM: cannot set
> > require_min_compat_client to reef: 2 connected >> client(s) look like
> > luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
> > anyway  Is it ok do confirm anyway? >> Which processes are still as
> > luminous?  Rafael. >> ___
> >>> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an
> > email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> She/Her/Hers
> >>> Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |
> > lflo...@redhat.com > M: +17087388804 >
> > ___ > ceph-users mailing list
> > -- ceph-users@ceph.io > To unsubscribe send an email to
> > ceph-users-le...@ceph.io ___
> > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email
> > to ceph-users-le...@ceph.io
> 
> 
>  ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
What is a “storage server”?
  These are machines that only have the operating system and ceph installed.
 



De: "Anthony D'Atri" 
Enviada: 2025/04/08 10:19:08
Para: quag...@bol.com.br
Cc:  ebl...@nde.ag, ceph-users@ceph.io
Assunto:  Re: [ceph-users] Ceph squid fresh install
 


> On Apr 8, 2025, at 9:13 AM, quag...@bol.com.br wrote:
>
> These 2 IPs are from the storage servers.

What is a “storage server”?


> There are no user processes running on them. It only has the operating system and ceph installed.

Nobody said anything about user processes.

>
>
> Rafael.
>
> De: "Eugen Block" 
> Enviada: 2025/04/08 09:35:35
> Para: quag...@bol.com.br
> Cc: ceph-users@ceph.io
> Assunto: Re: [ceph-users] Ceph squid fresh install
>
> These are your two Luminous clients:
>
> ---snip---
> {
> "name": "unknown.0",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> }
> ]
> },
> "socket_addr": {
> "type": "none",
> "addr": "172.27.254.7:0",
> "nonce": 443842330
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ...
>
> {
> "name": "client.104098",
> "entity_name": "client.admin",
> "addrs": {
> "addrvec": [
> {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> }
> ]
> },
> "socket_addr": {
> "type": "v1",
> "addr": "172.27.254.6:0",
> "nonce": 2027668300
> },
> "con_type": "client",
> "con_features": 3387146417253690110,
> "con_features_hex": "2f018fb87aa4aafe",
> "con_features_release": "luminous",
> ---snip---
>
> Zitat von quag...@bol.com.br:
>
> > Hi Eugen! Thanks a lot! I was able to find luminous connections,
> > but I still can't identify which client process. Here is the output:
> > Rafael.
> > ──
> > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
> > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> > the MON sessions to identify your older clients with: ceph tell mon.
> > sessions It will show you the IP address, con_features_release (Luminous)
> > and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
> > would not force the min_compat_client to be reef when there are still >
> > luminous clients connected, as it is important for all clients to be >=Reef
> >> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
> > for checking which processes are still luminous, I am copying @Radoslaw >
> > Zarzynski who may be able to help more with that. >> Thanks, > Laura
> > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
>  Hi, >> I just did a new Ceph installation and would like to enable the
> > "read >> balancer". >> However, the documentation requires that the minimum
> > client version >> be reef. I checked this information through "ceph
> > features" and came across >> the situation of having 2 luminous clients. >>
> > # ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",
> >>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
> >>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
> > "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
> > "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
> >>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
> > "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
> > version to reef and received the >> following alert: >> # ceph osd
> > set-require-min-compat-client reef >> Error EPERM: cannot set
> > require_min_compat_client to reef: 2 connected >> client(s) look like
> > luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
> > anyway  Is it ok do confirm anyway? >> Which processes are still as
> > luminous?  Rafael. >> ___
> >>> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an
> > email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> She/Her/Hers
> >>> Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |
> > lflo...@redhat.com > M: +17087388804 >
> > ___ > ceph-users mailing list
> > -- ceph-users@ceph.io > To unsubscribe send an email to
> > ceph-users-le...@ceph.io ___
> > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email
> > to ceph-users-le...@ceph.io
>
>
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
 ___
ceph-users mailing list --

[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quaglio
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Gorbachev
Hi Linas,

Is the intent of purging of this data mainly due to just cost concerns?  If
the goal is purely preservation of data, the likely cheapest and least
maintenance intensive way of doing this is a large scale tape archive.
Such archives (purely based on a google search) exist at LLNL and OU, and
there is a TAPAS service from SpectraLogic.

I would imagine questions would arise about custody of the data, legal
implications etc.  The easiest is for the organization already hosting the
data to just preserve it by archiving, and thereby claim a significant cost
reduction.

--
Alex Gorbachev




On Sun, Apr 6, 2025 at 11:08 PM Linas Vepstas 
wrote:

> OK what you will read below might sound insane but I am obliged to ask.
>
> There are 275 petabytes of NIH data at risk of being deleted. Cancer
> research, medical data, HIPAA type stuff. Currently unclear where it's
> located, how it's managed, who has access to what, but lets ignore
> that for now. It's presumably splattered across data centers, cloud,
> AWS, supercomputing labs, who knows. Everywhere.
>
> I'm talking to a biomed person in Australias that uses NCBI data
> daily, she's in talks w/ Australian govt to copy and preserve the
> datasets they use. Some multi-petabytes of stuff. I don't know.
>
> While bouncing around tech ideas, IPFS and Ceph came up. My experience
> with IPFS is that it's not a serious contender for anything. My
> experience with Ceph is that it's more-or-less A-list.
>
> OK. So here's the question: is it possible to (has anyone tried) set
> up an internet-wide Ceph cluster? Ticking off the typical checkboxes
> for "decentralized storage"? Stuff, like: internet connections need to
> be encrypted. Connections go down, come back up. Slow. Sure, national
> labs may have multi-terabit fiber, but little itty-bitty participants
> trying to contribute a small collection of disks to a large pool might
> only have a gigabit connection, of which maybe 10% is "usable".
> Barely. So, a hostile networking environment.
>
> Is this like, totally insane, run away now, can't do that, it won't
> work idea, or is there some glimmer of hope?
>
> Am I misunderstanding something about IPFS that merits taking a second
> look at it?
>
> Is there any other way of getting scalable reliable "decentralized"
> internet-wide storage?
>
> I mean, yes, of course, the conventional answer is that it could be
> copied to AWS or some national lab or two somewhere in the EU or Aus
> or UK or where-ever, That's the "obvious" answer. I'm looking for a
> non-obvious answer, an IPFS-like thing, but one that actually works.
> Could it work?
>
> -- Linas
>
>
> --
> Patrick: Are they laughing at us?
> Sponge Bob: No, Patrick, they are laughing next to us.
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: NIH Datasets

2025-04-08 Thread Anthony D'Atri
The intent is the US administration’s assault against science, Linas doesn’t 
*want* to do it, he wants to preserve for the hope of a better future.


> On Apr 8, 2025, at 9:28 AM, Alex Gorbachev  wrote:
> 
> Hi Linas,
> 
> Is the intent of purging of this data mainly due to just cost concerns?  If
> the goal is purely preservation of data, the likely cheapest and least
> maintenance intensive way of doing this is a large scale tape archive.
> Such archives (purely based on a google search) exist at LLNL and OU, and
> there is a TAPAS service from SpectraLogic.
> 
> I would imagine questions would arise about custody of the data, legal
> implications etc.  The easiest is for the organization already hosting the
> data to just preserve it by archiving, and thereby claim a significant cost
> reduction.
> 
> --
> Alex Gorbachev
> 
> 
> 
> 
> On Sun, Apr 6, 2025 at 11:08 PM Linas Vepstas 
> wrote:
> 
>> OK what you will read below might sound insane but I am obliged to ask.
>> 
>> There are 275 petabytes of NIH data at risk of being deleted. Cancer
>> research, medical data, HIPAA type stuff. Currently unclear where it's
>> located, how it's managed, who has access to what, but lets ignore
>> that for now. It's presumably splattered across data centers, cloud,
>> AWS, supercomputing labs, who knows. Everywhere.
>> 
>> I'm talking to a biomed person in Australias that uses NCBI data
>> daily, she's in talks w/ Australian govt to copy and preserve the
>> datasets they use. Some multi-petabytes of stuff. I don't know.
>> 
>> While bouncing around tech ideas, IPFS and Ceph came up. My experience
>> with IPFS is that it's not a serious contender for anything. My
>> experience with Ceph is that it's more-or-less A-list.
>> 
>> OK. So here's the question: is it possible to (has anyone tried) set
>> up an internet-wide Ceph cluster? Ticking off the typical checkboxes
>> for "decentralized storage"? Stuff, like: internet connections need to
>> be encrypted. Connections go down, come back up. Slow. Sure, national
>> labs may have multi-terabit fiber, but little itty-bitty participants
>> trying to contribute a small collection of disks to a large pool might
>> only have a gigabit connection, of which maybe 10% is "usable".
>> Barely. So, a hostile networking environment.
>> 
>> Is this like, totally insane, run away now, can't do that, it won't
>> work idea, or is there some glimmer of hope?
>> 
>> Am I misunderstanding something about IPFS that merits taking a second
>> look at it?
>> 
>> Is there any other way of getting scalable reliable "decentralized"
>> internet-wide storage?
>> 
>> I mean, yes, of course, the conventional answer is that it could be
>> copied to AWS or some national lab or two somewhere in the EU or Aus
>> or UK or where-ever, That's the "obvious" answer. I'm looking for a
>> non-obvious answer, an IPFS-like thing, but one that actually works.
>> Could it work?
>> 
>> -- Linas
>> 
>> 
>> --
>> Patrick: Are they laughing at us?
>> Sponge Bob: No, Patrick, they are laughing next to us.
>> ___
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] squid 19.2.2 hotfix QE validation status

2025-04-08 Thread Yuri Weinstein
This is a hotfix release to address only one issue:
https://github.com/ceph/ceph/pull/62711 and has limited testing.

Details of this release are summarized here:

https://tracker.ceph.com/issues/70822#note-1

Release Notes - TBD
LRC upgrade - excluded from this release
Gibba upgrade - excluded from this release

smoke - Adam E, Casey approved?
rgw - Adam E, Casey approved?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: NIH Datasets

2025-04-08 Thread Tim Holloway

I don't think Linus is only concerned with public data, no.

The United States Government has had in places for many years effective 
means of preserving their data. Some of those systems may be old and 
creaky, granted, and not always the most efficient, but they suffice.


The problem is that the current administration and their unelected 
henchmen are running rampant over them. Firing people with critical 
knowledge, and ordering the destruction of carefully amassed data.


So what we're looking at is more akin to wikileaks than to simple data 
archival. That is, for the good of the nation - and the world - that 
data should be preserved, not purged. And yes, that might include 
obtaining information by shady means because the administration would 
rather see some data completely expunged in the same way that the 
Taliban destroyed the Buddha statues.


A Ceph-like system would definitely be a good model, but for the sake of 
reliability and accessibility, we'd want something that could endure 
nodes popping in and dropping out, and to lessen the chances of a 
full-out invasion, that data would best be anonymously replicated over 
many nodes. Ceph loves very large datasets, so it should be sufficient 
were we to keep all our eggs in one basket. but again, that data has 
enemies, so extra measures are needful that Ceph wasn't designed to handle.


I think such a project is doable, but it should borrow from many other 
community service architectures as well. It's not something that is 
currently available off-the-shelf, but neither were a lot of the 
technologies we now depend on every day.


On 4/8/25 13:26, Alex Gorbachev wrote:

I was trying to analyze the original request, which seems to be something
of the following set:

- The goal is to archive a large amount of (presumably public) data on a
community run globally sharded or distributed storage.

- Can Ceph be used for this?  Seems no, at least not in a sense of running
lots of OSDs at different locations by different people loosely coupled
into one global public data repo.  Perhaps, there are some other ideas from
people who have done this kind of thing.

- Are there restrictions on obtaining the data?  If it's public and
accessible now, it should be able to be copied.  If not, what are the
restrictions on obtaining and copying the data?

- Organization: how will the storage and maintenance of data be organized
(and funded)?  A foundation, a SETI-at-home like network, a blockchain (to
preserve data veracity)?

- Legal support?

--
Alex Gorbachev




On Tue, Apr 8, 2025 at 9:41 AM Anthony D'Atri  wrote:


The intent is the US administration’s assault against science, Linas
doesn’t *want* to do it, he wants to preserve for the hope of a better
future.



On Apr 8, 2025, at 9:28 AM, Alex Gorbachev 

wrote:

Hi Linas,

Is the intent of purging of this data mainly due to just cost concerns?

If

the goal is purely preservation of data, the likely cheapest and least
maintenance intensive way of doing this is a large scale tape archive.
Such archives (purely based on a google search) exist at LLNL and OU, and
there is a TAPAS service from SpectraLogic.

I would imagine questions would arise about custody of the data, legal
implications etc.  The easiest is for the organization already hosting

the

data to just preserve it by archiving, and thereby claim a significant

cost

reduction.

--
Alex Gorbachev




On Sun, Apr 6, 2025 at 11:08 PM Linas Vepstas 
wrote:


OK what you will read below might sound insane but I am obliged to ask.

There are 275 petabytes of NIH data at risk of being deleted. Cancer
research, medical data, HIPAA type stuff. Currently unclear where it's
located, how it's managed, who has access to what, but lets ignore
that for now. It's presumably splattered across data centers, cloud,
AWS, supercomputing labs, who knows. Everywhere.

I'm talking to a biomed person in Australias that uses NCBI data
daily, she's in talks w/ Australian govt to copy and preserve the
datasets they use. Some multi-petabytes of stuff. I don't know.

While bouncing around tech ideas, IPFS and Ceph came up. My experience
with IPFS is that it's not a serious contender for anything. My
experience with Ceph is that it's more-or-less A-list.

OK. So here's the question: is it possible to (has anyone tried) set
up an internet-wide Ceph cluster? Ticking off the typical checkboxes
for "decentralized storage"? Stuff, like: internet connections need to
be encrypted. Connections go down, come back up. Slow. Sure, national
labs may have multi-terabit fiber, but little itty-bitty participants
trying to contribute a small collection of disks to a large pool might
only have a gigabit connection, of which maybe 10% is "usable".
Barely. So, a hostile networking environment.

Is this like, totally insane, run away now, can't do that, it won't
work idea, or is there some glimmer of hope?

Am I misunderstanding something about IPFS that merits taking a second
look at it?

Is 

[ceph-users] irc/slack/discord bridge failing

2025-04-08 Thread Alvaro Soto
Hi all,
Our IRC/Slack/Discord bridge started failing since ~2 days ago. Slack
deprecated the use of legacy bots[1], so at the moment, IRC/Discord is the
only bridge working.

[1]
~~~
time="2025-04-08T18:09:30Z" level=error msg="Could not retrieve channels:
slack.SlackErrorResponse{Err:"legacy_custom_bots_deprecated",
ResponseMetadata:slack.ResponseMetadata{Cursor:"", Messages:[]string(nil),
Warnings:[]string(nil)}}" prefix=slack
~~~

I'll keep you posted.
Cheers!
-- 

Alvaro Soto

*Note: My work hours may not be your work hours. Please do not feel the
need to respond during a time that is not convenient for you.*
--
Great people talk about ideas,
ordinary people talk about things,
small people talk... about other people.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Eugen Block

I think it's just hard-coded in the cephadm binary [0]:

_log_file_handler = {
'level': 'DEBUG',
'class': 'logging.handlers.WatchedFileHandler',
'formatter': 'cephadm',
'filename': '%s/cephadm.log' % LOG_DIR,
}

I haven't looked to deep yet if it can be overridden, but didn't find  
anything on a quick search.


[0]  
https://github.com/ceph/ceph/blob/46ba365307312980303de311cacbb49974dbc74f/src/cephadm/cephadmlib/logging.py#L77


Zitat von Alex :


Hi everyone.

My company has paid Ceph support.
The tech is saying:
"...cephadm package of ceph 5 is having bug. So It generate debug logs
even its set for "info" logs..."
I have two clusters running, one Ceph 5 and the other Ceph 6 (Quincy).
Both of them are sending "DEBUG -" messages to cephadm.log.
I checked everywhere I can think of and all the logs are set to INFO.

Is there really a bug? I can't find any mention of one anywhere on  
the internet,

besides, it seems that if there really is a bug it hasn't been fixed
in version 6 either.

I tried setting " mgr/cephadm/log_to_cluster_level" and  
"mgr/cephadm/log_level"

to WARN and noting changed.

/etc/ceph/ceph.conf has no mention of log level.

Any help would be appreciated,
Thanks!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Interesting. So it's like that for everybody?
Meaning cephadm.log logs debug messages.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quaglio
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread quag...@bol.com.br
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Awesome. Wish I knew that before I spent half a day trying to overwrite it.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Anthony D'Atri
Some of your messages come across without newlines which makes them hard to 
read.

Try

ceph tell mds.0 client ls | grep kernel_version

and

ceph daemon mon.whatever sessions


> On Apr 7, 2025, at 4:34 PM, quag...@bol.com.br wrote:
> 
> Hi Anthony, Thanks for your reply. I dont have any client connected yet. It's 
> a fresh install. I only made a CephFS mount point on each machine that I 
> installed ceph. How can I identify these clients? Can you point me? Rafael. 
> ── De: 
> "Anthony D'Atri" Enviada: 2025/04/07 17:11:50 Para: quag...@bol.com.br 
> Assunto: Re: [ceph-users] Ceph squid fresh install Clients, whatever your 
> clients are. RBD attachments, CephFS mounts. I’m guessing you have clients 
> mounting CephFS with a really old kernel. > On Apr 7, 2025, at 12:29 PM, 
> quag...@bol.com.br wrote: > > Hi, > I just did a new Ceph installation and 
> would like to enable the "read balancer". > However, the documentation 
> requires that the minimum client version be reef. I checked this information 
> through "ceph features" and came across the situation of having 2 luminous 
> clients. > # ceph features > { > "mon": [ > { > "features": 
> "0x3f03cffd", > "release": "squid", > "num": 2 > } > ], > "mds": [ > 
> { > "features": "0x3f03cffd", > "release": "squid", > "num": 2 > } > 
> ], > "osd": [ > { > "features": "0x3f03cffd", > "release": "squid", > 
> "num": 38 > } > ], > "client": [ > { > "features": "0x2f018fb87aa4aafe", > 
> "release": "luminous", > "num": 2 > }, > { > "features": 
> "0x3f03cffd", > "release": "squid", > "num": 5 > } > ], > "mgr": [ > 
> { > "features": "0x3f03cffd", > "release": "squid", > "num": 2 > } > 
> ] > } > > I tryed to configure the minimum version to reef and received the 
> following alert: > # ceph osd set-require-min-compat-client reef > Error 
> EPERM: cannot set require_min_compat_client to reef: 2 connected client(s) 
> look like luminous (missing 0x8000); add --yes-i-really-mean-it to do it 
> anyway > > Is it ok do confirm anyway? > Which processes are still as 
> luminous? > > Rafael. > ___ > 
> ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email 
> to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] RBD Block alignment 16k for Databases

2025-04-08 Thread filip Mutterer

Hi!

Is Block alignment for Databases a thing with RBD?

I heard something with 16K alignment ...

If yes how can be checked what alignment a RBD got.

And yes I need to run my Database in Ceph as I have no other storage 
with redundandency.


Hints for optimizing a single Block device for a Database are very welcome.


Greetings

filip
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: squid 19.2.2 hotfix QE validation status

2025-04-08 Thread Adam Emerson
On 08/04/2025, Yuri Weinstein wrote:
> This is a hotfix release to address only one issue:
> https://github.com/ceph/ceph/pull/62711 and has limited testing.
> 
> Details of this release are summarized here:
> 
> https://tracker.ceph.com/issues/70822#note-1
> 
> Release Notes - TBD
> LRC upgrade - excluded from this release
> Gibba upgrade - excluded from this release
> 
> smoke - Adam E, Casey approved?

Approved!

> rgw - Adam E, Casey approved?

Approved!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Laimis Juzeliūnas
Hi all,

This most likely boils down to the kernel version of connected clients. We
faced the same problem some time ago and just kicked away the client mds
connections with a follow up kernel update on the servers doing the mount.
I can check the version details later for more clarity.


Best,
Laimis J.

On Tue, Apr 8, 2025, 16:48 Anthony D'Atri  wrote:

> So, OSD nodes that do not have any OSDs on them yet?
>
> > On Apr 8, 2025, at 9:41 AM, quag...@bol.com.br wrote:
> >
> > More complete description:
> >
> > 1-) I formatted and installed the operating system
> >
> > 2-) This is "ceph installed":
> >
> > curl --silent --remote-name --location
> https://download.ceph.com/rpm-19.2.1/el9/noarch/cephadm
> > chmod +x cephadm
> >
> > ./cephadm add-repo --release squid
> > ./cephadm install
> >
> > cephadm -v bootstrap --mon-ip 172.27.254.6 --cluster-network
> 172.28.254.0/24 --log-to-file
> >
> > cephadm install ceph-common
> >
> >
> > De: "Anthony D'Atri" 
> > Enviada: 2025/04/08 10:35:22
> > Para: quag...@bol.com.br
> > Cc: ebl...@nde.ag, ceph-users@ceph.io
> > Assunto: Re: [ceph-users] Ceph squid fresh install
> >
> > What does “ceph installed” mean?  I suspect that this description is not
> complete.
> >
> >>
> >> On Apr 8, 2025, at 9:21 AM, quag...@bol.com.br wrote:
> >>
> >> What is a “storage server”?
> >>   These are machines that only have the operating system and
> ceph installed.
> >>
> >>
> >>
> >> De: "Anthony D'Atri" 
> >> Enviada: 2025/04/08 10:19:08
> >> Para: quag...@bol.com.br
> >> Cc: ebl...@nde.ag, ceph-users@ceph.io
> >> Assunto: Re: [ceph-users] Ceph squid fresh install
> >>
> >>
> >>
> >> > On Apr 8, 2025, at 9:13 AM, quag...@bol.com.br wrote:
> >> >
> >> > These 2 IPs are from the storage servers.
> >>
> >> What is a “storage server”?
> >>
> >>
> >> > There are no user processes running on them. It only has the
> operating system and ceph installed.
> >>
> >> Nobody said anything about user processes.
> >>
> >> >
> >> >
> >> > Rafael.
> >> >
> >> > De: "Eugen Block" 
> >> > Enviada: 2025/04/08 09:35:35
> >> > Para: quag...@bol.com.br
> >> > Cc: ceph-users@ceph.io
> >> > Assunto: Re: [ceph-users] Ceph squid fresh install
> >> >
> >> > These are your two Luminous clients:
> >> >
> >> > ---snip---
> >> > {
> >> > "name": "unknown.0",
> >> > "entity_name": "client.admin",
> >> > "addrs": {
> >> > "addrvec": [
> >> > {
> >> > "type": "none",
> >> > "addr": "172.27.254.7:0",
> >> > "nonce": 443842330
> >> > }
> >> > ]
> >> > },
> >> > "socket_addr": {
> >> > "type": "none",
> >> > "addr": "172.27.254.7:0",
> >> > "nonce": 443842330
> >> > },
> >> > "con_type": "client",
> >> > "con_features": 3387146417253690110,
> >> > "con_features_hex": "2f018fb87aa4aafe",
> >> > "con_features_release": "luminous",
> >> > ...
> >> >
> >> > {
> >> > "name": "client.104098",
> >> > "entity_name": "client.admin",
> >> > "addrs": {
> >> > "addrvec": [
> >> > {
> >> > "type": "v1",
> >> > "addr": "172.27.254.6:0",
> >> > "nonce": 2027668300
> >> > }
> >> > ]
> >> > },
> >> > "socket_addr": {
> >> > "type": "v1",
> >> > "addr": "172.27.254.6:0",
> >> > "nonce": 2027668300
> >> > },
> >> > "con_type": "client",
> >> > "con_features": 3387146417253690110,
> >> > "con_features_hex": "2f018fb87aa4aafe",
> >> > "con_features_release": "luminous",
> >> > ---snip---
> >> >
> >> > Zitat von quag...@bol.com.br:
> >> >
> >> > > Hi Eugen! Thanks a lot! I was able to find luminous connections,
> >> > > but I still can't identify which client process. Here is the output:
> >> > > Rafael.
> >> > > ──
> >> > > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para:
> ceph-users@ceph.io
> >> > > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> >> > > the MON sessions to identify your older clients with: ceph tell mon.
> >> > > sessions It will show you the IP address, con_features_release
> (Luminous)
> >> > > and a couple of other things. Zitat von Laura Flores : > Hi Rafael,
> >> I
> >> > > would not force the min_compat_client to be reef when there are
> still >
> >> > > luminous clients connected, as it is important for all clients to
> be >=Reef
> >> > >> to understand/encode the pg_upmap_primary feature in the osdmap.
> >> As
> >> > > for checking which processes are still luminous, I am copying
> @Radoslaw >
> >> > > Zarzynski who may be able to help more with that. >> Thanks, > Laura
> >> > > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br >
> wrote:
> >> >  Hi, >> I just did a new Ceph installation and would like to
> enable the
> >> > > "read >> balancer". >> However, the documentation requires that the
> minimum
> >> > > client version >> be reef. I checked this information through "ceph
> >> > > features" and came across >> the situation of having 2 luminous
> clients. >>
> >> > > # ceph features >> { >> "mon": [ >> { >> "features":
> "0x3f03cffd",
> >> > >>> "release": "squid", >> "num": 2 >> } >> ], >

[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Buie
If you have a line on the data, I have connections that can store it or I
can consult pro bono on building a system to store it.

However wan-ceph is not the answer here.

On Sun, Apr 6, 2025, 11:08 PM Linas Vepstas  wrote:

> OK what you will read below might sound insane but I am obliged to ask.
>
> There are 275 petabytes of NIH data at risk of being deleted. Cancer
> research, medical data, HIPAA type stuff. Currently unclear where it's
> located, how it's managed, who has access to what, but lets ignore
> that for now. It's presumably splattered across data centers, cloud,
> AWS, supercomputing labs, who knows. Everywhere.
>
> I'm talking to a biomed person in Australias that uses NCBI data
> daily, she's in talks w/ Australian govt to copy and preserve the
> datasets they use. Some multi-petabytes of stuff. I don't know.
>
> While bouncing around tech ideas, IPFS and Ceph came up. My experience
> with IPFS is that it's not a serious contender for anything. My
> experience with Ceph is that it's more-or-less A-list.
>
> OK. So here's the question: is it possible to (has anyone tried) set
> up an internet-wide Ceph cluster? Ticking off the typical checkboxes
> for "decentralized storage"? Stuff, like: internet connections need to
> be encrypted. Connections go down, come back up. Slow. Sure, national
> labs may have multi-terabit fiber, but little itty-bitty participants
> trying to contribute a small collection of disks to a large pool might
> only have a gigabit connection, of which maybe 10% is "usable".
> Barely. So, a hostile networking environment.
>
> Is this like, totally insane, run away now, can't do that, it won't
> work idea, or is there some glimmer of hope?
>
> Am I misunderstanding something about IPFS that merits taking a second
> look at it?
>
> Is there any other way of getting scalable reliable "decentralized"
> internet-wide storage?
>
> I mean, yes, of course, the conventional answer is that it could be
> copied to AWS or some national lab or two somewhere in the EU or Aus
> or UK or where-ever, That's the "obvious" answer. I'm looking for a
> non-obvious answer, an IPFS-like thing, but one that actually works.
> Could it work?
>
> -- Linas
>
>
> --
> Patrick: Are they laughing at us?
> Sponge Bob: No, Patrick, they are laughing next to us.
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Diskprediction_local mgr module removal - Call for feedback

2025-04-08 Thread Anthony D'Atri
I’ve done the same as well.

It doesn’t help that smartctl doesn’t even try to have consistent output for 
SATA, SAS, and NVMe drives, and especially that it doesn’t enforce uniform 
attribute labels.  Yes, the fundamental problem is that SMART is mechanism 
without policy, and inconsistently implemented, but smartctl doesn’t help like 
it could.  I’m working on a post processor for derived.h to at least debulk the 
madness, e.g. casting all of the SSD life remaining/used metrics into the same 
name, and adding a primitive to subtract from 100.  If upstream won’t take that 
addition, I guess I’ll be forking smartmoxntools in toto.

I had hopes for smartctl_exporter, but have given up on that project as too 
disorganized and contentious.

So the last time I did this, I looked up the details of all 76 drive SKUs in my 
fleet and hardcoded every one into my collection script.

Many systems cheap out and just use the overall pass/fail SMART health status 
attribute, which is very prone to reporting that clearly failing drives are 
just smurfy, and thus worthless.  This is what BMCs seem to do, for example.

Grown defects - on a spinner RMA or shred it if there are say more than 1 per 
2TB.

SATA downshift errors, though these can be HBA issues as well.

UDMA/CRC errors - silently slow, but can be addressed by reseating in most 
cases, and using OEM carriers for SATA/SAS drives.

Rate of reallocated blocks, alert if it isn’t quite slow

Since Nautilus we are much less prone to grown errors then we used to be, the 
OSD will retry a failed write, so a remapped LBA will succeed the second time.  
There is IIRC a limit to how many of these are tolerated, but it does 
underscore the need to look deeper.

Similarly one can alert on drives with high latency via primal query.

Oh we were talking about the module.  I vote to remove it.  Note that the RH 
model is binary blobs as well as the ProphetStor model.  Incomplete and 
impossible to maintain in the current state.

It would be terrific to have a fully normalized and maintained / maintainable 
metric and prediction subsystem, but I haven’t seen anyone step up.  It would 
be too much for me to do myself, especially without access to hardware, and I 
fear that without multiple people involved we won’t have continuity and it’ll 
lapse again.  If at least one SWE with staying power steps up I’d be willing to 
entertain an idea for reimplementation eschewing opaque madness.

anthonydatri@Mac models % pwd
/Users/anthonydatri/git/ceph/src/pybind/mgr/diskprediction_local/models
anthonydatri@Mac models % file redhat/*
redhat/config.json:   JSON data
redhat/hgst_predictor.pkl:data
redhat/hgst_scaler.pkl:   data
redhat/seagate_predictor.pkl: data
redhat/seagate_scaler.pkl:data
anthonydatri@Mac models %





> On Apr 8, 2025, at 5:30 AM, Lukasz Borek  wrote:
> 
> +1
> 
> I wasn't aware that this module is obsolete and was trying to start it a
> few weeks ago.
> 
> We develop a home-made  solution some time ago to monitor smart data from
> both HDD (uncorrected errors, grown defect list) and SSD (WLC/TBW). But
> keeping it up to date with non-unified disk models is a nightmare.
> 
> Alert : "OSD.12 is going to fail. Replace it soon" before seeing SLOW_OPS
> would be a game changer!
> 
> Thanks!
> 
> On Tue, 8 Apr 2025 at 10:00, Michal Strnad  wrote:
> 
>> Hi.
>> 
>> From our point of view, it's important to keep disk failure prediction
>> tool as part of Ceph, ideally as an MGR module. In environments with
>> hundreds or thousands of disks, it's crucial to know whether, for
>> example, a significant number of them are likely to fail within a month
>> - which, in the best-case scenario, would mean performance degradation,
>> and in the worst-case, data loss.
>> 
>> Some have already responded to the deprecation of diskprediction by
>> starting to develop their own solutions. For instance, just yesterday,
>> Daniel Persson published a solution [1] on his website that addresses
>> the same problem.
>> 
>> Would it be possible to join forces and try to revive that module?
>> 
>> [1] https://www.youtube.com/watch?v=Gr_GtC9dcMQ
>> 
>> Thanks,
>> Michal
>> 
>> 
>> On 4/8/25 01:18, Yaarit Hatuka wrote:
>>> Hi everyone,
>>> 
>>> On today's Ceph Steering Committee call we discussed the idea of removing
>>> the diskprediction_local mgr module, as the current prediction model is
>>> obsolete and not maintained.
>>> 
>>> We would like to gather feedback from the community about the usage of
>> this
>>> module, and find out if anyone is interested in maintaining it.
>>> 
>>> Thanks,
>>> Yaarit
>>> ___
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
>> 
>> ___
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
> 
> 
> -- 
> Łukasz Borek
> luk...@borek.org.pl
>

[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Anthony D'Atri
So, OSD nodes that do not have any OSDs on them yet?

> On Apr 8, 2025, at 9:41 AM, quag...@bol.com.br wrote:
> 
> More complete description:
> 
> 1-) I formatted and installed the operating system
> 
> 2-) This is "ceph installed":
> 
> curl --silent --remote-name --location 
> https://download.ceph.com/rpm-19.2.1/el9/noarch/cephadm
> chmod +x cephadm
> 
> ./cephadm add-repo --release squid
> ./cephadm install
> 
> cephadm -v bootstrap --mon-ip 172.27.254.6 --cluster-network 172.28.254.0/24 
> --log-to-file
> 
> cephadm install ceph-common
>  
> 
> De: "Anthony D'Atri" 
> Enviada: 2025/04/08 10:35:22
> Para: quag...@bol.com.br
> Cc: ebl...@nde.ag, ceph-users@ceph.io
> Assunto: Re: [ceph-users] Ceph squid fresh install
>  
> What does “ceph installed” mean?  I suspect that this description is not 
> complete.
>  
>> 
>> On Apr 8, 2025, at 9:21 AM, quag...@bol.com.br wrote:
>>  
>> What is a “storage server”?
>>   These are machines that only have the operating system and ceph 
>> installed.
>> 
>>  
>> 
>> De: "Anthony D'Atri" 
>> Enviada: 2025/04/08 10:19:08
>> Para: quag...@bol.com.br
>> Cc: ebl...@nde.ag, ceph-users@ceph.io
>> Assunto: Re: [ceph-users] Ceph squid fresh install
>>  
>> 
>> 
>> > On Apr 8, 2025, at 9:13 AM, quag...@bol.com.br wrote:
>> >
>> > These 2 IPs are from the storage servers.
>> 
>> What is a “storage server”?
>> 
>> 
>> > There are no user processes running on them. It only has the operating 
>> > system and ceph installed.
>> 
>> Nobody said anything about user processes.
>> 
>> >
>> >
>> > Rafael.
>> >
>> > De: "Eugen Block" 
>> > Enviada: 2025/04/08 09:35:35
>> > Para: quag...@bol.com.br
>> > Cc: ceph-users@ceph.io
>> > Assunto: Re: [ceph-users] Ceph squid fresh install
>> >
>> > These are your two Luminous clients:
>> >
>> > ---snip---
>> > {
>> > "name": "unknown.0",
>> > "entity_name": "client.admin",
>> > "addrs": {
>> > "addrvec": [
>> > {
>> > "type": "none",
>> > "addr": "172.27.254.7:0",
>> > "nonce": 443842330
>> > }
>> > ]
>> > },
>> > "socket_addr": {
>> > "type": "none",
>> > "addr": "172.27.254.7:0",
>> > "nonce": 443842330
>> > },
>> > "con_type": "client",
>> > "con_features": 3387146417253690110,
>> > "con_features_hex": "2f018fb87aa4aafe",
>> > "con_features_release": "luminous",
>> > ...
>> >
>> > {
>> > "name": "client.104098",
>> > "entity_name": "client.admin",
>> > "addrs": {
>> > "addrvec": [
>> > {
>> > "type": "v1",
>> > "addr": "172.27.254.6:0",
>> > "nonce": 2027668300
>> > }
>> > ]
>> > },
>> > "socket_addr": {
>> > "type": "v1",
>> > "addr": "172.27.254.6:0",
>> > "nonce": 2027668300
>> > },
>> > "con_type": "client",
>> > "con_features": 3387146417253690110,
>> > "con_features_hex": "2f018fb87aa4aafe",
>> > "con_features_release": "luminous",
>> > ---snip---
>> >
>> > Zitat von quag...@bol.com.br:
>> >
>> > > Hi Eugen! Thanks a lot! I was able to find luminous connections,
>> > > but I still can't identify which client process. Here is the output:
>> > > Rafael.
>> > > ──
>> > > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
>> > > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
>> > > the MON sessions to identify your older clients with: ceph tell mon.
>> > > sessions It will show you the IP address, con_features_release (Luminous)
>> > > and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
>> > > would not force the min_compat_client to be reef when there are still >
>> > > luminous clients connected, as it is important for all clients to be 
>> > > >=Reef
>> > >> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
>> > > for checking which processes are still luminous, I am copying @Radoslaw >
>> > > Zarzynski who may be able to help more with that. >> Thanks, > Laura
>> > > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
>> >  Hi, >> I just did a new Ceph installation and would like to enable the
>> > > "read >> balancer". >> However, the documentation requires that the 
>> > > minimum
>> > > client version >> be reef. I checked this information through "ceph
>> > > features" and came across >> the situation of having 2 luminous clients. 
>> > > >>
>> > > # ceph features >> { >> "mon": [ >> { >> "features": 
>> > > "0x3f03cffd",
>> > >>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
>> > > "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
>> > >>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
>> > > "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
>> > > "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
>> > > "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
>> > >>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
>> > > "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
>> > > v

[ceph-users] Re: File immutability on CephFS ?

2025-04-08 Thread Venky Shankar
Hi Fabien,

[cc Milind (who owns the tracker ticket)]

On Tue, Apr 8, 2025 at 8:16 PM Fabien Sirjean  wrote:
>
> Dear all,
>
> If I am to believe this issue [1], it seems that it is still not
> possible to make files immutable (chattr +i) in cephfs.

That's correct.

>
> Do you have any update on this matter ?

Unfortunately, no. The tracker ticket hasn't been prioritized and no
progress has been made. Will see what can be done to have it
implemented. Thanks for the nudge on this.

>
> Thanks a lot for all the good work!
>
> Cheers,
>
> Fabien
>
> [1] : https://tracker.ceph.com/issues/10679
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Cheers,
Venky
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Use cephfs when cephfs-data-scan cleanup is not finished

2025-04-08 Thread Venky Shankar
Hi Jiří,

On Mon, Apr 7, 2025 at 8:36 AM Jiří Župka  wrote:
>
> Dear conference,
>I would like to ask if is secure use cephfs when cephfs-data-scan
> cleanup is not
> finished yet?

You could interrupt the cleanup task and use the file system.

 BTW, you could also speed up the cleanup process by running it via
multiple worker threads as with other data scan commands.

>
> Best regards,
>Jiri
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Cheers,
Venky
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Diskprediction_local mgr module removal - Call for feedback

2025-04-08 Thread Konstantin Shalygin
Hi,

It's will be very nice, if this module will be removed. Everything that Ceph 
operator need can be covered via smartctl_exporter [1]


Thanks,
k
[1] https://github.com/prometheus-community/smartctl_exporter


Sent from my iPhone

> On 8 Apr 2025, at 02:20, Yaarit Hatuka  wrote:
> 
> We would like to gather feedback from the community about the usage of this
> module, and find out if anyone is interested in maintaining it.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Eugen Block

I guess it's debatable as almost everything. ;-)
One of the advantages is that you usually see immediately what is  
failing, you don't have to turn on debug first, retry the deployment  
or whatever again to reproduce. The file doesn't really grow to huge  
sizes (around 2 MB per day or so) and gets logrotated. Most people  
seem okay with it, at least I haven't seen many complaints. But of  
course, you're always welcome to propose a feature or PR.


Zitat von Alex :


Thanks Eugen!

I think you're right since support had me grep for the same code.
Seems crazy though that it's hardcoded doesn't it?
I guess we can mod the Python file, but you'd think that wouldn't be  
necessary.


Should we make a feature request or modify the code ourselves and make
a pull request?



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: NIH Datasets

2025-04-08 Thread Alex Gorbachev
I was trying to analyze the original request, which seems to be something
of the following set:

- The goal is to archive a large amount of (presumably public) data on a
community run globally sharded or distributed storage.

- Can Ceph be used for this?  Seems no, at least not in a sense of running
lots of OSDs at different locations by different people loosely coupled
into one global public data repo.  Perhaps, there are some other ideas from
people who have done this kind of thing.

- Are there restrictions on obtaining the data?  If it's public and
accessible now, it should be able to be copied.  If not, what are the
restrictions on obtaining and copying the data?

- Organization: how will the storage and maintenance of data be organized
(and funded)?  A foundation, a SETI-at-home like network, a blockchain (to
preserve data veracity)?

- Legal support?

--
Alex Gorbachev




On Tue, Apr 8, 2025 at 9:41 AM Anthony D'Atri  wrote:

> The intent is the US administration’s assault against science, Linas
> doesn’t *want* to do it, he wants to preserve for the hope of a better
> future.
>
>
> > On Apr 8, 2025, at 9:28 AM, Alex Gorbachev 
> wrote:
> >
> > Hi Linas,
> >
> > Is the intent of purging of this data mainly due to just cost concerns?
> If
> > the goal is purely preservation of data, the likely cheapest and least
> > maintenance intensive way of doing this is a large scale tape archive.
> > Such archives (purely based on a google search) exist at LLNL and OU, and
> > there is a TAPAS service from SpectraLogic.
> >
> > I would imagine questions would arise about custody of the data, legal
> > implications etc.  The easiest is for the organization already hosting
> the
> > data to just preserve it by archiving, and thereby claim a significant
> cost
> > reduction.
> >
> > --
> > Alex Gorbachev
> >
> >
> >
> >
> > On Sun, Apr 6, 2025 at 11:08 PM Linas Vepstas 
> > wrote:
> >
> >> OK what you will read below might sound insane but I am obliged to ask.
> >>
> >> There are 275 petabytes of NIH data at risk of being deleted. Cancer
> >> research, medical data, HIPAA type stuff. Currently unclear where it's
> >> located, how it's managed, who has access to what, but lets ignore
> >> that for now. It's presumably splattered across data centers, cloud,
> >> AWS, supercomputing labs, who knows. Everywhere.
> >>
> >> I'm talking to a biomed person in Australias that uses NCBI data
> >> daily, she's in talks w/ Australian govt to copy and preserve the
> >> datasets they use. Some multi-petabytes of stuff. I don't know.
> >>
> >> While bouncing around tech ideas, IPFS and Ceph came up. My experience
> >> with IPFS is that it's not a serious contender for anything. My
> >> experience with Ceph is that it's more-or-less A-list.
> >>
> >> OK. So here's the question: is it possible to (has anyone tried) set
> >> up an internet-wide Ceph cluster? Ticking off the typical checkboxes
> >> for "decentralized storage"? Stuff, like: internet connections need to
> >> be encrypted. Connections go down, come back up. Slow. Sure, national
> >> labs may have multi-terabit fiber, but little itty-bitty participants
> >> trying to contribute a small collection of disks to a large pool might
> >> only have a gigabit connection, of which maybe 10% is "usable".
> >> Barely. So, a hostile networking environment.
> >>
> >> Is this like, totally insane, run away now, can't do that, it won't
> >> work idea, or is there some glimmer of hope?
> >>
> >> Am I misunderstanding something about IPFS that merits taking a second
> >> look at it?
> >>
> >> Is there any other way of getting scalable reliable "decentralized"
> >> internet-wide storage?
> >>
> >> I mean, yes, of course, the conventional answer is that it could be
> >> copied to AWS or some national lab or two somewhere in the EU or Aus
> >> or UK or where-ever, That's the "obvious" answer. I'm looking for a
> >> non-obvious answer, an IPFS-like thing, but one that actually works.
> >> Could it work?
> >>
> >> -- Linas
> >>
> >>
> >> --
> >> Patrick: Are they laughing at us?
> >> Sponge Bob: No, Patrick, they are laughing next to us.
> >> ___
> >> ceph-users mailing list -- ceph-users@ceph.io
> >> To unsubscribe send an email to ceph-users-le...@ceph.io
> >>
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
Can someone paste in here their copy of logrotate?
The trick always with rotating logs is that the service writing to it needs
to be restarted or told to stop writing so the file handle gets closed.
Otherwise it stays open and the free disc space isn't recovered.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Eugen Block

It's the same across all versions I have used so far.

Zitat von Alex :


What about Pacific and Quincy?



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
What about Pacific and Quincy?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Diskprediction_local mgr module removal - Call for feedback

2025-04-08 Thread Robert Sander

Hi,

Am 4/8/25 um 14:30 schrieb Anthony D'Atri:


anthonydatri@Mac models % pwd
/Users/anthonydatri/git/ceph/src/pybind/mgr/diskprediction_local/models
anthonydatri@Mac models % file redhat/*
redhat/config.json:   JSON data
redhat/hgst_predictor.pkl:data
redhat/hgst_scaler.pkl:   data
redhat/seagate_predictor.pkl: data
redhat/seagate_scaler.pkl:data
anthonydatri@Mac models %


These are Python pickle files from 2019 containing ML models made with a 
version of sklearn from 2019.


It really looks like this part of the Ceph project is dead.

Regards
--
Robert Sander
Linux Consultant

Heinlein Consulting GmbH
Schwedter Str. 8/9b, 10119 Berlin

https://www.heinlein-support.de

Tel: +49 30 405051 - 0
Fax: +49 30 405051 - 19

Amtsgericht Berlin-Charlottenburg - HRB 220009 B
Geschäftsführer: Peer Heinlein - Sitz: Berlin
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] File immutability on CephFS ?

2025-04-08 Thread Fabien Sirjean

Dear all,

If I am to believe this issue [1], it seems that it is still not 
possible to make files immutable (chattr +i) in cephfs.


Do you have any update on this matter ?

Thanks a lot for all the good work!

Cheers,

Fabien

[1] : https://tracker.ceph.com/issues/10679
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph squid fresh install

2025-04-08 Thread Anthony D'Atri
What does “ceph installed” mean?  I suspect that this description is not 
complete.

> On Apr 8, 2025, at 9:21 AM, quag...@bol.com.br wrote:
> 
> What is a “storage server”?
>   These are machines that only have the operating system and ceph 
> installed.
> 
>  
> 
> De: "Anthony D'Atri" 
> Enviada: 2025/04/08 10:19:08
> Para: quag...@bol.com.br
> Cc: ebl...@nde.ag, ceph-users@ceph.io
> Assunto: Re: [ceph-users] Ceph squid fresh install
>  
> 
> 
> > On Apr 8, 2025, at 9:13 AM, quag...@bol.com.br wrote:
> >
> > These 2 IPs are from the storage servers.
> 
> What is a “storage server”?
> 
> 
> > There are no user processes running on them. It only has the operating 
> > system and ceph installed.
> 
> Nobody said anything about user processes.
> 
> >
> >
> > Rafael.
> >
> > De: "Eugen Block" 
> > Enviada: 2025/04/08 09:35:35
> > Para: quag...@bol.com.br
> > Cc: ceph-users@ceph.io
> > Assunto: Re: [ceph-users] Ceph squid fresh install
> >
> > These are your two Luminous clients:
> >
> > ---snip---
> > {
> > "name": "unknown.0",
> > "entity_name": "client.admin",
> > "addrs": {
> > "addrvec": [
> > {
> > "type": "none",
> > "addr": "172.27.254.7:0",
> > "nonce": 443842330
> > }
> > ]
> > },
> > "socket_addr": {
> > "type": "none",
> > "addr": "172.27.254.7:0",
> > "nonce": 443842330
> > },
> > "con_type": "client",
> > "con_features": 3387146417253690110,
> > "con_features_hex": "2f018fb87aa4aafe",
> > "con_features_release": "luminous",
> > ...
> >
> > {
> > "name": "client.104098",
> > "entity_name": "client.admin",
> > "addrs": {
> > "addrvec": [
> > {
> > "type": "v1",
> > "addr": "172.27.254.6:0",
> > "nonce": 2027668300
> > }
> > ]
> > },
> > "socket_addr": {
> > "type": "v1",
> > "addr": "172.27.254.6:0",
> > "nonce": 2027668300
> > },
> > "con_type": "client",
> > "con_features": 3387146417253690110,
> > "con_features_hex": "2f018fb87aa4aafe",
> > "con_features_release": "luminous",
> > ---snip---
> >
> > Zitat von quag...@bol.com.br:
> >
> > > Hi Eugen! Thanks a lot! I was able to find luminous connections,
> > > but I still can't identify which client process. Here is the output:
> > > Rafael.
> > > ──
> > > De: "Eugen Block" Enviada: 2025/04/08 04:37:47 Para: ceph-users@ceph.io
> > > Assunto: [ceph-users] Re: Ceph squid fresh install Hi, you can query
> > > the MON sessions to identify your older clients with: ceph tell mon.
> > > sessions It will show you the IP address, con_features_release (Luminous)
> > > and a couple of other things. Zitat von Laura Flores : > Hi Rafael, >> I
> > > would not force the min_compat_client to be reef when there are still >
> > > luminous clients connected, as it is important for all clients to be 
> > > >=Reef
> > >> to understand/encode the pg_upmap_primary feature in the osdmap. >> As
> > > for checking which processes are still luminous, I am copying @Radoslaw >
> > > Zarzynski who may be able to help more with that. >> Thanks, > Laura
> > > Flores >> On Mon, Apr 7, 2025 at 11:30 AM quag...@bol.com.br > wrote:
> >  Hi, >> I just did a new Ceph installation and would like to enable the
> > > "read >> balancer". >> However, the documentation requires that the 
> > > minimum
> > > client version >> be reef. I checked this information through "ceph
> > > features" and came across >> the situation of having 2 luminous clients. 
> > > >>
> > > # ceph features >> { >> "mon": [ >> { >> "features": "0x3f03cffd",
> > >>> "release": "squid", >> "num": 2 >> } >> ], >> "mds": [ >> { >>
> > > "features": "0x3f03cffd", >> "release": "squid", >> "num": 2 >> }
> > >>> ], >> "osd": [ >> { >> "features": "0x3f03cffd", >> "release":
> > > "squid", >> "num": 38 >> } >> ], >> "client": [ >> { >> "features":
> > > "0x2f018fb87aa4aafe", >> "release": "luminous", >> "num": 2 >> }, >> { >>
> > > "features": "0x3f03cffd", >> "release": "squid", >> "num": 5 >> }
> > >>> ], >> "mgr": [ >> { >> "features": "0x3f03cffd", >> "release":
> > > "squid", >> "num": 2 >> } >> ] >> }  I tryed to configure the minimum
> > > version to reef and received the >> following alert: >> # ceph osd
> > > set-require-min-compat-client reef >> Error EPERM: cannot set
> > > require_min_compat_client to reef: 2 connected >> client(s) look like
> > > luminous (missing 0x8000); add >> --yes-i-really-mean-it to do it
> > > anyway  Is it ok do confirm anyway? >> Which processes are still as
> > > luminous?  Rafael. >> ___
> > >>> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an
> > > email to ceph-users-le...@ceph.io >> -- >> Laura Flores >> 
> > > She/Her/Hers
> > >>> Software Engineer, Ceph Storage >>> Chicago, IL >> lflo...@ibm.com |
> > > lflo...@redhat.com > M: +17087388804 >
> > > ___ > ceph-users mailing list
> > > -- ceph-users@ceph.io > To unsubscribe send an email 

[ceph-users] Re: Cephadm upgrade from 16.2.15 -> 17.2.0

2025-04-08 Thread Jeremy Hansen
This seems to have worked to get the orch back up and put me back to 16.2.15. 
Thank you. Debating on waiting for 18.2.5 to move forward.

-jeremy

> On Monday, Apr 07, 2025 at 1:26 AM, Eugen Block  (mailto:ebl...@nde.ag)> wrote:
> Still no, just edit the unit.run file for the MGRs to use a different
> image. See Frédéric's instructions (now that I'm re-reading it,
> there's a little mistake with dots and hyphens):
>
> # Backup the unit.run file
> $ cp /var/lib/ceph/$(ceph fsid)/mgr.ceph01.eydqvm/unit.run{,.bak}
>
> # Change container image's signature. You can get the signature of the
> version you
> want to reach from https://quay.io/repository/ceph/ceph?tab=tags. It's
> in the URL of a
> version.
> $ sed -i
> 's/ceph@sha256:e40c19cd70e047d14d70f5ec3cf501da081395a670cd59ca881ff56119660c8f/ceph@sha256:d26c11e20773704382946e34f0d3d2c0b8bb0b7b37d9017faa9dc11a0196c7d9/g'
> /var/lib/ceph/$(ceph fsid)/mgr.ceph01.eydqvm/unit.run
>
> # Restart the container (systemctl daemon-reload not needed)
> $ systemctl restart ceph-$(ceph fsid)(a)mgr.ceph01.eydqvm.service
>
> # Run this command a few times and it should show the new version
> ceph orch ps --refresh --hostname ceph01 | grep mgr
>
> To get the image signature, you can also look into the other unit.run
> files, a version tag would also work.
>
> It depends on how often you need the orchestrator to maintain the
> cluster. If you have the time, you could wait a bit longer for other
> responses. If you need the orchestrator in the meantime, you can roll
> back the MGRs.
>
> https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/32APKOXKRAIZ7IDCNI25KVYFCCCF6RJG/
>
> Zitat von Jeremy Hansen :
>
> > Thank you. The only thing I’m unclear on is the rollback to pacific.
> >
> > Are you referring to
> >
> > > > >
> > > https://docs.ceph.com/en/quincy/cephadm/troubleshooting/#manually-deploying-a-manager-daemon
> >
> > Thank you. I appreciate all the help. Should I wait for Adam to
> > comment? At the moment, the cluster is functioning enough to
> > maintain running vms, so if it’s wise to wait, I can do that.
> >
> > -jeremy
> >
> > > On Monday, Apr 07, 2025 at 12:23 AM, Eugen Block  > > (mailto:ebl...@nde.ag)> wrote:
> > > I haven't tried it this way yet, and I had hoped that Adam would chime
> > > in, but my approach would be to remove this key (it's not present when
> > > no upgrade is in progress):
> > >
> > > ceph config-key rm mgr/cephadm/upgrade_state
> > >
> > > Then rollback the two newer MGRs to Pacific as described before. If
> > > they come up healthy, test if the orchestrator works properly first.
> > > For example, remove a node-exporter or crash or anything else
> > > uncritical and let it redeploy.
> > > If that works, try a staggered upgrade, starting with the MGRs only:
> > >
> > > ceph orch upgrade start --image  --daemon-types mgr
> > >
> > > Since there's no need to go to Quincy, I suggest to upgrade to Reef
> > > 18.2.4 (or you wait until 18.2.5 is released, which should be very
> > > soon), so set the respective  in the above command.
> > >
> > > If all three MGRs successfully upgrade, you can continue with the
> > > MONs, or with the entire rest.
> > >
> > > In production clusters, I usually do staggered upgrades, e. g. I limit
> > > the number of OSD daemons first just to see if they come up healthy,
> > > then I let it upgrade all other OSDs automatically.
> > >
> > > https://docs.ceph.com/en/latest/cephadm/upgrade/#staggered-upgrade
> > >
> > > Zitat von Jeremy Hansen :
> > >
> > > > Snipped some of the irrelevant logs to keep message size down.
> > > >
> > > > ceph config-key get mgr/cephadm/upgrade_state
> > > >
> > > > {"target_name": "quay.io/ceph/ceph:v17.2.0", "progress_id":
> > > > "e7e1a809-558d-43a7-842a-c6229fdc57af", "target_id":
> > > > "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9",
> > > > "target_digests":
> > > >
> > > ["quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a",
> > >  
> > > "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667"],
> > >  "target_version": "17.2.0", "fs_original_max_mds": null, 
> > > "fs_original_allow_standby_replay": null, "error": null, "paused": false, 
> > > "daemon_types": null, "hosts": null, "services": null, "total_count": 
> > > null,
> > > "remaining_count":
> > > > null}
> > > >
> > > > What should I do next?
> > > >
> > > > Thank you!
> > > > -jeremy
> > > >
> > > > > On Sunday, Apr 06, 2025 at 1:38 AM, Eugen Block  > > > > (mailto:ebl...@nde.ag)> wrote:
> > > > > Can you check if you have this config-key?
> > > > >
> > > > > ceph config-key get mgr/cephadm/upgrade_state
> > > > >
> > > > > If you reset the MGRs, it might be necessary to clear this key,
> > > > > otherwise you might end up in some inconsistency. Just to be sure.
> > > > >
> > > > > Zitat von Jeremy Hansen :
> > > > >
> > > > > > Thanks. I’m trying to be extra careful since this cluster is
> > > > > > actually in use

[ceph-users] Re: Cephadm flooding /var/log/ceph/cephadm.log

2025-04-08 Thread Alex
I mean this bit

'log_file': {
'level': 'DEBUG',
'class': 'logging.handlers.WatchedFileHandler',
'formatter': 'cephadm',
'filename': '%s/cephadm.log' % LOG_DIR,
}
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io