Dear Adam, 
thank you very much for your reply. 
In /var/log/ceph/cephadm.log i saw lots of entries like this 

2025-01-08 10:00:22,045 7ff021d8c000 DEBUG 
--------------------------------------------------------------------------------
 
cephadm ['--image', 'harborregistry/quay.io/ceph/ceph', '--timeout', '895', 
'pull'] 
2025-01-08 10:00:22,172 7ff021d8c000 INFO Pulling container image 
harborregistry/quay.io/ceph/ceph... 
2025-01-08 10:00:27,176 7ff021d8c000 INFO Non-zero exit code 125 from 
/usr/bin/podman pull harborregistry/quay.io/ceph/ceph 
2025-01-08 10:00:27,176 7ff021d8c000 INFO /usr/bin/podman: stderr Trying to 
pull harborregistry/quay.io/ceph/ceph:latest... 
2025-01-08 10:00:27,176 7ff021d8c000 INFO /usr/bin/podman: stderr 
time="2025-01-08T10:00:22+01:00" level=warning msg="failed, retrying in 1s ... 
(1/3). Error: initializing source 
docker://harborregistry/quay.io/ceph/ceph:latest: reading manifest latest in 
harborregistry/quay.io/ ceph/ceph: unknown: resource not found: repo 
quay.io/ceph/ceph, tag latest not found" 
... 
2025-01-08 10:00:27,176 7ff021d8c000 INFO /usr/bin/podman: stderr Error: 
initializing source docker://harborregistry/quay.io/ceph/ceph:latest: reading 
manifest latest in harborregistry/quay.io/ceph/ceph: unknown: resource not 
found: repo quay.io/ceph/ceph, tag latest not found 
2025-01-08 10:00:27,177 7ff021d8c000 ERROR ERROR: Failed command: 
/usr/bin/podman pull harborregistry/quay.io/ceph/ceph 
2025-01-08 10:01:27,459 7f5f185d0000 DEBUG 
--------------------------------------------------------------------------------
 

In the meantime i was given a hint, to 
ceph config set mgr container_image harborregistry/quay.io/ceph/ceph:v18.2.4 
which indeed changed things to 

2025-01-08 17:12:45,952 7ffb1da9b000 DEBUG 
--------------------------------------------------------------------------------
 
cephadm ['--image', 'harborregistry/quay.io/ceph/ceph:v18.2.4', '--timeout', 
'895', 'inspect-image'] 
2025-01-08 17:12:46,219 7ffb1da9b000 DEBUG /usr/bin/podman: stdout 
2bc0b0f4375ddf4270a9a865dfd4e53063acc8e6c3afd7a2546507cafd2ec86a,[quay.io/ceph/ceph@sha256:6ac7f923aa1d23b43248ce0ddec
 
7e1388855ee3d00813b52c3172b0b23b37906 
quay.io/ceph/ceph@sha256:ac06cdca6f2512a763f1ace8553330e454152b82f95a2b6bf33c3f3ec2eeac77
 harborregistry/quay.io/ceph/ceph@sha256:6ac7f923aa1d23 
b43248ce0ddec7e1388855ee3d00813b52c3172b0b23b37906 
harborregistry/quay.io/ceph/ceph@sha256:ac06cdca6f2512a763f1ace8553330e454152b82f95a2b6bf33c3f3ec2eeac77]
 
2025-01-08 17:12:46,649 7ffb1da9b000 DEBUG ceph: stdout ceph version 18.2.4 
(e7ad5345525c7aa95470c26863873b581076945d) reef (stable) 
2025-01-08 17:12:50,852 7f21649cf000 DEBUG 
--------------------------------------------------------------------------------
 

Only to now encounter log entries 

2025-01-09 00:01:20,077 7fe3a719e000 DEBUG 
--------------------------------------------------------------------------------
 
cephadm ['--image', 'docker.io/ceph/daemon-base:latest-master-devel', 
'--timeout', '895', '_orch', 'deploy', '--fsid', 'xxxxx'] 
2025-01-09 00:01:20,210 7fe3a719e000 DEBUG Loaded deploy configuration: 
{'fsid': 'xxxxx', 'name': 'mon.monitor0x', 'image': '', 'deploy_arguments': [], 
'params': {}, 'meta': {'service_name': 'mon', 'ports': [], 'ip': None, 
'deployed_by': 
['quay.io/ceph/ceph@sha256:6ac7f923aa1d23b43248ce0ddec7e1388855ee3d00813b52c3172b0b23b37906',
 
'quay.io/ceph/ceph@sha256:ac06cdca6f2512a763f1ace8553330e454152b82f95a2b6bf33c3f3ec2eeac77'],
 'rank': None, 'rank_generation': None, 'extra_container_args': None, 
'extra_entrypoint_args': None}, 'config_blobs': {'config': '# minimal ceph.conf 
for xxxxx\n[global]\n\tfsid = xxxxx\n\tmon_host = 
[v2:x.x.x.x:3300/0,v1:x.x.x.x:6789/0] [v2:x.x.x.x:3300/0,v1:x.x.x.x:6789/0] 
[v2:x.x.x.x:3300/0,v1:x.x.x.x:6789/0]\n[mon.monitor0x]\npublic network = 
x.x.x.0/22\n', 'keyring': '[mon.]\n\tkey = xxxxx\n\tcaps mon = "allow *"\n', 
'files': {'config': '[mon.monitor0x]\npublic network = x.x.x.0/22\n'}}} 
2025-01-09 00:01:20,210 7fe3a719e000 DEBUG Determined image: 
'docker.io/ceph/daemon-base:latest-master-devel' 
2025-01-09 00:01:20,218 7fe3a719e000 INFO Redeploy daemon mon.monitor0x ... 
2025-01-09 00:02:20,255 7fe3a719e000 INFO Non-zero exit code 125 from 
/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM --net=host 
--entrypoint stat --init -e 
CONTAINER_IMAGE=docker.io/ceph/daemon-base:latest-master-devel -e 
NODE_NAME=monitor0x -e CEPH_USE_RANDOM_NONCE=1 
docker.io/ceph/daemon-base:latest-master-devel -c %u %g /var/lib/ceph 
2025-01-09 00:02:20,255 7fe3a719e000 INFO stat: stderr Trying to pull 
docker.io/ceph/daemon-base:latest-master-devel... 
2025-01-09 00:02:20,255 7fe3a719e000 INFO stat: stderr Error: initializing 
source docker://ceph/daemon-base:latest-master-devel: pinging container 
registry registry-1.docker.io: Get "https://registry-1.docker.io/v2/": dial tcp 
54.236.113.205:443: i/o timeout 
2025-01-09 00:02:20,256 7fe3a719e000 ERROR ERROR: Failed to extract uid/gid for 
path /var/lib/ceph: Failed command: /usr/bin/podman run --rm --ipc=host 
--stop-signal=SIGTERM --net=host --entrypoint stat --init -e 
CONTAINER_IMAGE=docker.io/ceph/daemon-base:latest-master-devel -e 
NODE_NAME=monitor0x -e CEPH_USE_RANDOM_NONCE=1 
docker.io/ceph/daemon-base:latest-master-devel -c %u %g /var/lib/ceph: Trying 
to pull docker.io/ceph/daemon-base:latest-master-devel... 
Error: initializing source docker://ceph/daemon-base:latest-master-devel: 
pinging container registry registry-1.docker.io: Get 
"https://registry-1.docker.io/v2/": dial tcp 54.236.113.205:443: i/o timeout 


After that I was also directed to a config setting "mgr 
mgr/cephadm/default_registry" which up to now is absent from our configuration 
(nothing shows up in "ceph config dump | grep registry") 
but i yet have no idea what to set here ... 

Again thank you very much, 
cheers, toBias 



From: "Adam King" <adk...@redhat.com> 
To: "Tobias Tempel" <tobias.tem...@desy.de> 
Cc: "ceph-users" <ceph-users@ceph.io> 
Sent: Wednesday, 8 January, 2025 20:15:51 
Subject: [ceph-users] Re: ceph orch upgrade tries to pull latest? 

It looks like the "resource not found" message is being directly output by 
podman. Is there anything in the cephadm.log (/var/log/ceph/cephadm.log) on 
one of the hosts where this is happening that says what podman command 
cephadm was running that hit this error? 

On Wed, Jan 8, 2025 at 5:27 AM tobias tempel <tobias.tem...@desy.de> wrote: 

> Dear all, 
> i'm trying to cephadm-upgrade in an airgapped environment from 18.2.2 to 
> 18.2.4 ... yet to no avail. 
> local image registry is a harbor instance, I start the upgrade process with 
> 
> ceph orch upgrade start --image harborregistry/quay.io/ceph/ceph:v18.2.4 
> 
> and status looks good 
> 
> ceph orch upgrade status 
> { 
> "target_image": "harborregistry/quay.io/ceph/ceph:v18.2.4", 
> "in_progress": true, 
> "which": "Upgrading all daemon types on all hosts", 
> "services_complete": [], 
> "progress": "", 
> "message": "", 
> "is_paused": false 
> } 
> 
> In the cephadm log i can see messages like 
> 
> cephadm ['--image', 'harborregistry/quay.io/ceph/ceph:v18.2.4', 
> '--timeout', '895', 'inspect-image'] 
> 
> which is fine (works on the commandline), but also 
> 
> 2025-01-08 10:33:53,911 7f9c66d50000 INFO /usr/bin/podman: stderr Error: 
> initializing source docker://harborregistry/quay.io/ceph/ceph:latest: 
> reading manifest latest in harborregistry/quay.io/ceph/ceph: unknown: 
> resource not found: repo quay.io/ceph/ceph, tag latest not found 
> 
> so for some reason cephadm keeps trying to pull tag "latest" - which i did 
> not specify - and this fails ... again and again and again. 
> what am i missing? 
> can anyone give me a hint, where to look at? 
> 
> Thank you very much, 
> cheers, toBias 
> 
> PS: ceph config get mgr 
> WHO MASK LEVEL OPTION 
> VALUE RO 
> mgr basic container_image 
> harborregistry/quay.io/ceph/ceph * 
> mgr advanced mgr/cephadm/container_image_alertmanager 
> harborregistry/quay.io/prometheus/alertmanager * 
> mgr advanced mgr/cephadm/container_image_base 
> harborregistry/quay.io/ceph/ceph 
> mgr advanced mgr/cephadm/container_image_grafana 
> harborregistry/quay.io/ceph/ceph-grafana * 
> mgr advanced mgr/cephadm/container_image_node_exporter 
> harborregistry/quay.io/prometheus/node-exporter * 
> mgr advanced mgr/cephadm/container_image_prometheus 
> harborregistry/quay.io/prometheus/prometheus * 
> 
> _______________________________________________ 
> ceph-users mailing list -- ceph-users@ceph.io 
> To unsubscribe send an email to ceph-users-le...@ceph.io 
> 

Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to