GitHub user weizhouapache added a comment to the discussion: Failed to create RBD storage pool after KVM agent upgrade from 4.20 to 4.22: "org.libvirt.LibvirtException: failed to create the RBD IoCTX"
> @weizhouapache Today I just discovered a strange issue as follows, which is > related to this issue. I had to correct a wrong slave interface name inside a > bond. After editing the netplan YAML and running `netplan apply`, the bond > came up correctly and network was fully restored and then I restart > cloudstack-agent. However, immediately after that, CloudStack permanently > lost the Ceph RBD primary storage mount with this repeated agent error: > > ``` > Failed to create RBD storage pool: org.libvirt.LibvirtException: failed to > create the RBD IoCTX. Does the pool 'cloudstack-prod' exist?: No such file or > directory > ``` > > and output of virsh pool-list --all showed the line: > > ``` > error: Could not retrieve pool information > ``` > > While others agent still connected normally so I think the issue is not on > the Ceph side. Interestingly, if I reboot the entire host, the problem > disappears and the RBD pool is mounted normally again. Maybe, does a simple > netplan apply (which only briefly interrupts the network) permanently break > libvirt’s ability to create the RBD pool, while a full host reboot fixes it? hmm, when you change the network configuration, nobody knows what will happen unexpectedly ... before that , you'd better move all vms away and remove all storage pools which are impacted by the network change GitHub link: https://github.com/apache/cloudstack/discussions/12154#discussioncomment-15095645 ---- This is an automatically sent email for [email protected]. To unsubscribe, please send an email to: [email protected]
