Re: [pve-devel] PVE child process behavior question

2025-05-27 Thread Denis Kanchev via pve-devel
--- Begin Message --- Here is the task log 2025-04-11 03:45:42 starting migration of VM 2282 to node 'telpr01pve05' (10.10.17.5) 2025-04-11 03:45:42 starting VM 2282 on remote node 'telpr01pve05' 2025-04-11 03:45:45 [telpr01pve05] Warning: sch_htb: quantum of class 10001 is big. Consider r2q change

Re: [pve-devel] PVE child process behavior question

2025-05-21 Thread Denis Kanchev via pve-devel
very helpful. On 22.05.25 г. 9:30 ч., Fabian Grünbichler wrote: Denis Kanchev via pve-devel hat am 21.05.2025 15:13 CEST geschrieben: Hello, We had an issue with a customer migrating a VM between nodes using our shared storage solution. On the target host the OOM killer killed the main migration

[pve-devel] PVE child process behavior question

2025-05-21 Thread Denis Kanchev via pve-devel
--- Begin Message --- Hello, We had an issue with a customer migrating a VM between nodes using our shared storage solution. On the target host the OOM killer killed the main migration process, but the child process (which actually performs the migration) kept on working, which we did not ex

Re: [pve-devel] PVE child process behavior question

2025-06-02 Thread Denis Kanchev via pve-devel
--- Begin Message --- > I thought your storage plugin is a shared storage, so there is no storage migration at all, yet you keep talking about storage migration? It's a shared storage indeed, the issue was that the migration process on the destination host got OOM killed and the migration failed, m

Re: [pve-devel] PVE child process behavior question

2025-06-02 Thread Denis Kanchev via pve-devel
--- Begin Message --- My bad :) in terms of Proxmox it must be hand-overing the storage control - the storage plugin function activate_volume() is called in our case, which moves the storage to the new VM. So no data is moved across the nodes and only the volumes get re-attached. Thanks for the ple

Re: [pve-devel] PVE child process behavior question

2025-06-02 Thread Denis Kanchev via pve-devel
--- Begin Message --- We tend to prevent having a volume active on two nodes, as may lead to data corruption, so we detach the volume from all nodes ( except the target one ) via our shared storage system. In the sub activate_volume() our logic is to not detach the volume from other hosts in case o

Re: [pve-devel] PVE child process behavior question

2025-05-29 Thread Denis Kanchev via pve-devel
--- Begin Message --- The issue here is that the storage plugin activate_volume() is called after migration cancel which in case of network shared storage can make things bad. This is a sort of race condition, because migration_cancel wont stop the storage migration on the remote server. As you can

Re: [pve-devel] PVE child process behavior question

2025-06-04 Thread Denis Kanchev via pve-devel
--- Begin Message --- > how are you hooking the migration state to know whether deactivation should be done or not? By using the VM property "lock" which must be "migrate": PVE::Cluster::get_guest_config_properties(['lock']); > qm start (over SSH, is this being killed?) > -> start_vm task worker (