On Tue, Dec 17, 2024 at 04:54:43PM -0500, Steven Sistare wrote: > On 12/16/2024 1:19 PM, Peter Xu wrote: > > On Fri, Dec 13, 2024 at 11:41:45AM -0500, Steven Sistare wrote: > > > On 12/12/2024 4:22 PM, Peter Xu wrote: > > > > On Thu, Dec 12, 2024 at 03:38:00PM -0500, Steven Sistare wrote: > > > > > On 12/9/2024 2:42 PM, Peter Xu wrote: > > > > > > On Mon, Dec 02, 2024 at 05:19:54AM -0800, Steve Sistare wrote: > > > > > > > @@ -2089,13 +2154,23 @@ RAMBlock > > > > > > > *qemu_ram_alloc_internal(ram_addr_t size, ram_addr_t max_size, > > > > > > > new_block->page_size = qemu_real_host_page_size(); > > > > > > > new_block->host = host; > > > > > > > new_block->flags = ram_flags; > > > > > > > + > > > > > > > + if (!host && !xen_enabled()) { > > > > > > > > > > > > Adding one more xen check is unnecessary. This patch needed it > > > > > > could mean > > > > > > that the patch can be refactored.. because we have xen checks in > > > > > > both > > > > > > ram_block_add() and also in the fd allocation path. > > > > > > > > > > > > At the meantime, see: > > > > > > > > > > > > qemu_ram_alloc_from_fd(): > > > > > > if (kvm_enabled() && !kvm_has_sync_mmu()) { > > > > > > error_setg(errp, > > > > > > "host lacks kvm mmu notifiers, -mem-path > > > > > > unsupported"); > > > > > > return NULL; > > > > > > } > > > > > > > > > > > > I don't think any decent kernel could hit this, but that could be > > > > > > another > > > > > > sign that this patch duplicated some file allocations. > > > > > > > > > > > > > + if ((new_block->flags & RAM_SHARED) && > > > > > > > + !qemu_ram_alloc_shared(new_block, &local_err)) { > > > > > > > + goto err; > > > > > > > + } > > > > > > > + } > > > > > > > + > > > > > > > ram_block_add(new_block, &local_err); > > > > > > > - if (local_err) { > > > > > > > - g_free(new_block); > > > > > > > - error_propagate(errp, local_err); > > > > > > > - return NULL; > > > > > > > + if (!local_err) { > > > > > > > + return new_block; > > > > > > > } > > > > > > > - return new_block; > > > > > > > + > > > > > > > +err: > > > > > > > + g_free(new_block); > > > > > > > + error_propagate(errp, local_err); > > > > > > > + return NULL; > > > > > > > } > > > > > > > > > > > > IIUC we only need to conditionally convert an anon-allocation into > > > > > > an > > > > > > fd-allocation, and then we don't need to mostly duplicate > > > > > > qemu_ram_alloc_from_fd(), instead we reuse it. > > > > > > > > > > > > I do have a few other comments elsewhere, but when I was trying to > > > > > > comment. > > > > > > E.g., we either shouldn't need to bother caching qemu_memfd_check() > > > > > > results, or do it in qemu_memfd_check() directly.. and some more. > > > > > > > > > > Someone thought it a good idea to cache the result of > > > > > qemu_memfd_alloc_check, > > > > > and qemu_memfd_check will be called more often. I'll cache the > > > > > result inside > > > > > qemu_memfd_check for the special case of flags=0. > > > > > > > > OK. > > > > > > > > > > > > > > > Then I think it's easier I provide a patch, and also show that it > > > > > > can be > > > > > > also smaller changes to do the same thing, with everything fixed up > > > > > > (e.g. addressing above mmu notifier missing issue). What do you > > > > > > think as > > > > > > below? > > > > > > > > > > The key change you make is calling qemu_ram_alloc_from_fd instead of > > > > > file_ram_alloc, > > > > > which buys the xen and kvm checks for free. Sounds good, I will do > > > > > that in the > > > > > context of my patch. > > > > > > > > > > Here are some other changes in your patch, and my responses: > > > > > > > > > > I will drop the "Retrying using MAP_ANON|MAP_SHARED" message, as you > > > > > did. > > > > > > > > > > However, I am keeping QEMU_VMALLOC_ALIGN, qemu_set_cloexec, and > > > > > trace_qemu_ram_alloc_shared. > > > > > > > > I guess no huge deal on these, however since we're talking.. Is that > > > > QEMU_VMALLOC_ALIGN from qemu_anon_ram_alloc()? > > > > > > > > A quick dig tells me that it was used to be for anon THPs.. > > > > > > > > commit 36b586284e678da28df3af9fd0907d2b16f9311c > > > > Author: Avi Kivity <a...@redhat.com> > > > > Date: Mon Sep 5 11:07:05 2011 +0300 > > > > > > > > qemu_vmalloc: align properly for transparent hugepages and KVM > > > > > > > > And I'm guessing if at that time was also majorly for guest ram. > > > > > > > > Considering that this path won't make an effect until the new aux mem > > > > option is on, I'd think it better to stick without anything special like > > > > QEMU_VMALLOC_ALIGN, until it's justified to be worthwhile. E.g., Avi > > > > used > > > > to explicitly mention this in that commit message: > > > > > > > > Adjust qemu_vmalloc() to honor that requirement. Ignore it for > > > > small regions > > > > to avoid fragmentation. > > > > > > > > And this is exactly mostly small regions when it's AUX.. probably except > > > > VGA, but it'll be SHARED on top of shmem not PRIVATE on anon anyway... > > > > so > > > > it'll be totally different things. > > > > > > > > So I won't worry on that 2M alignment, and I will try to not carry over > > > > that, because then trying to remove it will be harder.. even when we > > > > want. > > > > > > Yes, currently the aux allocations get QEMU_VMALLOC_ALIGN alignment in > > > qemu_anon_ram_alloc. I do the same for the shared fd mappings to > > > guarantee > > > no performance regression, > > > > I don't know how we could guarantee that at all - anon and shmem uses > > different knobs to enable/disable THPs after all.. For example: > > > > $ ls /sys/kernel/mm/transparent_hugepage/*enabled > > /sys/kernel/mm/transparent_hugepage/enabled > > /sys/kernel/mm/transparent_hugepage/shmem_enabled > > Yes, but at least shmem_enabled is something the end user can fix. If > we bake a poor alignment into qemu, the user has no recourse. By setting > it to QEMU_VMALLOC_ALIGN, I eliminate alignment as a potential performance > issue. There is no practical downside. We should just do it, especially if > you believe "no huge deal on these" as written above :)
I'd wager nobody will be able to notice the anon/shmem difference at all, so if it really regressed nobody will be able fix it. :) Not to mention it's a global knob, and IMHO it doesn't make a lot of sense to change it for an aux mem not aligned.. while changing a global knob could OTOH break other things. But sure, if you do prefer having that I'm ok. Please still consider adding a comment then explaining where it came from.. > > > And their default values normally differ too... it means after switching to > > fd based we do face the possibility that thp can be gone at least on the > > 1st 2mb. > > > > When I was suggesting it, I was hoping thp doesn't really matter that lot > > on aux mem, even for VGA. > > > > Btw, I don't even think the alignment will affect THP allocations for the > > whole vma, anyway? I mean, it's only about the initial 2MB portion.. IOW, > > when not aligned, I think the worst case is we have <2MB at start address > > that is not using THP, but later on when it starts to align with 2MB, THPs > > will be allocated again. > > It depends on the kernel version/implementation. In 6.13, it is not that > clever for memfd_create + mmap. An unaligned start means no huge pages > anywhere > in the allocation, as shown by the page-types utility. Add > QEMU_VMALLOC_ALIGN, > and I get huge pages. > > > The challenge is more on the "fd-based" side, where shmem on most distros > > will disable THP completely. > > > > > as some of them are larger than 2M and would > > > benefit from using huge pages. The VA fragmentation is trivial for this > > > small > > > number of aux blocks in a 64-bit address space, and is no different than > > > it was > > > for qemu_anon_ram_alloc. > > > > > > > For the 2nd.. Any quick answer on why explicit qemu_set_cloexec() > > > > needed? > > > > > > qemu sets cloexec for all descriptors it opens to prevent them from > > > accidentally > > > being leaked to another process via fork+exec. > > > > But my question is why this is special? For example, we don't do that for > > "-object memory-backend-memfd", am I right? > > We should, the backends also need to set cloexec when they use a cpr fd. > I'll delete the call here and push it into cpr_find_fd. Maybe we already have that? As CPR receives fds from iochannels. I am looking at qio_channel_socket_copy_fds(), where we have: #ifndef MSG_CMSG_CLOEXEC qemu_set_cloexec(fd); #endif > > > > > For 3rd, tracepoint would definitely be fine whenever you feel > > > > necessary. > > > > > > > > > Also, when qemu_memfd_create + qemu_ram_alloc_from_fd fails, qemu > > > > > should fail and exit, > > > > > and not fall back, because something unexpected went wrong. David > > > > > said the same. > > > > > > > > Why? I was trying to rely on such fallback to make it work on e.g. Xen. > > > > In that case, Xen fails there and fallback to xen_ram_alloc() inside the > > > > later call to ram_block_add(), no? > > > > > > Why -- because something went wrong that should have worked, and we > > > should report the > > > first fault so its cause can be fixed and cpr can be used. > > > > Ahh so it's only about the corner cases where CPR could raise an error? > > Can we rely on the failure later on "migrate" command to tell which > > ramblock doesn't support it, so the user could be aware as well? > > The ramblock migration blocker will indeed tell us which block is a problem. > > But, we are throwing away potentially useful information by dropping the > first error message on the floor. We should only fall back for expected > failures. Unexpected failures mean there is something to fix. > > I can compromise and fail on errors from these: > qemu_memfd_create(name, 0, 0, 0, 0, errp); > qemu_shm_alloc(0, errp); How are we going to be sure all existing systems using RAM_SHARED ramblocks will always succeed on either memfd or sysv shm? IOW, what if there's a system that can only support mmap(MAP_SHARED) but none of the two? That's my major concern, on start failing some systems where it used to work, even if they're corner cases. > > but ignore errors from the subsequent call to qemu_ram_alloc_from_fd, > and fall back. That keeps the code simple. > > > > However, to do the above, but still quietly fallback if > > > qemu_ram_alloc_from_fd > > > fails because of xen or kvm, I would need to return different error codes > > > from > > > qemu_ram_alloc_from_fd. Doable, but requires tweaks to all occurrences of > > > qemu_ram_alloc_from_fd. > > > > > > And BTW, qemu_ram_alloc_from_fd is defined for CONFIG_POSIX only. I need > > > to modify the call site in the patch accordingly. > > > > Yep, I was thinking maybe qemu_ram_alloc_from_fd() had a stub function, > > indeed looks not.. "allocating the fd" part definitely has, which I > > remember I checked.. > > > > > Overall, I am not convinced that using qemu_ram_alloc_from_fd in this > > > patch > > > is better/simpler than my V4 patch using file_ram_alloc, plus adding xen > > > and > > > kvm_has_sync_mmu checks in qemu_ram_alloc_internal. > > > > As long as you don't need to duplicate these two checks (or duplicate any > > such check..) I'm ok. > > > > Reusing qemu_ram_alloc_from_fd() still sounds like the easiest to go. Yes > > we'll need to teach it about resize(), used_length etc. to it, but they all > > look sane to me. We didn't have those simply because we don't have use of > > them, now we want to have resizable fd-based mem, that's the right thing to > > do to support that on fd allocations. > > > > OTOH, duplicating xen/mmu checks isn't sane to me.. :( It will make the > > code harder to maintain because the 3rd qemu_ram_alloc_from_fd() in the > > future will need to duplicate it once more (or worse, forget it again until > > xen / old kernels reports a failure).. > > I'll make the necessary changes to use qemu_ram_alloc_from_fd. Thanks. -- Peter Xu