On 7/14/22 10:54, Joao Martins wrote: > On 7/14/22 10:28, Igor Mammedov wrote: >> On Tue, 12 Jul 2022 12:35:49 +0100 >> Joao Martins <joao.m.mart...@oracle.com> wrote: >>> On 7/12/22 11:01, Joao Martins wrote: >>>> On 7/12/22 10:06, Igor Mammedov wrote: >>>>> On Mon, 11 Jul 2022 21:03:28 +0100 >>>>> Joao Martins <joao.m.mart...@oracle.com> wrote: >>>>>> On 7/11/22 16:31, Joao Martins wrote: >>>>>>> On 7/11/22 15:52, Joao Martins wrote: >>>>>>>> On 7/11/22 13:56, Igor Mammedov wrote: >>>>>>>>> On Fri, 1 Jul 2022 17:10:13 +0100 >>>>>>>>> Joao Martins <joao.m.mart...@oracle.com> wrote: >>>>>> void pc_memory_init(PCMachineState *pcms, >>>>>> MemoryRegion *system_memory, >>>>>> MemoryRegion *rom_memory, >>>>>> @@ -897,6 +953,7 @@ void pc_memory_init(PCMachineState *pcms, >>>>>> PCMachineClass *pcmc = PC_MACHINE_GET_CLASS(pcms); >>>>>> X86MachineState *x86ms = X86_MACHINE(pcms); >>>>>> hwaddr cxl_base, cxl_resv_end = 0; >>>>>> + X86CPU *cpu = X86_CPU(first_cpu); >>>>>> >>>>>> assert(machine->ram_size == x86ms->below_4g_mem_size + >>>>>> x86ms->above_4g_mem_size); >>>>>> @@ -904,6 +961,29 @@ void pc_memory_init(PCMachineState *pcms, >>>>>> linux_boot = (machine->kernel_filename != NULL); >>>>>> >>>>>> /* >>>>>> + * The HyperTransport range close to the 1T boundary is unique to >>>>>> AMD >>>>>> + * hosts with IOMMUs enabled. Restrict the ram-above-4g relocation >>>>>> + * to above 1T to AMD vCPUs only. >>>>>> + */ >>>>>> + if (IS_AMD_CPU(&cpu->env) && x86ms->above_4g_mem_size) { >>>>> >>>>> it has the same issue as pc_max_used_gpa(), i.e. >>>>> x86ms->above_4g_mem_size != 0 >>>>> doesn't mean that there isn't any memory above 4Gb nor that there aren't >>>>> any MMIO (sgx/cxl/pci64hole), that's was the reason we were are >>>>> considering >>>>> max_used_gpa >>>>> I'd prefer to keep pc_max_used_gpa(), >>>>> idea but make it work for above cases and be more generic (i.e. not to be >>>>> tied to AMD only) since 'pc_max_used_gpa() < physbits' >>>>> applies to equally >>>>> to AMD and Intel (and to trip it, one just have to configure small enough >>>>> physbits or large enough hotpluggable RAM/CXL/PCI64HOLE) >>>>> >>>> I can reproduce the issue you're thinking with basic memory hotplug. >>> >>> I was mislead by a bug that only existed in v6. Which I fixed now. >>> So any bug possibility with hotplug, SGX and CXL, or pcihole64 is simply >>> covered with: >>> >>> pc_pci_hole64_start() + pci_hole64_size; >>> >>> which is what pc_max_used_gpa() does. This works fine /without/ >>> above_4g_mem_size != 0 >>> check even without above_4g_mem_size (e.g. mem=2G,maxmem=1024G). >>> >>> And as a reminder: SGX, hotplug, CXL and pci-hole64 *require* memory above >>> 4G[*]. And part >>> of the point of us moving to pc_pci_hole64_start() was to make these all >>> work in a generic >>> way. >>> >>> So I've removed the x86ms->above_4g_mem_size != 0 check. Current patch diff >>> pasted at the end. >>> >>> [*] As reiterated here: >>> >>>> Let me see >>>> what I can come up in pc_max_used_gpa() to cover this one. I'll respond >>>> here with a proposal. >>>> >>> >>> I was over-complicating things here. It turns out nothing else is needed >>> aside in the >>> context of 1T hole. >>> >>> This is because I only need to check address space limits (as consequence of >>> pc_set_amd_above_4g_mem_start()) when pc_max_used_gpa() surprasses >>> HT_START. Which >>> requires fundamentally a value closer to 1T well beyond what 32-bit can >>> cover. So on >>> 32-bit guests this is never true and thus it things don't change behaviour >>> from current >>> default for these guests. And thus I won't break qtests and things fail >>> correctly in the >>> right places. >>> >>> Now I should say that pc_max_used_gpa() is still not returning the accurate >>> 32-bit guest >>> max used GPA value, given that I return pci hole64 end (essentially). Do >>> you still that >>> addressed out of correctness even if it doesn't matter much for the 64-bit >>> 1T case? >>> >>> If so, our only option seems to be to check phys_bits <= 32 and return max >>> CPU >>> boundary there? Unless you have something enterily different in mind? >>> >>>> I would really love to have v7.1.0 with this issue fixed but I am not very >>>> confident it is going to make it :( >>>> >>>> Meanwhile, let me know if you have thoughts on this one: >>>> >>>> https://lore.kernel.org/qemu-devel/1b2fa957-74f6-b5a9-3fc1-65c5d6830...@oracle.com/ >>>> >>>> I am going to assume that if no comments on the above that I'll keep >>>> things as is. >>>> >>>> And also, whether I can retain your ack with Bernhard's suggestion here: >>>> >>>> https://lore.kernel.org/qemu-devel/0eefb382-4ac6-4335-ca61-035babb95...@oracle.com/ >>>> >>> >>>
[...] >>> /* >>> + * The HyperTransport range close to the 1T boundary is unique to AMD >>> + * hosts with IOMMUs enabled. Restrict the ram-above-4g relocation >>> + * to above 1T to AMD vCPUs only. >>> + */ >>> + if (IS_AMD_CPU(&cpu->env)) { >>> + /* Bail out if max possible address does not cross HT range */ >>> + if (pc_max_used_gpa(pcms, pci_hole64_size) >= AMD_HT_START) { >>> + pc_set_amd_above_4g_mem_start(pcms, pci_hole64_size); >> >> I'd replace call with >> x86ms->above_4g_mem_start = AMD_ABOVE_1TB_START; >> > See below. > >>> + } >>> + >>> + /* >>> + * Advertise the HT region if address space covers the reserved >>> + * region or if we relocate. >>> + */ >>> + if (x86ms->above_4g_mem_start == AMD_ABOVE_1TB_START || >>> + cpu->phys_bits >= 40) { >>> + e820_add_entry(AMD_HT_START, AMD_HT_SIZE, E820_RESERVED); >>> + } >>> + } >> >> and then here check that pc_max_used_gpa() fits into phys_bits >> which should cover AMD case and case where pci64_hole goes beyond >> supported address range even without 1TB hole >> > > When you say 'here' you mean outside IS_AMD_CPU() ? > > If we put outside (and thus generic) where it was ... it will break qtests > as pc_max_used_gpa() does not handle 32-bit case, as mentioned earlier. > Hence why it is inside pc_set_amd_above_4g_mem_start(), or in other words > inside the scope of: > > if (pc_max_used_gpa(pcms, pci_hole64_size) >= AMD_HT_START) > > Which means I will for sure have a pci_hole64. > Making it generic to /outside/ this conditional requires addressing this > earlier comment I made: > > our only option seems to be to check phys_bits <= 32 and return max CPU > boundary there? > Here's how this patch looks like, after your comments and the above issue I am talking. The added part is inside pc_max_used_gpa(). diff --git a/hw/i386/pc.c b/hw/i386/pc.c index 668e15c8f2a6..2d85c66502d5 100644 --- a/hw/i386/pc.c +++ b/hw/i386/pc.c @@ -881,6 +881,51 @@ static uint64_t pc_get_cxl_range_end(PCMachineState *pcms) return start; } +static hwaddr pc_max_used_gpa(PCMachineState *pcms, uint64_t pci_hole64_size) +{ + X86CPU *cpu = X86_CPU(first_cpu); + + if (cpu->phys_bits <= 32) { + return (1ULL << cpu->phys_bits) - 1ULL; + } + + return pc_pci_hole64_start() + pci_hole64_size; +} + +/* + * AMD systems with an IOMMU have an additional hole close to the + * 1Tb, which are special GPAs that cannot be DMA mapped. Depending + * on kernel version, VFIO may or may not let you DMA map those ranges. + * Starting Linux v5.4 we validate it, and can't create guests on AMD machines + * with certain memory sizes. It's also wrong to use those IOVA ranges + * in detriment of leading to IOMMU INVALID_DEVICE_REQUEST or worse. + * The ranges reserved for Hyper-Transport are: + * + * FD_0000_0000h - FF_FFFF_FFFFh + * + * The ranges represent the following: + * + * Base Address Top Address Use + * + * FD_0000_0000h FD_F7FF_FFFFh Reserved interrupt address space + * FD_F800_0000h FD_F8FF_FFFFh Interrupt/EOI IntCtl + * FD_F900_0000h FD_F90F_FFFFh Legacy PIC IACK + * FD_F910_0000h FD_F91F_FFFFh System Management + * FD_F920_0000h FD_FAFF_FFFFh Reserved Page Tables + * FD_FB00_0000h FD_FBFF_FFFFh Address Translation + * FD_FC00_0000h FD_FDFF_FFFFh I/O Space + * FD_FE00_0000h FD_FFFF_FFFFh Configuration + * FE_0000_0000h FE_1FFF_FFFFh Extended Configuration/Device Messages + * FE_2000_0000h FF_FFFF_FFFFh Reserved + * + * See AMD IOMMU spec, section 2.1.2 "IOMMU Logical Topology", + * Table 3: Special Address Controls (GPA) for more information. + */ +#define AMD_HT_START 0xfd00000000UL +#define AMD_HT_END 0xffffffffffUL +#define AMD_ABOVE_1TB_START (AMD_HT_END + 1) +#define AMD_HT_SIZE (AMD_ABOVE_1TB_START - AMD_HT_START) + void pc_memory_init(PCMachineState *pcms, MemoryRegion *system_memory, MemoryRegion *rom_memory, @@ -895,7 +940,9 @@ void pc_memory_init(PCMachineState *pcms, MachineClass *mc = MACHINE_GET_CLASS(machine); PCMachineClass *pcmc = PC_MACHINE_GET_CLASS(pcms); X86MachineState *x86ms = X86_MACHINE(pcms); + hwaddr maxphysaddr, maxusedaddr; hwaddr cxl_base, cxl_resv_end = 0; + X86CPU *cpu = X86_CPU(first_cpu); assert(machine->ram_size == x86ms->below_4g_mem_size + x86ms->above_4g_mem_size); @@ -903,6 +950,40 @@ void pc_memory_init(PCMachineState *pcms, linux_boot = (machine->kernel_filename != NULL); /* + * The HyperTransport range close to the 1T boundary is unique to AMD + * hosts with IOMMUs enabled. Restrict the ram-above-4g relocation + * to above 1T to AMD vCPUs only. + */ + if (IS_AMD_CPU(&cpu->env)) { + /* Bail out if max possible address does not cross HT range */ + if (pc_max_used_gpa(pcms, pci_hole64_size) >= AMD_HT_START) { + x86ms->above_4g_mem_start = AMD_ABOVE_1TB_START; + } + + /* + * Advertise the HT region if address space covers the reserved + * region or if we relocate. + */ + if (cpu->phys_bits >= 40) { + e820_add_entry(AMD_HT_START, AMD_HT_SIZE, E820_RESERVED); + } + } + + /* + * Relocating ram-above-4G requires more than TCG_PHYS_ADDR_BITS (40). + * So make sure phys-bits is required to be appropriately sized in order + * to proceed with the above-4g-region relocation and thus boot. + */ + maxusedaddr = pc_max_used_gpa(pcms, pci_hole64_size); + maxphysaddr = ((hwaddr)1 << cpu->phys_bits) - 1; + if (maxphysaddr < maxusedaddr) { + error_report("Address space limit 0x%"PRIx64" < 0x%"PRIx64 + " phys-bits too low (%u)", + maxphysaddr, maxusedaddr, cpu->phys_bits); + exit(EXIT_FAILURE); + } + + /* * Split single memory region and use aliases to address portions of it, * done for backwards compatibility with older qemus. */