> This solved all the problems with BAR allocation.
Nice!

> Do we have already some code for ROM BAR somewhere?
Henrik Gulbrandsen was working on expansion ROM support last July (CC'd). He 
posted his work up on https://www.gulbra.net/freebsd-bhyve/.


‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Thursday, 19 March 2020 03:52, Alex Erley <erl...@gmail.com> wrote:

> Hello,
>
> You are right about mapping 64-bit BARs.
>
> === (1) ===
> The initial value of PCI_EMUL_MEMBASE64 is 0xD000000000 and it
> corresponds to the 64-bit mmio window set by QWordMemory(...)
> in DSDT from pci_bus_write_dsdt(), so guest VM gets from
> UEFI BIOS this 64-bit window:
> 0x000000D000000000-0x000000D0000FFFFF
> and it is ignored by guest kernel during boot:
> ACPI: PCI Root Bridge [PC00] (domain 0000 [bus 00-ff])
> acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM
> Segments MSI]
> acpi PNP0A03:00: _OSC failed (AE_NOT_FOUND); disabling ASPM
> acpi PNP0A03:00: host bridge window [mem 0xd000000000-0xd0000fffff
> window] (ignored, not CPU addressable)
> acpi PNP0A03:00: host bridge window expanded to [io 0x0000-0x0cf7];
> [io 0x0000-0x0cf7 window] ignored
>
> 64-bit BAR code must be fixed to set proper memory region depending on
> host and guest resources.
> In general it makes sense only for huge BAR sizes, but in most cases all
> the guest devices should fit in available 32-bit mmio window.
> I decided to put it aside and concentrate my efforts on 32-bit BARs.
>
> So, I increased the max size in pci_emul_alloc_pbar() from 32 to 256 Mb
> to have all BARs allocated in 32-bit mmio window (for me it still looks
> like a hack).
> As I already tried it before, BHyve failed on VM start and I turned to
> debug what happened.
>
> Allocation goes in 32-bit mmio window 0xc0000000-0xdfffffff (512Mb).
> Keeping in mind these two requirements from PCI standard:
>
> -   size of each region must be a power of two (it is already OK),
>
> -   base address must be aligned on a boundary equal to the region size,
>     BARs are allocated in the order of their indices:
>     BAR[0]: type=2, size=0x1000000 => addr=0xc0000000, nextaddr=0xc1000000
>     BAR[1]: type=3, size=0x10000000=> addr=0xd0000000, nextaddr=0xe0000000
>     BAR[3]: type=3, size=0x2000000 => KO, 0xe0000000+size > 0xdfffffff
>     BAR[5]: type=1, size=0x80
>
>     I fixed cfginitbar() in pci_passthru.c to allocate BARs in different
>     order to make allocations more compact. They have to be allocated from
>     bigger claimed size to smaller (btw, as it is done on host system).
>     For this, in cfginitbar() I reordered BAR indices before calling
>     pci_emul_alloc_pbar().
>     This solved all the problems with BAR allocation.
>     (will share my patch when review it a bit more)
>
>     === (2) ===
>     As you pointed before, now I faced the ROM mapping problem.
>
>     All the ROM BARs are disabled in guest VM and quick look shows that
>     as PCI_BARMAX==5 (=PCIR_MAX_BAR_0 from <dev/pci/pcireg.h>),
>
>
> so only 6 BARs initialized in cfginitbar() at pci_passthru.c.
> It corresponds to registers 10h, 14h, 18h, 1ch, 20h, 24h.
> Expansion ROM Base Address register 30h is not initialized at all.
>
> I'd like to add missing code for it in pci_passthru.c.
> Physical guest memory map described in pci_emul.c says all ROM BARs
> must be allocated in 8Mb window 0xf0000000-0xf07fffff and it already
> matches generated DSDT.
>
> From PCI documentation dealing with ROM doesn't seem to be very
> complicated, but anyway, I'm interested in
>
> -   What pitfalls will be on that way?
> -   Do we have already some code for ROM BAR somewhere?
>
>     Any help is welcome.
>
>     Have a nice day,
>     Alex
>
>     On 3/15/20 3:20 PM, Robert Crowston wrote:
>
>
> > I suggest you map the BAR into the 32 bit address space, unless you have so 
> > many PCI devices that this is not feasible. Just raise the limit of the 
> > special 64 bit handling to 1 GB or something big.
> >
> > -   Many/most(?) consumer BIOS/UEFIs map 64 bit bars into the 32 bit 
> > address space by default, so this configuration is much more tested for 
> > device drivers and guest operating systems.
> > -   Passthrough doesn't work for me at all above the 4GB memory window on 
> > my recent AMD system. It exits back to the hypervisor, which then crashes 
> > with a failed assert because it doesn't expect to handle this.
> >
> > With this tweak it is possible to use the open source nVidia driver on 
> > Linux.
> > However, your next problem---if you want to use proprietary drivers---is 
> > that access to the ROM BAR is not supported or emulated. If that could be 
> > fixed, it would be a big step forwards.
> > ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
> > On Saturday, 14 March 2020 12:50, Alex Erley erl...@gmail.com wrote:
> >
> > > Hello,
> > > Some new findings to share.
> > >
> > > 1.  Changing PCI_EMUL_MEMBASE64 from 0xD000000000 to any value
> > >     below 0x0440000000 makes bhyve fail when starting VM with message:
> > >     bhyve: failed to initialize BARs for PCI 1/0/0
> > >     device emulation initialization error: Cannot allocate memory
> > >
> > > 2.  Having PCI_EMUL_MEMBASE64 set to 0x0440000000 (or above) guest VM
> > >     can not configure BARs of pass-through device properly.
> > >     == (a) ==
> > >     On BHyve host ppt device is:
> > >
> > >
> > > > devinfo -rv
> > >
> > > ...
> > > pci0
> > > hostb0 at slot=0 function=0 dbsf=pci0:0:0:0
> > > pcib1 at slot=1 function=0 dbsf=pci0:0:1:0 handle=\SB.PCI0.P0P2
> > > I/O ports: 0xe000-0xefff
> > > I/O memory addresses:
> > > 0x00c0000000-0x00d30fffff <-- covers all child mem windows
> > > pci1
> > > ppt0 at slot=0 function=0 dbsf=pci0:1:0:0
> > > pcib1 I/O port window: 0xe000-0xe07f
> > > pcib1 memory window:
> > > 0x00c0000000-0x00cfffffff <-- 256M
> > > 0x00d0000000-0x00d1ffffff <-- 32M
> > > 0x00d2000000-0x00d2ffffff <-- 16M
> > > ppt1 at slot=0 function=1 dbsf=pci0:1:0:1
> > > pcib1 memory window:
> > > 0xd3080000-0xd3083fff <-- 16K
> > > ...
> > > and there is no other device attached to pci1.
> > > == (b) ==
> > > On guest VM dmesg shows (timestamps are removed):
> > > ...
> > > BIOS-provided physical RAM map:
> > > BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable
> > > BIOS-e820: [mem 0x0000000000100000-0x00000000bea95fff] usable
> > > BIOS-e820: [mem 0x00000000bea96000-0x00000000bea97fff] reserved
> > > BIOS-e820: [mem 0x00000000bea98000-0x00000000bea99fff] ACPI data
> > > BIOS-e820: [mem 0x00000000bea9a000-0x00000000beaa8fff] reserved
> > > BIOS-e820: [mem 0x00000000beaa9000-0x00000000bfb28fff] usable
> > > BIOS-e820: [mem 0x00000000bfb29000-0x00000000bfb58fff] type 20
> > > BIOS-e820: [mem 0x00000000bfb59000-0x00000000bfb7cfff] reserved
> > > BIOS-e820: [mem 0x00000000bfb7d000-0x00000000bfb81fff] usable
> > > BIOS-e820: [mem 0x00000000bfb82000-0x00000000bfb88fff] ACPI data
> > > BIOS-e820: [mem 0x00000000bfb89000-0x00000000bfb8cfff] ACPI NVS
> > > BIOS-e820: [mem 0x00000000bfb8d000-0x00000000bffcffff] usable
> > > BIOS-e820: [mem 0x00000000bffd0000-0x00000000bffeffff] reserved
> > > BIOS-e820: [mem 0x00000000bfff0000-0x00000000bfffffff] usable
> > > BIOS-e820: [mem 0x0000000100000000-0x000000043fffffff] usable
> > > ^^^-upper limit for adressable memory
> > > ...
> > > PM: Registered nosave memory: [mem 0xc0000000-0xffffffff]
> > > [mem 0xc0000000-0xffffffff] available for PCI devices
> > > ...
> > > pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7]
> > > pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window]
> > > pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
> > > ^-- 128K
> > > pci_bus 0000:00: root bus resource [mem 0xc0000000-0xdfffffff window]
> > > ^-- 512M
> > > pci_bus 0000:00: root bus resource [mem 0xf0000000-0xf07fffff window]
> > > ^-- 8M
> > > pci_bus 0000:00: root bus resource [bus 00-ff]
> > > == (c) ==
> > > Until now all runs OK.
> > > Guest Linux then allocates memory regions for devices.
> > > Allocation is done from lower reg (0x10) to higher (0x30)
> > > for each device (i.e. from 00.0 to 1f.0) on PCI bus.
> > > Here I reordered dmesg output to groups to show continuous RAM regions:
> > > (pass-through device is marked with )
> > > pci 0000:00:01.0: reg 0x24: [io 0x2000-0x207f]
> > > pci 0000:00:02.0: reg 0x10: [io 0x2080-0x209f]
> > > pci 0000:00:03.0: reg 0x10: [io 0x20c0-0x20ff]
> > > ...
> > > pci 0000:00:00.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > pci 0000:00:02.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > pci 0000:00:03.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > pci 0000:00:1d.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > pci 0000:00:1e.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > pci 0000:00:1f.0: reg 0x30: [mem 0x00000000-0x000007ff pref]
> > > ...
> > > pci 0000:00:01.0: reg 0x10:[mem 0xc0000000-0xc0ffffff] 16M... 
> > > 0xc1000000-0xc1ffffff 16M gap
> > > pci 0000:00:01.0: reg 0x1c:[mem 0xc2000000-0xc3ffffff 64bit pref] 32M
> > > pci 0000:00:01.1: reg 0x10:[mem 0xc4000000-0xc4003fff]pci 0000:00:02.0: 
> > > reg 0x14: [mem 0xc4004000-0xc4005fff]
> > > pci 0000:00:03.0: reg 0x14: [mem 0xc4006000-0xc4007fff]
> > > pci 0000:00:1d.0: reg 0x10: [mem 0xc4008000-0xc400807f]
> > > ... 0xc4008080-0xc4ffffff <16M gap
> > > pci 0000:00:1d.0: reg 0x14: [mem 0xc5000000-0xc5ffffff] 16M
> > > pci 0000:00:1e.0: reg 0x10: [mem 0xc6000000-0xc6000fff]
> > > ... 0xc6001000-0xd2ffffff <208M gap
> > > pci 0000:00:01.0: reg 0x30:[mem 0xd3000000-0xd307ffff pref] 512K
> > > 0xd3080000-0xdfffffff <208M gap
> > > pci 0000:00:01.0: reg0x14:[mem 0x440000000-0x44fffffff 64bit pref] 
> > > 256M^^^- this value is outside allowed range
> > > == (d) ==
> > > So, there is no window for 256M BAR, although there are 2 big gapes
> > > of 208M in 512M space provided for BAR allocation by PCI bus.
> > > So, BAR reg 0x14 of size 256M for device 01.0 must be inside provisioned
> > > 512M region 0xc0000000-0xdfffffff.
> > > But refering to (1) above, setting base address to any value below
> > > 0x440000000 breaks bhyve on start.
> > > According to (b), this value corresponds to upper addressable memory
> > > limit in guest VM.
> > > So I'm blocked here at the moment:
> > >
> > > -   Guest VM requires a value which BHyve doesn't like.
> > >
> > > -   Guest VM allocates BARs with huge gapes.
> > >     I have little knowledge about PCI bus internals, although I already 
> > > read
> > >     some articles on internet.
> > >     Could it be some ACPI trick to do?
> > >     I'd be happy to hear any ideas...
> > >     PS
> > >     I suspect that if I take other OS as a guest VM or other pass-through
> > >     GPU model, it would probably allocate BARs properly.
> > >     But this is not what I want for this config.
> > >     There should be a way to allocate 256M BAR in guest Linux.
> > >     Have a nice day,
> > >     Alex
> > >
> > >
> > > freebsd-virtualization@freebsd.org mailing list
> > > https://lists.freebsd.org/mailman/listinfo/freebsd-virtualization
> > > To unsubscribe, send any mail to 
> > > "freebsd-virtualization-unsubscr...@freebsd.org"


_______________________________________________
freebsd-virtualization@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-virtualization
To unsubscribe, send any mail to 
"freebsd-virtualization-unsubscr...@freebsd.org"

Reply via email to