qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: QEMU PCI subsystem: what code is responsible for making accesses to


From: Michael S. Tsirkin
Subject: Re: QEMU PCI subsystem: what code is responsible for making accesses to non-mapped addresses read as -1?
Date: Sat, 20 Mar 2021 17:50:13 -0400

On Sat, Mar 20, 2021 at 08:40:13PM +0000, Peter Maydell wrote:
> On Sat, 20 Mar 2021 at 18:59, Michael S. Tsirkin <mst@redhat.com> wrote:
> >
> > On Fri, Mar 19, 2021 at 12:35:31PM +0000, Peter Maydell wrote:
> > > I'm looking at a bug reported against the QEMU arm virt board's pci-gpex
> > > PCI controller: https://bugs.launchpad.net/qemu/+bug/1918917
> > > where an attempt to write to an address within the PCI IO window
> > > where the guest hasn't mapped a BAR causes a CPU exception rather than
> > > (what I believe is) the PCI-required behaviour of writes-ignored, reads
> > > return -1.
> > >
> > > What in the QEMU PCI code is responsible for giving the PCI-spec
> > > behaviour for accesses to the PCI IO and memory windows where there
> > > is no BAR? I was expecting the generic PCI code to map a background
> > > memory region over the whole window to do this, but it looks like it
> > > doesn't...
> 
> > As far as I know, at the PCI level what happens is Master Abort
> > on PCI/PCI-X and Unsupported Request on Express.
> > PCI spec says:
> >         The host bus bridge, in PC compatible systems, must return all 1's 
> > on a read transaction and
> >         discard data on a write transaction when terminated with 
> > Master-Abort.
> >
> > We thus implement this per host e.g. on pc compatible systems by
> > calling pc_pci_as_mapping_init.
> 
> Isn't pc_pci_as_mapping_init() "put the PCI space into the system
> address space", rather than "define the default behaviour for
> accesses in PCI space" ? IIRC x86 has -1/discard for everywhere,
> though, so maybe you get that without having to do anything special.

Hmm ... I guess it's too late in the night here, and I am confused ..
will try to recall tomorrow.

> Q: if PCI device A does a bus-mastering DMA read to a PCI address
> where no other device has been mapped, does the spec require
> it to (a) get back a "transaction failed" response or (b) get
> back read-data of -1 ? It sounds like the answer based on what
> you write above is (a), device A gets a Master Abort.
> (Put another way, is the -1/discard behaviour general to PCI
> transactions or is it strictly something that happens at the host
> bridge where the host bridge turns host CPU transactions into PCI
> transactions ?)

IIUC it's the host that is doing the conversion.

> If this is host-bridge specific then I guess our current
> implementation of "leave it up to the host bridge code" makes
> sense, but it also seems like a recipe for all our host bridges
> forgetting this corner case, in the absence of support from
> the common code for making it easy/the default...

Right at some level ... I'm not against just doing this in the pci core.

> Anyway, I think that for hw/pci-host/gpex.c we would need
> to change the current
>     memory_region_init(&s->io_mmio, OBJECT(s), "gpex_mmio", UINT64_MAX);
>     [...]
>     sysbus_init_mmio(sbd, &s->io_mmio);
>     [...]
>     pci->bus = pci_register_root_bus(dev, "pcie.0", gpex_set_irq,
>                                      pci_swizzle_map_irq_fn, s, &s->io_mmio,
>                                      &s->io_ioport, 0, 4, TYPE_PCIE_BUS);
> 
> to also create a container MR with a background set of io read/write
> functions to give the -1/discard behaviour, map s->io_mmio into
> that container, and return the container as the sysbus MMIO region.
> (and same again for the IO window).
> 
> thanks
> -- PMM

Hmm. But see
commit 0fbf50b6ec126600dca115adb1563c657cc27695
Author: Marcel Apfelbaum <marcel.a@redhat.com>
Date:   Sun Nov 10 11:49:57 2013 +0200

    Revert "hw/pci: partially handle pci master abort"
    
    This reverts commit a53ae8e934cd54686875b5bcfc2f434244ee55d6.
    
    The patch being reverted introduced a low-priority memory region
    covering all 64 bit pci address space.  This exposed the following bugs
    elsewhere in the code:
     1. Some memory regions have INT64_MAX size, where the
        intent was all 64 bit address space.
        This results in a sub-page region, should be UINT64_MAX.
     2. page table rendering in exec.c ignores physical address bits
        above TARGET_PHYS_ADDR_SPACE_BITS.
        Access outside this range (e.g. from device DMA, or gdb stub)
        ends up with a wrong region.  Registering a region outside this
        range leads to page table corruption.
     3. Some regions overlap PCI hole and have same priority.
        This only works as long as no device uses the overlapping address.
    
    It doesn't look like we can resolve all issues in time for 1.7.
    Let's fix the bugs first and apply afterwards for 1.8.
    
    Signed-off-by: Marcel Apfelbaum <marcel.a@redhat.com>
    Reviewed-by: Michael S. Tsirkin <mst@redhat.com>
    Signed-off-by: Michael S. Tsirkin <mst@redhat.com>

I wonder whether we fixed all of these by now ...  I think we basically
fixed 3 and some greps could not find any more instances of 1.  Not sure
about 2 though. Should we just limit the region to
TARGET_PHYS_ADDR_SPACE_BITS?

-- 
MST




reply via email to

[Prev in Thread] Current Thread [Next in Thread]