Hi Joerg,

Recently I am free and can try to work out the amd-iommu support for
kdump kernel. Now I have some plans and draft them into codes and debugging.
And also there are prlblems. I brief them here, could you please have a
look and give some suggestions?

Two parts:

1) IO page mapping
 .> Checking if it's in kdump kernel and previously enabled
 .> If yes do below operatons:
        .> Do not disable amd iommu
        .> Copy dev table form old kernel and set the old domain id in 
amd_iommu_pd_alloc_bitmap
        .> Don't call update_domain() to update device table until the first 
__map_single() is called by device driver init

2)interrupt remapping
 .> I didn't think of this well. Now I only copy the old irq table when it 
first calls get_irq_table(). 


Attach the patches here, the first 2 patches are clean up patch,
attach them too for better code understanding.

The problem happened in check_timer(). Seems timer interrupt doesn't
work well after modify_irte(). I don't know why it happened. Though I
have copied the old irte tables.

[   12.296525] ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1
[   12.302513] CPU: 0 PID: 1 Comm: swapper/0 Not tainted 4.2.0+ #18
[   12.308500] Hardware name: AMD Dinar/Dinar, BIOS RDN1505B 06/05/2013
[   12.314832]  0000000000000000 0000000085c693e9 ffff880030d6fd58
ffffffff8139746f
[   12.322239]  00000000000000a0 ffff880030d6fd90 ffffffff814b4813
ffff880030d283c0
[   12.329645]  ffff880030d2e100 0000000000000002 0000000000000000
ffff880030c29808
[   12.337052] Call Trace:
[   12.339493]  [<ffffffff8139746f>] dump_stack+0x44/0x55
[   12.344616]  [<ffffffff814b4813>] modify_irte+0x23/0xc0
[   12.349827]  [<ffffffff814b48cc>] irq_remapping_deactivate+0x1c/0x20
[   12.356162]  [<ffffffff814b48de>] irq_remapping_activate+0xe/0x10
[   12.362238]  [<ffffffff810fa6b1>] irq_domain_activate_irq+0x41/0x50
[   12.368486]  [<ffffffff810fa69b>] irq_domain_activate_irq+0x2b/0x50
[   12.374736]  [<ffffffff81d6ccbb>] setup_IO_APIC+0x33e/0x7e4
[   12.380294]  [<ffffffff81052039>] ? clear_IO_APIC+0x39/0x60
[   12.385853]  [<ffffffff81d6b82c>] apic_bsp_setup+0xa1/0xac
[   12.391323]  [<ffffffff81d69463>] native_smp_prepare_cpus+0x25f/0x2db
[   12.397747]  [<ffffffff81d550ee>] kernel_init_freeable+0xc9/0x228
[   12.403824]  [<ffffffff81762370>] ? rest_init+0x80/0x80
[   12.409034]  [<ffffffff8176237e>] kernel_init+0xe/0xe0
[   12.414158]  [<ffffffff8176e19f>] ret_from_fork+0x3f/0x70
[   12.419541]  [<ffffffff81762370>] ? rest_init+0x80/0x80
[   12.424751]   modify_irte     devid: 00:14.0 index: 2, vector:48
[   12.440491] Kernel panic - not syncing: timer doesn't work through
Interrupt-remapped IO-APIC
[   12.449022] CPU: 0 PID: 1 Comm: swapper/0 Not tainted 4.2.0+ #18
[   12.455008] Hardware name: AMD Dinar/Dinar, BIOS RDN1505B 06/05/2013
[   12.461340]  0000000000000000 0000000085c693e9 ffff880030d6fd58
ffffffff8139746f
[   12.468753]  ffffffff81a3cdf8 ffff880030d6fde0 ffffffff8119e921
0000000000000008
[   12.476165]  ffff880030d6fdf0 ffff880030d6fd88 0000000085c693e9
ffffffff813a41b5
[   12.483577] Call Trace:
[   12.486018]  [<ffffffff8139746f>] dump_stack+0x44/0x55
[   12.491142]  [<ffffffff8119e921>] panic+0xd3/0x20b
[   12.495919]  [<ffffffff813a41b5>] ? delay_tsc+0x25/0x60
[   12.501129]  [<ffffffff814bfaba>] panic_if_irq_remap+0x1a/0x20
[   12.506947]  [<ffffffff81d6ccf2>] setup_IO_APIC+0x375/0x7e4
[   12.512503]  [<ffffffff81052039>] ? clear_IO_APIC+0x39/0x60
[   12.518060]  [<ffffffff81d6b82c>] apic_bsp_setup+0xa1/0xac
[   12.523530]  [<ffffffff81d69463>] native_smp_prepare_cpus+0x25f/0x2db
[   12.529952]  [<ffffffff81d550ee>] kernel_init_freeable+0xc9/0x228
[   12.536030]  [<ffffffff81762370>] ? rest_init+0x80/0x80
[   12.541238]  [<ffffffff8176237e>] kernel_init+0xe/0xe0
[   12.546361]  [<ffffffff8176e19f>] ret_from_fork+0x3f/0x70
[   12.551745]  [<ffffffff81762370>] ? rest_init+0x80/0x80
[   12.556957] Rebooting in 10 seconds..
>From 09943d6354ee1626426f6ff060d92173bb164279 Mon Sep 17 00:00:00 2001
From: Baoquan He <b...@redhat.com>
Date: Thu, 25 Jun 2015 16:46:16 +0800
Subject: [PATCH 1/3] iommu/amd: Fix a code bug of bitmap operation

Signed-off-by: Baoquan He <b...@redhat.com>
---
 drivers/iommu/amd_iommu.c      | 2 +-
 drivers/iommu/amd_iommu_init.c | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
index 45b7581..552730b 100644
--- a/drivers/iommu/amd_iommu.c
+++ b/drivers/iommu/amd_iommu.c
@@ -1901,7 +1901,7 @@ static struct dma_ops_domain *dma_ops_domain_alloc(void)
         * mark the first page as allocated so we never return 0 as
         * a valid dma-address. So we can use 0 as error value
         */
-       dma_dom->aperture[0]->bitmap[0] = 1;
+       __set_bit(0, dma_dom->aperture[0]->bitmap);
        dma_dom->next_address = 0;
 
 
diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index f954ae8..0fe7eb4 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -21,6 +21,7 @@
 #include <linux/pci.h>
 #include <linux/acpi.h>
 #include <linux/list.h>
+#include <linux/bitmap.h>
 #include <linux/slab.h>
 #include <linux/syscore_ops.h>
 #include <linux/interrupt.h>
@@ -1939,8 +1940,7 @@ static int __init early_amd_iommu_init(void)
         * never allocate domain 0 because its used as the non-allocated and
         * error value placeholder
         */
-       amd_iommu_pd_alloc_bitmap[0] = 1;
-
+        __set_bit(0, amd_iommu_pd_alloc_bitmap);
        spin_lock_init(&amd_iommu_pd_lock);
 
        /*
-- 
2.5.0

>From 997465ee1d23f1895bcf574df7a8b19ee0dde03d Mon Sep 17 00:00:00 2001
From: Baoquan He <b...@redhat.com>
Date: Fri, 26 Jun 2015 21:06:27 +0800
Subject: [PATCH 2/3] get the device range

Signed-off-by: Baoquan He <b...@redhat.com>
---
 drivers/iommu/amd_iommu_init.c | 45 +++++++++++++++++++++++++++---------------
 1 file changed, 29 insertions(+), 16 deletions(-)

diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index 0fe7eb4..ae5f35a 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -413,14 +413,32 @@ static inline int ivhd_entry_length(u8 *ivhd)
  * This function reads the last device id the IOMMU has to handle from the PCI
  * capability header for this IOMMU
  */
-static int __init find_last_devid_on_pci(int bus, int dev, int fn, int cap_ptr)
+static int __init find_first_devid_on_pci(struct ivhd_header *h)
 {
        u32 cap;
 
-       cap = read_pci_config(bus, dev, fn, cap_ptr+MMIO_RANGE_OFFSET);
-       update_last_devid(PCI_DEVID(MMIO_GET_BUS(cap), MMIO_GET_LD(cap)));
+       cap = read_pci_config(PCI_BUS_NUM(h->devid),
+                               PCI_SLOT(h->devid),
+                               PCI_FUNC(h->devid),
+                               h->cap_ptr+MMIO_RANGE_OFFSET);
 
-       return 0;
+       return PCI_DEVID(MMIO_GET_BUS(range), MMIO_GET_FD(range));
+}
+
+/*
+ * This function reads the last device id the IOMMU has to handle from the PCI
+ * capability header for this IOMMU
+ */
+static int __init find_last_devid_on_pci(struct ivhd_header *h)
+{
+       u32 cap;
+
+       cap = read_pci_config(PCI_BUS_NUM(h->devid),
+                               PCI_SLOT(h->devid),
+                               PCI_FUNC(h->devid),
+                               h->cap_ptr+MMIO_RANGE_OFFSET);
+
+       return PCI_DEVID(MMIO_GET_BUS(range), MMIO_GET_LD(range));
 }
 
 /*
@@ -431,15 +449,14 @@ static int __init find_last_devid_from_ivhd(struct 
ivhd_header *h)
 {
        u8 *p = (void *)h, *end = (void *)h;
        struct ivhd_entry *dev;
+       u16 devid;
+
+       devid = find_last_devid_on_pci(h);
+       update_last_devid(devid);
 
        p += sizeof(*h);
        end += h->length;
 
-       find_last_devid_on_pci(PCI_BUS_NUM(h->devid),
-                       PCI_SLOT(h->devid),
-                       PCI_FUNC(h->devid),
-                       h->cap_ptr);
-
        while (p < end) {
                dev = (struct ivhd_entry *)p;
                switch (dev->type) {
@@ -1099,6 +1116,9 @@ static int __init init_iommu_one(struct amd_iommu *iommu, 
struct ivhd_header *h)
        iommu->pci_seg = h->pci_seg;
        iommu->mmio_phys = h->mmio_phys;
 
+       iommu->first_device = find_first_devid_on_pci(h);
+       iommu->last_device = find_last_devid_on_pci(h);
+
        /* Check if IVHD EFR contains proper max banks/counters */
        if ((h->efr != 0) &&
            ((h->efr & (0xF << 13)) != 0) &&
@@ -1260,16 +1280,9 @@ static int iommu_init_pci(struct amd_iommu *iommu)
 
        pci_read_config_dword(iommu->dev, cap_ptr + MMIO_CAP_HDR_OFFSET,
                              &iommu->cap);
-       pci_read_config_dword(iommu->dev, cap_ptr + MMIO_RANGE_OFFSET,
-                             &range);
        pci_read_config_dword(iommu->dev, cap_ptr + MMIO_MISC_OFFSET,
                              &misc);
 
-       iommu->first_device = PCI_DEVID(MMIO_GET_BUS(range),
-                                        MMIO_GET_FD(range));
-       iommu->last_device = PCI_DEVID(MMIO_GET_BUS(range),
-                                       MMIO_GET_LD(range));
-
        if (!(iommu->cap & (1 << IOMMU_CAP_IOTLB)))
                amd_iommu_iotlb_sup = false;
 
-- 
2.5.0

>From b8bbc536151e5134d2e41ab01daaf1d79fe2c1fb Mon Sep 17 00:00:00 2001
From: Baoquan He <b...@redhat.com>
Date: Sun, 28 Jun 2015 21:05:16 +0800
Subject: [PATCH 3/3] check if it's pre enabled

Signed-off-by: Baoquan He <b...@redhat.com>

latest
---
 drivers/iommu/amd_iommu.c       |  49 ++++++++++++++-
 drivers/iommu/amd_iommu_init.c  | 131 +++++++++++++++++++++++++++++++---------
 drivers/iommu/amd_iommu_types.h |  15 +++++
 3 files changed, 162 insertions(+), 33 deletions(-)

diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
index 552730b..286ff4e 100644
--- a/drivers/iommu/amd_iommu.c
+++ b/drivers/iommu/amd_iommu.c
@@ -2351,7 +2351,7 @@ static void update_device_table(struct protection_domain 
*domain)
 
 static void update_domain(struct protection_domain *domain)
 {
-       if (!domain->updated)
+       if (!domain->updated || translation_pre_enabled())
                return;
 
        update_device_table(domain);
@@ -2470,6 +2470,7 @@ static dma_addr_t __map_single(struct device *dev,
        unsigned long align_mask = 0;
        int i;
 
+       clear_translation_pre_enabled();
        pages = iommu_num_pages(paddr, size, PAGE_SIZE);
        paddr &= PAGE_MASK;
 
@@ -2919,6 +2920,7 @@ static int protection_domain_init(struct 
protection_domain *domain)
        domain->id = domain_id_alloc();
        if (!domain->id)
                return -ENOMEM;
+       pr_info("#########protection_domain_init() domain->id=%d \n", 
domain->id);
        INIT_LIST_HEAD(&domain->dev_list);
 
        return 0;
@@ -3641,6 +3643,7 @@ static struct irq_remap_table *get_irq_table(u16 devid, 
bool ioapic)
        struct amd_iommu *iommu;
        unsigned long flags;
        u16 alias;
+       u64 dte;
 
        write_lock_irqsave(&amd_iommu_devtable_lock, flags);
 
@@ -3682,12 +3685,33 @@ static struct irq_remap_table *get_irq_table(u16 devid, 
bool ioapic)
 
        memset(table->table, 0, MAX_IRQS_PER_TABLE * sizeof(u32));
 
+#if 0
        if (ioapic) {
                int i;
 
                for (i = 0; i < 32; ++i)
                        table->table[i] = IRTE_ALLOCATED;
        }
+#else
+       if (translation_pre_enabled()) {
+               u64 old_intr_virt;
+               dte     = amd_iommu_dev_table[devid].data[2];
+               dte &= DTE_IRQ_PHYS_ADDR_MASK;
+               old_intr_virt = ioremap_cache(dte, MAX_IRQS_PER_TABLE * 
sizeof(u32));
+               memcpy_fromio(table->table, old_intr_virt, MAX_IRQS_PER_TABLE * 
sizeof(u32));
+               iounmap(old_intr_virt);
+       }
+       else {
+
+               if (ioapic) {
+                       int i;
+
+                       for (i = 0; i < 32; ++i)
+                               table->table[i] = IRTE_ALLOCATED;
+               }
+       }
+
+#endif
 
        irq_lookup_table[devid] = table;
        set_dte_irq_entry(devid, table);
@@ -3751,6 +3775,15 @@ static int modify_irte(u16 devid, int index, union irte 
irte)
        struct amd_iommu *iommu;
        unsigned long flags;
 
+       dump_stack();
+       pr_info("  modify_irte\t devid: %02x:%02x.%x "
+                                    "index: %d, vector:%u\n",
+                                    PCI_BUS_NUM(devid),
+                                    PCI_SLOT(devid),
+                                    PCI_FUNC(devid),
+                                    index,
+                                        irte.fields.vector);
+
        iommu = amd_iommu_rlookup_table[devid];
        if (iommu == NULL)
                return -EINVAL;
@@ -3945,6 +3978,14 @@ static int irq_remapping_alloc(struct irq_domain 
*domain, unsigned int virq,
        if (devid < 0)
                return -EINVAL;
 
+       pr_info("  irq_remapping_alloc\t devid: %02x:%02x.%x "
+                                    "info->type: %02x, virq:%u,nr_irqs:%u\n",
+                                    PCI_BUS_NUM(devid),
+                                    PCI_SLOT(devid),
+                                    PCI_FUNC(devid),
+                                    info->type,
+                                       virq,
+                                       nr_irqs);
        ret = irq_domain_alloc_irqs_parent(domain, virq, nr_irqs, arg);
        if (ret < 0)
                return ret;
@@ -3954,9 +3995,10 @@ static int irq_remapping_alloc(struct irq_domain 
*domain, unsigned int virq,
                        index = info->ioapic_pin;
                else
                        ret = -ENOMEM;
-       } else {
+       } else
                index = alloc_irq_index(devid, nr_irqs);
-       }
+
+       pr_info("########### irq_remapping_alloc index =%d. \n", index);
        if (index < 0) {
                pr_warn("Failed to allocate IRTE\n");
                goto out_free_parent;
@@ -3979,6 +4021,7 @@ static int irq_remapping_alloc(struct irq_domain *domain, 
unsigned int virq,
                irq_data->chip_data = data;
                irq_data->chip = &amd_ir_chip;
                irq_remapping_prepare_irte(data, cfg, info, devid, index, i);
+               pr_info("########### irq_remapping_alloc vector:%d. \n", 
cfg->vector);
                irq_set_status_flags(virq + i, IRQ_MOVE_PCNTXT);
        }
 
diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index ae5f35a..64a80a2 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -226,6 +226,27 @@ static bool __initdata cmdline_maps;
 
 static enum iommu_init_state init_state = IOMMU_START_STATE;
 
+u8 g_pre_enabled;
+
+bool translation_pre_enabled(void)
+{
+        return !!g_pre_enabled;
+}
+
+void clear_translation_pre_enabled(void)
+{
+        g_pre_enabled = 0;
+}
+
+static void init_translation_status(struct amd_iommu *iommu)
+{
+       u32 ctrl;
+
+       ctrl = readl(iommu->mmio_base + MMIO_CONTROL_OFFSET);
+       if (ctrl & (1<<CONTROL_IOMMU_EN))
+               g_pre_enabled = 1;
+}
+
 static int amd_iommu_enable_interrupts(void);
 static int __init iommu_go_to_state(enum iommu_init_state state);
 static void init_device_table_dma(void);
@@ -422,7 +443,7 @@ static int __init find_first_devid_on_pci(struct 
ivhd_header *h)
                                PCI_FUNC(h->devid),
                                h->cap_ptr+MMIO_RANGE_OFFSET);
 
-       return PCI_DEVID(MMIO_GET_BUS(range), MMIO_GET_FD(range));
+       return PCI_DEVID(MMIO_GET_BUS(cap), MMIO_GET_FD(cap));
 }
 
 /*
@@ -438,7 +459,7 @@ static int __init find_last_devid_on_pci(struct ivhd_header 
*h)
                                PCI_FUNC(h->devid),
                                h->cap_ptr+MMIO_RANGE_OFFSET);
 
-       return PCI_DEVID(MMIO_GET_BUS(range), MMIO_GET_LD(range));
+       return PCI_DEVID(MMIO_GET_BUS(cap), MMIO_GET_LD(cap));
 }
 
 /*
@@ -712,22 +733,24 @@ static void __init set_iommu_for_device(struct amd_iommu 
*iommu, u16 devid)
 static void __init set_dev_entry_from_acpi(struct amd_iommu *iommu,
                                           u16 devid, u32 flags, u32 ext_flags)
 {
-       if (flags & ACPI_DEVFLAG_INITPASS)
-               set_dev_entry_bit(devid, DEV_ENTRY_INIT_PASS);
-       if (flags & ACPI_DEVFLAG_EXTINT)
-               set_dev_entry_bit(devid, DEV_ENTRY_EINT_PASS);
-       if (flags & ACPI_DEVFLAG_NMI)
-               set_dev_entry_bit(devid, DEV_ENTRY_NMI_PASS);
-       if (flags & ACPI_DEVFLAG_SYSMGT1)
-               set_dev_entry_bit(devid, DEV_ENTRY_SYSMGT1);
-       if (flags & ACPI_DEVFLAG_SYSMGT2)
-               set_dev_entry_bit(devid, DEV_ENTRY_SYSMGT2);
-       if (flags & ACPI_DEVFLAG_LINT0)
-               set_dev_entry_bit(devid, DEV_ENTRY_LINT0_PASS);
-       if (flags & ACPI_DEVFLAG_LINT1)
-               set_dev_entry_bit(devid, DEV_ENTRY_LINT1_PASS);
-
-       amd_iommu_apply_erratum_63(devid);
+       if ( !translation_pre_enabled()) {
+               if (flags & ACPI_DEVFLAG_INITPASS)
+                       set_dev_entry_bit(devid, DEV_ENTRY_INIT_PASS);
+               if (flags & ACPI_DEVFLAG_EXTINT)
+                       set_dev_entry_bit(devid, DEV_ENTRY_EINT_PASS);
+               if (flags & ACPI_DEVFLAG_NMI)
+                       set_dev_entry_bit(devid, DEV_ENTRY_NMI_PASS);
+               if (flags & ACPI_DEVFLAG_SYSMGT1)
+                       set_dev_entry_bit(devid, DEV_ENTRY_SYSMGT1);
+               if (flags & ACPI_DEVFLAG_SYSMGT2)
+                       set_dev_entry_bit(devid, DEV_ENTRY_SYSMGT2);
+               if (flags & ACPI_DEVFLAG_LINT0)
+                       set_dev_entry_bit(devid, DEV_ENTRY_LINT0_PASS);
+               if (flags & ACPI_DEVFLAG_LINT1)
+                       set_dev_entry_bit(devid, DEV_ENTRY_LINT1_PASS);
+
+               amd_iommu_apply_erratum_63(devid);
+       }
 
        set_iommu_for_device(iommu, devid);
 }
@@ -811,7 +834,8 @@ static void __init set_device_exclusion_range(u16 devid, 
struct ivmd_header *m)
                 * per device. But we can enable the exclusion range per
                 * device. This is done here
                 */
-               set_dev_entry_bit(devid, DEV_ENTRY_EX);
+               if (!translation_pre_enabled())
+                       set_dev_entry_bit(devid, DEV_ENTRY_EX);
                iommu->exclusion_start = m->range_start;
                iommu->exclusion_length = m->range_length;
        }
@@ -1093,6 +1117,7 @@ static void amd_iommu_erratum_746_workaround(struct 
amd_iommu *iommu)
 static int __init init_iommu_one(struct amd_iommu *iommu, struct ivhd_header 
*h)
 {
        int ret;
+       u32 ctrl;
 
        spin_lock_init(&iommu->lock);
 
@@ -1143,6 +1168,9 @@ static int __init init_iommu_one(struct amd_iommu *iommu, 
struct ivhd_header *h)
 
        iommu->int_enabled = false;
 
+       init_translation_status(iommu);
+       pr_info("###g_pre_enabled=%d\n", g_pre_enabled);
+
        ret = init_iommu_from_acpi(iommu, h);
        if (ret)
                return ret;
@@ -1405,7 +1433,8 @@ static int __init amd_iommu_init_pci(void)
                        break;
        }
 
-       init_device_table_dma();
+       if ( !translation_pre_enabled() )
+               init_device_table_dma();
 
        for_each_iommu(iommu)
                iommu_flush_all_caches(iommu);
@@ -1689,6 +1718,38 @@ static void iommu_apply_resume_quirks(struct amd_iommu 
*iommu)
                               iommu->stored_addr_lo | 1);
 }
 
+static void copy_dev_tables(void)
+{
+       u64 entry;
+       u32 lo, hi;
+       phys_addr_t old_devtb_phys;
+       u64 old_devtb_virt;
+       struct dev_table_entry *old_devtb;
+       struct amd_iommu *iommu;
+       u16 dom_id;
+       u32 devid;
+
+       for_each_iommu(iommu) {
+               lo = readl(iommu->mmio_base + MMIO_DEV_TABLE_OFFSET);
+               hi = readl(iommu->mmio_base + MMIO_DEV_TABLE_OFFSET + 4);
+               entry = (((u64) hi) << 32) + lo;
+               old_devtb_phys = entry & PAGE_MASK;
+               old_devtb_virt = ioremap_cache(old_devtb_phys, dev_table_size);
+               old_devtb = (struct dev_table_entry*) old_devtb_virt;
+               for (devid = 0; devid <= amd_iommu_last_bdf; ++devid) {
+                       amd_iommu_dev_table[devid] = old_devtb[devid];
+                       dom_id = amd_iommu_dev_table[devid].data[1] & 
DEV_DOMID_MASK;
+                       pr_info("###### copy_dev_tables 
dom_id=%u,data[1]=0x%llx old.data[1]=0x%llx\n",
+                                       dom_id, 
amd_iommu_dev_table[devid].data[1], old_devtb[devid].data[1]);
+                       __set_bit(dom_id, amd_iommu_pd_alloc_bitmap);
+                       pr_info("###### copy_dev_tables bitmap=0x%lx\n", 
amd_iommu_pd_alloc_bitmap[0]);
+               }
+               iounmap(old_devtb);
+       }
+
+}
+
+
 /*
  * This function finally enables all IOMMUs found in the system after
  * they have been initialized
@@ -1698,14 +1759,21 @@ static void early_enable_iommus(void)
        struct amd_iommu *iommu;
 
        for_each_iommu(iommu) {
-               iommu_disable(iommu);
-               iommu_init_flags(iommu);
-               iommu_set_device_table(iommu);
-               iommu_enable_command_buffer(iommu);
-               iommu_enable_event_buffer(iommu);
-               iommu_set_exclusion_range(iommu);
-               iommu_enable(iommu);
-               iommu_flush_all_caches(iommu);
+               if ( !translation_pre_enabled() ) {
+                       iommu_disable(iommu);
+                       iommu_init_flags(iommu);
+                       iommu_set_device_table(iommu);
+                       iommu_enable_command_buffer(iommu);
+                       iommu_enable_event_buffer(iommu);
+                       iommu_set_exclusion_range(iommu);
+                       iommu_enable(iommu);
+                       iommu_flush_all_caches(iommu); 
+               } else {
+                       copy_dev_tables();
+                       iommu_enable_command_buffer(iommu);
+                       iommu_enable_event_buffer(iommu);
+                       //iommu_flush_all_caches(iommu); 
+               }
        }
 }
 
@@ -1989,10 +2057,11 @@ static int __init early_amd_iommu_init(void)
 
        ret = init_memory_definitions(ivrs_base);
        if (ret)
-               goto out;
+               goto out; 
 
        /* init the device table */
-       init_device_table();
+       if (!g_pre_enabled)
+               init_device_table();
 
 out:
        /* Don't leak any ACPI memory */
@@ -2002,6 +2071,8 @@ out:
        return ret;
 }
 
+
+
 static int amd_iommu_enable_interrupts(void)
 {
        struct amd_iommu *iommu;
diff --git a/drivers/iommu/amd_iommu_types.h b/drivers/iommu/amd_iommu_types.h
index f659088..5f5e1ed 100644
--- a/drivers/iommu/amd_iommu_types.h
+++ b/drivers/iommu/amd_iommu_types.h
@@ -194,6 +194,12 @@
 #define DEV_ENTRY_MODE_MASK    0x07
 #define DEV_ENTRY_MODE_SHIFT   0x09
 
+
+/* kdump fix */
+#define DEV_PTE_ROOT_LEN 52
+#define DEV_DOMID_MASK        0xffff
+
+
 #define MAX_DEV_TABLE_ENTRIES  0xffff
 
 /* constants to configure the command buffer */
@@ -205,6 +211,7 @@
 
 /* constants for event buffer handling */
 #define EVT_BUFFER_SIZE                8192 /* 512 entries */
+#define MMIO_EVT_SIZE_SHIFT 56
 #define EVT_LEN_MASK           (0x9ULL << 56)
 
 /* Constants for PPR Log handling */
@@ -294,6 +301,7 @@
 #define IOMMU_PTE_FC (1ULL << 60)
 #define IOMMU_PTE_IR (1ULL << 61)
 #define IOMMU_PTE_IW (1ULL << 62)
+#define IOMMU_PTE_IW (1ULL << 62)
 
 #define DTE_FLAG_IOTLB (0x01UL << 32)
 #define DTE_FLAG_GV    (0x01ULL << 55)
@@ -463,6 +471,9 @@ struct dma_ops_domain {
        bool need_flush;
 };
 
+#define IOMMU_FLAG_TRANS_PRE_ENABLED     (1 << 0)
+#define IOMMU_FLAG_IRQ_REMAP_PRE_ENABLED (1 << 1)
+
 /*
  * Structure where we save information about one hardware AMD IOMMU in the
  * system.
@@ -573,6 +584,7 @@ struct amd_iommu {
        struct irq_domain *ir_domain;
        struct irq_domain *msi_domain;
 #endif
+       u32     flags;
 };
 
 struct devid_map {
@@ -686,6 +698,9 @@ extern bool amd_iommu_force_isolation;
 /* Max levels of glxval supported */
 extern int amd_iommu_max_glx_val;
 
+extern u8 g_pre_enabled;
+extern bool translation_pre_enabled(void);
+extern void clear_translation_pre_enabled(void);
 /*
  * This function flushes all internal caches of
  * the IOMMU used by this driver.
-- 
2.5.0

_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to