Loading arch/x86/include/asm/amd_iommu_types.h +4 −0 Original line number Diff line number Diff line Loading @@ -316,6 +316,10 @@ struct iommu_dev_data { struct protection_domain *domain; /* Domain the device is bound to */ atomic_t bind; /* Domain attach reverent count */ u16 devid; /* PCI Device ID */ struct { bool enabled; int qdep; } ats; /* ATS state */ }; /* Loading arch/x86/kernel/amd_iommu.c +24 −20 Original line number Diff line number Diff line Loading @@ -695,17 +695,16 @@ void iommu_flush_all_caches(struct amd_iommu *iommu) */ static int device_flush_iotlb(struct device *dev, u64 address, size_t size) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; struct amd_iommu *iommu; struct iommu_cmd cmd; u16 devid; int qdep; qdep = pci_ats_queue_depth(pdev); devid = get_device_id(dev); iommu = amd_iommu_rlookup_table[devid]; dev_data = get_dev_data(dev); qdep = dev_data->ats.qdep; iommu = amd_iommu_rlookup_table[dev_data->devid]; build_inv_iotlb_pages(&cmd, devid, qdep, address, size); build_inv_iotlb_pages(&cmd, dev_data->devid, qdep, address, size); return iommu_queue_command(iommu, &cmd); } Loading @@ -728,7 +727,7 @@ static int device_flush_dte(struct device *dev) if (ret) return ret; if (pci_ats_enabled(pdev)) if (dev_data->ats.enabled) ret = device_flush_iotlb(dev, 0, ~0UL); return ret; Loading Loading @@ -760,9 +759,8 @@ static void __domain_flush_pages(struct protection_domain *domain, } list_for_each_entry(dev_data, &domain->dev_list, list) { struct pci_dev *pdev = to_pci_dev(dev_data->dev); if (!pci_ats_enabled(pdev)) if (!dev_data->ats.enabled) continue; ret |= device_flush_iotlb(dev_data->dev, address, size); Loading Loading @@ -1576,8 +1574,7 @@ static void do_attach(struct device *dev, struct protection_domain *domain) iommu = amd_iommu_rlookup_table[dev_data->devid]; pdev = to_pci_dev(dev); if (amd_iommu_iotlb_sup) ats = pci_ats_enabled(pdev); ats = dev_data->ats.enabled; /* Update data structures */ dev_data->domain = domain; Loading Loading @@ -1674,11 +1671,16 @@ static int attach_device(struct device *dev, struct protection_domain *domain) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; unsigned long flags; int ret; if (amd_iommu_iotlb_sup) pci_enable_ats(pdev, PAGE_SHIFT); dev_data = get_dev_data(dev); if (amd_iommu_iotlb_sup && pci_enable_ats(pdev, PAGE_SHIFT) == 0) { dev_data->ats.enabled = true; dev_data->ats.qdep = pci_ats_queue_depth(pdev); } write_lock_irqsave(&amd_iommu_devtable_lock, flags); ret = __attach_device(dev, domain); Loading Loading @@ -1736,7 +1738,7 @@ static void __detach_device(struct device *dev) */ static void detach_device(struct device *dev) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; unsigned long flags; /* lock device table */ Loading @@ -1744,8 +1746,12 @@ static void detach_device(struct device *dev) __detach_device(dev); write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); if (amd_iommu_iotlb_sup && pci_ats_enabled(pdev)) pci_disable_ats(pdev); dev_data = get_dev_data(dev); if (dev_data->ats.enabled) { pci_disable_ats(to_pci_dev(dev)); dev_data->ats.enabled = false; } } /* Loading Loading @@ -1890,10 +1896,8 @@ static void update_device_table(struct protection_domain *domain) { struct iommu_dev_data *dev_data; list_for_each_entry(dev_data, &domain->dev_list, list) { struct pci_dev *pdev = to_pci_dev(dev_data->dev); set_dte_entry(dev_data->devid, domain, pci_ats_enabled(pdev)); } list_for_each_entry(dev_data, &domain->dev_list, list) set_dte_entry(dev_data->devid, domain, dev_data->ats.enabled); } static void update_domain(struct protection_domain *domain) Loading Loading
arch/x86/include/asm/amd_iommu_types.h +4 −0 Original line number Diff line number Diff line Loading @@ -316,6 +316,10 @@ struct iommu_dev_data { struct protection_domain *domain; /* Domain the device is bound to */ atomic_t bind; /* Domain attach reverent count */ u16 devid; /* PCI Device ID */ struct { bool enabled; int qdep; } ats; /* ATS state */ }; /* Loading
arch/x86/kernel/amd_iommu.c +24 −20 Original line number Diff line number Diff line Loading @@ -695,17 +695,16 @@ void iommu_flush_all_caches(struct amd_iommu *iommu) */ static int device_flush_iotlb(struct device *dev, u64 address, size_t size) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; struct amd_iommu *iommu; struct iommu_cmd cmd; u16 devid; int qdep; qdep = pci_ats_queue_depth(pdev); devid = get_device_id(dev); iommu = amd_iommu_rlookup_table[devid]; dev_data = get_dev_data(dev); qdep = dev_data->ats.qdep; iommu = amd_iommu_rlookup_table[dev_data->devid]; build_inv_iotlb_pages(&cmd, devid, qdep, address, size); build_inv_iotlb_pages(&cmd, dev_data->devid, qdep, address, size); return iommu_queue_command(iommu, &cmd); } Loading @@ -728,7 +727,7 @@ static int device_flush_dte(struct device *dev) if (ret) return ret; if (pci_ats_enabled(pdev)) if (dev_data->ats.enabled) ret = device_flush_iotlb(dev, 0, ~0UL); return ret; Loading Loading @@ -760,9 +759,8 @@ static void __domain_flush_pages(struct protection_domain *domain, } list_for_each_entry(dev_data, &domain->dev_list, list) { struct pci_dev *pdev = to_pci_dev(dev_data->dev); if (!pci_ats_enabled(pdev)) if (!dev_data->ats.enabled) continue; ret |= device_flush_iotlb(dev_data->dev, address, size); Loading Loading @@ -1576,8 +1574,7 @@ static void do_attach(struct device *dev, struct protection_domain *domain) iommu = amd_iommu_rlookup_table[dev_data->devid]; pdev = to_pci_dev(dev); if (amd_iommu_iotlb_sup) ats = pci_ats_enabled(pdev); ats = dev_data->ats.enabled; /* Update data structures */ dev_data->domain = domain; Loading Loading @@ -1674,11 +1671,16 @@ static int attach_device(struct device *dev, struct protection_domain *domain) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; unsigned long flags; int ret; if (amd_iommu_iotlb_sup) pci_enable_ats(pdev, PAGE_SHIFT); dev_data = get_dev_data(dev); if (amd_iommu_iotlb_sup && pci_enable_ats(pdev, PAGE_SHIFT) == 0) { dev_data->ats.enabled = true; dev_data->ats.qdep = pci_ats_queue_depth(pdev); } write_lock_irqsave(&amd_iommu_devtable_lock, flags); ret = __attach_device(dev, domain); Loading Loading @@ -1736,7 +1738,7 @@ static void __detach_device(struct device *dev) */ static void detach_device(struct device *dev) { struct pci_dev *pdev = to_pci_dev(dev); struct iommu_dev_data *dev_data; unsigned long flags; /* lock device table */ Loading @@ -1744,8 +1746,12 @@ static void detach_device(struct device *dev) __detach_device(dev); write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); if (amd_iommu_iotlb_sup && pci_ats_enabled(pdev)) pci_disable_ats(pdev); dev_data = get_dev_data(dev); if (dev_data->ats.enabled) { pci_disable_ats(to_pci_dev(dev)); dev_data->ats.enabled = false; } } /* Loading Loading @@ -1890,10 +1896,8 @@ static void update_device_table(struct protection_domain *domain) { struct iommu_dev_data *dev_data; list_for_each_entry(dev_data, &domain->dev_list, list) { struct pci_dev *pdev = to_pci_dev(dev_data->dev); set_dte_entry(dev_data->devid, domain, pci_ats_enabled(pdev)); } list_for_each_entry(dev_data, &domain->dev_list, list) set_dte_entry(dev_data->devid, domain, dev_data->ats.enabled); } static void update_domain(struct protection_domain *domain) Loading