x86/amd-iommu: Store ATS state in dev_data
This allows the low-level functions to operate on dev_data exclusivly later. Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
This commit is contained in:
		
							parent
							
								
									f62dda66b5
								
							
						
					
					
						commit
						ea61cddb9d
					
				@ -316,6 +316,10 @@ struct iommu_dev_data {
 | 
			
		||||
	struct protection_domain *domain; /* Domain the device is bound to */
 | 
			
		||||
	atomic_t bind;			  /* Domain attach reverent count */
 | 
			
		||||
	u16 devid;			  /* PCI Device ID */
 | 
			
		||||
	struct {
 | 
			
		||||
		bool enabled;
 | 
			
		||||
		int qdep;
 | 
			
		||||
	} ats;				  /* ATS state */
 | 
			
		||||
};
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 | 
			
		||||
@ -695,17 +695,16 @@ void iommu_flush_all_caches(struct amd_iommu *iommu)
 | 
			
		||||
 */
 | 
			
		||||
static int device_flush_iotlb(struct device *dev, u64 address, size_t size)
 | 
			
		||||
{
 | 
			
		||||
	struct pci_dev *pdev = to_pci_dev(dev);
 | 
			
		||||
	struct iommu_dev_data *dev_data;
 | 
			
		||||
	struct amd_iommu *iommu;
 | 
			
		||||
	struct iommu_cmd cmd;
 | 
			
		||||
	u16 devid;
 | 
			
		||||
	int qdep;
 | 
			
		||||
 | 
			
		||||
	qdep  = pci_ats_queue_depth(pdev);
 | 
			
		||||
	devid = get_device_id(dev);
 | 
			
		||||
	iommu = amd_iommu_rlookup_table[devid];
 | 
			
		||||
	dev_data = get_dev_data(dev);
 | 
			
		||||
	qdep     = dev_data->ats.qdep;
 | 
			
		||||
	iommu    = amd_iommu_rlookup_table[dev_data->devid];
 | 
			
		||||
 | 
			
		||||
	build_inv_iotlb_pages(&cmd, devid, qdep, address, size);
 | 
			
		||||
	build_inv_iotlb_pages(&cmd, dev_data->devid, qdep, address, size);
 | 
			
		||||
 | 
			
		||||
	return iommu_queue_command(iommu, &cmd);
 | 
			
		||||
}
 | 
			
		||||
@ -728,7 +727,7 @@ static int device_flush_dte(struct device *dev)
 | 
			
		||||
	if (ret)
 | 
			
		||||
		return ret;
 | 
			
		||||
 | 
			
		||||
	if (pci_ats_enabled(pdev))
 | 
			
		||||
	if (dev_data->ats.enabled)
 | 
			
		||||
		ret = device_flush_iotlb(dev, 0, ~0UL);
 | 
			
		||||
 | 
			
		||||
	return ret;
 | 
			
		||||
@ -760,9 +759,8 @@ static void __domain_flush_pages(struct protection_domain *domain,
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	list_for_each_entry(dev_data, &domain->dev_list, list) {
 | 
			
		||||
		struct pci_dev *pdev = to_pci_dev(dev_data->dev);
 | 
			
		||||
 | 
			
		||||
		if (!pci_ats_enabled(pdev))
 | 
			
		||||
		if (!dev_data->ats.enabled)
 | 
			
		||||
			continue;
 | 
			
		||||
 | 
			
		||||
		ret |= device_flush_iotlb(dev_data->dev, address, size);
 | 
			
		||||
@ -1576,8 +1574,7 @@ static void do_attach(struct device *dev, struct protection_domain *domain)
 | 
			
		||||
	iommu    = amd_iommu_rlookup_table[dev_data->devid];
 | 
			
		||||
	pdev     = to_pci_dev(dev);
 | 
			
		||||
 | 
			
		||||
	if (amd_iommu_iotlb_sup)
 | 
			
		||||
		ats = pci_ats_enabled(pdev);
 | 
			
		||||
	ats = dev_data->ats.enabled;
 | 
			
		||||
 | 
			
		||||
	/* Update data structures */
 | 
			
		||||
	dev_data->domain = domain;
 | 
			
		||||
@ -1674,11 +1671,16 @@ static int attach_device(struct device *dev,
 | 
			
		||||
			 struct protection_domain *domain)
 | 
			
		||||
{
 | 
			
		||||
	struct pci_dev *pdev = to_pci_dev(dev);
 | 
			
		||||
	struct iommu_dev_data *dev_data;
 | 
			
		||||
	unsigned long flags;
 | 
			
		||||
	int ret;
 | 
			
		||||
 | 
			
		||||
	if (amd_iommu_iotlb_sup)
 | 
			
		||||
		pci_enable_ats(pdev, PAGE_SHIFT);
 | 
			
		||||
	dev_data = get_dev_data(dev);
 | 
			
		||||
 | 
			
		||||
	if (amd_iommu_iotlb_sup && pci_enable_ats(pdev, PAGE_SHIFT) == 0) {
 | 
			
		||||
		dev_data->ats.enabled = true;
 | 
			
		||||
		dev_data->ats.qdep    = pci_ats_queue_depth(pdev);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	write_lock_irqsave(&amd_iommu_devtable_lock, flags);
 | 
			
		||||
	ret = __attach_device(dev, domain);
 | 
			
		||||
@ -1736,7 +1738,7 @@ static void __detach_device(struct device *dev)
 | 
			
		||||
 */
 | 
			
		||||
static void detach_device(struct device *dev)
 | 
			
		||||
{
 | 
			
		||||
	struct pci_dev *pdev = to_pci_dev(dev);
 | 
			
		||||
	struct iommu_dev_data *dev_data;
 | 
			
		||||
	unsigned long flags;
 | 
			
		||||
 | 
			
		||||
	/* lock device table */
 | 
			
		||||
@ -1744,8 +1746,12 @@ static void detach_device(struct device *dev)
 | 
			
		||||
	__detach_device(dev);
 | 
			
		||||
	write_unlock_irqrestore(&amd_iommu_devtable_lock, flags);
 | 
			
		||||
 | 
			
		||||
	if (amd_iommu_iotlb_sup && pci_ats_enabled(pdev))
 | 
			
		||||
		pci_disable_ats(pdev);
 | 
			
		||||
	dev_data = get_dev_data(dev);
 | 
			
		||||
 | 
			
		||||
	if (dev_data->ats.enabled) {
 | 
			
		||||
		pci_disable_ats(to_pci_dev(dev));
 | 
			
		||||
		dev_data->ats.enabled = false;
 | 
			
		||||
	}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
@ -1890,10 +1896,8 @@ static void update_device_table(struct protection_domain *domain)
 | 
			
		||||
{
 | 
			
		||||
	struct iommu_dev_data *dev_data;
 | 
			
		||||
 | 
			
		||||
	list_for_each_entry(dev_data, &domain->dev_list, list) {
 | 
			
		||||
		struct pci_dev *pdev = to_pci_dev(dev_data->dev);
 | 
			
		||||
		set_dte_entry(dev_data->devid, domain, pci_ats_enabled(pdev));
 | 
			
		||||
	}
 | 
			
		||||
	list_for_each_entry(dev_data, &domain->dev_list, list)
 | 
			
		||||
		set_dte_entry(dev_data->devid, domain, dev_data->ats.enabled);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static void update_domain(struct protection_domain *domain)
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user