Commit 94f6d190 authored by Joerg Roedel's avatar Joerg Roedel

x86/amd-iommu: Simplify get_device_resources()

With the previous changes the get_device_resources function
can be simplified even more. The only important information
for the callers is the protection domain.
This patch renames the function to get_domain() and let it
only return the protection domain for a device.
Signed-off-by: default avatarJoerg Roedel <joerg.roedel@amd.com>
parent 15898bbc
...@@ -1463,6 +1463,7 @@ static struct dma_ops_domain *find_protection_domain(u16 devid) ...@@ -1463,6 +1463,7 @@ static struct dma_ops_domain *find_protection_domain(u16 devid)
{ {
struct dma_ops_domain *entry, *ret = NULL; struct dma_ops_domain *entry, *ret = NULL;
unsigned long flags; unsigned long flags;
u16 alias = amd_iommu_alias_table[devid];
if (list_empty(&iommu_pd_list)) if (list_empty(&iommu_pd_list))
return NULL; return NULL;
...@@ -1470,7 +1471,8 @@ static struct dma_ops_domain *find_protection_domain(u16 devid) ...@@ -1470,7 +1471,8 @@ static struct dma_ops_domain *find_protection_domain(u16 devid)
spin_lock_irqsave(&iommu_pd_list_lock, flags); spin_lock_irqsave(&iommu_pd_list_lock, flags);
list_for_each_entry(entry, &iommu_pd_list, list) { list_for_each_entry(entry, &iommu_pd_list, list) {
if (entry->target_dev == devid) { if (entry->target_dev == devid ||
entry->target_dev == alias) {
ret = entry; ret = entry;
break; break;
} }
...@@ -1488,33 +1490,31 @@ static struct dma_ops_domain *find_protection_domain(u16 devid) ...@@ -1488,33 +1490,31 @@ static struct dma_ops_domain *find_protection_domain(u16 devid)
* If the device is not yet associated with a domain this is also done * If the device is not yet associated with a domain this is also done
* in this function. * in this function.
*/ */
static bool get_device_resources(struct device *dev, static struct protection_domain *get_domain(struct device *dev)
struct protection_domain **domain,
u16 *bdf)
{ {
struct protection_domain *domain;
struct dma_ops_domain *dma_dom; struct dma_ops_domain *dma_dom;
struct amd_iommu *iommu; u16 devid = get_device_id(dev);
if (!check_device(dev)) if (!check_device(dev))
return false; return ERR_PTR(-EINVAL);
*bdf = get_device_id(dev); domain = domain_for_device(dev);
*domain = domain_for_device(dev); if (domain != NULL && !dma_ops_domain(domain))
iommu = amd_iommu_rlookup_table[*bdf]; return ERR_PTR(-EBUSY);
if (*domain != NULL) if (domain != NULL)
return true; return domain;
/* Device not bount yet - bind it */ /* Device not bount yet - bind it */
dma_dom = find_protection_domain(*bdf); dma_dom = find_protection_domain(devid);
if (!dma_dom) if (!dma_dom)
dma_dom = iommu->default_dom; dma_dom = amd_iommu_rlookup_table[devid]->default_dom;
*domain = &dma_dom->domain; attach_device(dev, &dma_dom->domain);
attach_device(dev, *domain);
DUMP_printk("Using protection domain %d for device %s\n", DUMP_printk("Using protection domain %d for device %s\n",
(*domain)->id, dev_name(dev)); dma_dom->domain.id, dev_name(dev));
return true; return &dma_dom->domain;
} }
static void update_device_table(struct protection_domain *domain) static void update_device_table(struct protection_domain *domain)
...@@ -1825,23 +1825,22 @@ static dma_addr_t map_page(struct device *dev, struct page *page, ...@@ -1825,23 +1825,22 @@ static dma_addr_t map_page(struct device *dev, struct page *page,
{ {
unsigned long flags; unsigned long flags;
struct protection_domain *domain; struct protection_domain *domain;
u16 devid;
dma_addr_t addr; dma_addr_t addr;
u64 dma_mask; u64 dma_mask;
phys_addr_t paddr = page_to_phys(page) + offset; phys_addr_t paddr = page_to_phys(page) + offset;
INC_STATS_COUNTER(cnt_map_single); INC_STATS_COUNTER(cnt_map_single);
if (!get_device_resources(dev, &domain, &devid)) domain = get_domain(dev);
/* device not handled by any AMD IOMMU */ if (PTR_ERR(domain) == -EINVAL)
return (dma_addr_t)paddr; return (dma_addr_t)paddr;
else if (IS_ERR(domain))
return DMA_ERROR_CODE;
dma_mask = *dev->dma_mask; dma_mask = *dev->dma_mask;
if (!dma_ops_domain(domain))
return DMA_ERROR_CODE;
spin_lock_irqsave(&domain->lock, flags); spin_lock_irqsave(&domain->lock, flags);
addr = __map_single(dev, domain->priv, paddr, size, dir, false, addr = __map_single(dev, domain->priv, paddr, size, dir, false,
dma_mask); dma_mask);
if (addr == DMA_ERROR_CODE) if (addr == DMA_ERROR_CODE)
...@@ -1863,15 +1862,11 @@ static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size, ...@@ -1863,15 +1862,11 @@ static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size,
{ {
unsigned long flags; unsigned long flags;
struct protection_domain *domain; struct protection_domain *domain;
u16 devid;
INC_STATS_COUNTER(cnt_unmap_single); INC_STATS_COUNTER(cnt_unmap_single);
if (!get_device_resources(dev, &domain, &devid)) domain = get_domain(dev);
/* device not handled by any AMD IOMMU */ if (IS_ERR(domain))
return;
if (!dma_ops_domain(domain))
return; return;
spin_lock_irqsave(&domain->lock, flags); spin_lock_irqsave(&domain->lock, flags);
...@@ -1911,7 +1906,6 @@ static int map_sg(struct device *dev, struct scatterlist *sglist, ...@@ -1911,7 +1906,6 @@ static int map_sg(struct device *dev, struct scatterlist *sglist,
{ {
unsigned long flags; unsigned long flags;
struct protection_domain *domain; struct protection_domain *domain;
u16 devid;
int i; int i;
struct scatterlist *s; struct scatterlist *s;
phys_addr_t paddr; phys_addr_t paddr;
...@@ -1920,14 +1914,14 @@ static int map_sg(struct device *dev, struct scatterlist *sglist, ...@@ -1920,14 +1914,14 @@ static int map_sg(struct device *dev, struct scatterlist *sglist,
INC_STATS_COUNTER(cnt_map_sg); INC_STATS_COUNTER(cnt_map_sg);
if (!get_device_resources(dev, &domain, &devid)) domain = get_domain(dev);
if (PTR_ERR(domain) == -EINVAL)
return map_sg_no_iommu(dev, sglist, nelems, dir); return map_sg_no_iommu(dev, sglist, nelems, dir);
else if (IS_ERR(domain))
return 0;
dma_mask = *dev->dma_mask; dma_mask = *dev->dma_mask;
if (!dma_ops_domain(domain))
return 0;
spin_lock_irqsave(&domain->lock, flags); spin_lock_irqsave(&domain->lock, flags);
for_each_sg(sglist, s, nelems, i) { for_each_sg(sglist, s, nelems, i) {
...@@ -1974,15 +1968,12 @@ static void unmap_sg(struct device *dev, struct scatterlist *sglist, ...@@ -1974,15 +1968,12 @@ static void unmap_sg(struct device *dev, struct scatterlist *sglist,
unsigned long flags; unsigned long flags;
struct protection_domain *domain; struct protection_domain *domain;
struct scatterlist *s; struct scatterlist *s;
u16 devid;
int i; int i;
INC_STATS_COUNTER(cnt_unmap_sg); INC_STATS_COUNTER(cnt_unmap_sg);
if (!get_device_resources(dev, &domain, &devid)) domain = get_domain(dev);
return; if (IS_ERR(domain))
if (!dma_ops_domain(domain))
return; return;
spin_lock_irqsave(&domain->lock, flags); spin_lock_irqsave(&domain->lock, flags);
...@@ -2007,17 +1998,18 @@ static void *alloc_coherent(struct device *dev, size_t size, ...@@ -2007,17 +1998,18 @@ static void *alloc_coherent(struct device *dev, size_t size,
unsigned long flags; unsigned long flags;
void *virt_addr; void *virt_addr;
struct protection_domain *domain; struct protection_domain *domain;
u16 devid;
phys_addr_t paddr; phys_addr_t paddr;
u64 dma_mask = dev->coherent_dma_mask; u64 dma_mask = dev->coherent_dma_mask;
INC_STATS_COUNTER(cnt_alloc_coherent); INC_STATS_COUNTER(cnt_alloc_coherent);
if (!get_device_resources(dev, &domain, &devid)) { domain = get_domain(dev);
if (PTR_ERR(domain) == -EINVAL) {
virt_addr = (void *)__get_free_pages(flag, get_order(size)); virt_addr = (void *)__get_free_pages(flag, get_order(size));
*dma_addr = __pa(virt_addr); *dma_addr = __pa(virt_addr);
return virt_addr; return virt_addr;
} } else if (IS_ERR(domain))
return NULL;
dma_mask = dev->coherent_dma_mask; dma_mask = dev->coherent_dma_mask;
flag &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32); flag &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32);
...@@ -2029,9 +2021,6 @@ static void *alloc_coherent(struct device *dev, size_t size, ...@@ -2029,9 +2021,6 @@ static void *alloc_coherent(struct device *dev, size_t size,
paddr = virt_to_phys(virt_addr); paddr = virt_to_phys(virt_addr);
if (!dma_ops_domain(domain))
goto out_free;
if (!dma_mask) if (!dma_mask)
dma_mask = *dev->dma_mask; dma_mask = *dev->dma_mask;
...@@ -2066,14 +2055,11 @@ static void free_coherent(struct device *dev, size_t size, ...@@ -2066,14 +2055,11 @@ static void free_coherent(struct device *dev, size_t size,
{ {
unsigned long flags; unsigned long flags;
struct protection_domain *domain; struct protection_domain *domain;
u16 devid;
INC_STATS_COUNTER(cnt_free_coherent); INC_STATS_COUNTER(cnt_free_coherent);
if (!get_device_resources(dev, &domain, &devid)) domain = get_domain(dev);
goto free_mem; if (IS_ERR(domain))
if (!dma_ops_domain(domain))
goto free_mem; goto free_mem;
spin_lock_irqsave(&domain->lock, flags); spin_lock_irqsave(&domain->lock, flags);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment