Commit d834c5ab authored by Marek Szyprowski's avatar Marek Szyprowski Committed by Linus Torvalds

kernel/dma: remove unsupported gfp_mask parameter from dma_alloc_from_contiguous()

The CMA memory allocator doesn't support standard gfp flags for memory
allocation, so there is no point having it as a parameter for
dma_alloc_from_contiguous() function.  Replace it by a boolean no_warn
argument, which covers all the underlaying cma_alloc() function
supports.

This will help to avoid giving false feeling that this function supports
standard gfp flags and callers can pass __GFP_ZERO to get zeroed buffer,
what has already been an issue: see commit dd65a941 ("arm64:
dma-mapping: clear buffers allocated with FORCE_CONTIGUOUS flag").

Link: http://lkml.kernel.org/r/20180709122020eucas1p21a71b092975cb4a3b9954ffc63f699d1~-sqUFoa-h2939329393eucas1p2Y@eucas1p2.samsung.comSigned-off-by: default avatarMarek Szyprowski <m.szyprowski@samsung.com>
Acked-by: default avatarMichał Nazarewicz <mina86@mina86.com>
Acked-by: default avatarVlastimil Babka <vbabka@suse.cz>
Reviewed-by: default avatarChristoph Hellwig <hch@lst.de>
Cc: Laura Abbott <labbott@redhat.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Joonsoo Kim <js1304@gmail.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 65182029
...@@ -594,7 +594,7 @@ static void *__alloc_from_contiguous(struct device *dev, size_t size, ...@@ -594,7 +594,7 @@ static void *__alloc_from_contiguous(struct device *dev, size_t size,
struct page *page; struct page *page;
void *ptr = NULL; void *ptr = NULL;
page = dma_alloc_from_contiguous(dev, count, order, gfp); page = dma_alloc_from_contiguous(dev, count, order, gfp & __GFP_NOWARN);
if (!page) if (!page)
return NULL; return NULL;
...@@ -1299,7 +1299,8 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size, ...@@ -1299,7 +1299,8 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size,
unsigned long order = get_order(size); unsigned long order = get_order(size);
struct page *page; struct page *page;
page = dma_alloc_from_contiguous(dev, count, order, gfp); page = dma_alloc_from_contiguous(dev, count, order,
gfp & __GFP_NOWARN);
if (!page) if (!page)
goto error; goto error;
......
...@@ -355,7 +355,7 @@ static int __init atomic_pool_init(void) ...@@ -355,7 +355,7 @@ static int __init atomic_pool_init(void)
if (dev_get_cma_area(NULL)) if (dev_get_cma_area(NULL))
page = dma_alloc_from_contiguous(NULL, nr_pages, page = dma_alloc_from_contiguous(NULL, nr_pages,
pool_size_order, GFP_KERNEL); pool_size_order, false);
else else
page = alloc_pages(GFP_DMA32, pool_size_order); page = alloc_pages(GFP_DMA32, pool_size_order);
...@@ -573,7 +573,7 @@ static void *__iommu_alloc_attrs(struct device *dev, size_t size, ...@@ -573,7 +573,7 @@ static void *__iommu_alloc_attrs(struct device *dev, size_t size,
struct page *page; struct page *page;
page = dma_alloc_from_contiguous(dev, size >> PAGE_SHIFT, page = dma_alloc_from_contiguous(dev, size >> PAGE_SHIFT,
get_order(size), gfp); get_order(size), gfp & __GFP_NOWARN);
if (!page) if (!page)
return NULL; return NULL;
......
...@@ -137,7 +137,7 @@ static void *xtensa_dma_alloc(struct device *dev, size_t size, ...@@ -137,7 +137,7 @@ static void *xtensa_dma_alloc(struct device *dev, size_t size,
if (gfpflags_allow_blocking(flag)) if (gfpflags_allow_blocking(flag))
page = dma_alloc_from_contiguous(dev, count, get_order(size), page = dma_alloc_from_contiguous(dev, count, get_order(size),
flag); flag & __GFP_NOWARN);
if (!page) if (!page)
page = alloc_pages(flag, get_order(size)); page = alloc_pages(flag, get_order(size));
......
...@@ -2620,7 +2620,7 @@ static void *alloc_coherent(struct device *dev, size_t size, ...@@ -2620,7 +2620,7 @@ static void *alloc_coherent(struct device *dev, size_t size,
return NULL; return NULL;
page = dma_alloc_from_contiguous(dev, size >> PAGE_SHIFT, page = dma_alloc_from_contiguous(dev, size >> PAGE_SHIFT,
get_order(size), flag); get_order(size), flag & __GFP_NOWARN);
if (!page) if (!page)
return NULL; return NULL;
} }
......
...@@ -3758,7 +3758,8 @@ static void *intel_alloc_coherent(struct device *dev, size_t size, ...@@ -3758,7 +3758,8 @@ static void *intel_alloc_coherent(struct device *dev, size_t size,
if (gfpflags_allow_blocking(flags)) { if (gfpflags_allow_blocking(flags)) {
unsigned int count = size >> PAGE_SHIFT; unsigned int count = size >> PAGE_SHIFT;
page = dma_alloc_from_contiguous(dev, count, order, flags); page = dma_alloc_from_contiguous(dev, count, order,
flags & __GFP_NOWARN);
if (page && iommu_no_mapping(dev) && if (page && iommu_no_mapping(dev) &&
page_to_phys(page) + size > dev->coherent_dma_mask) { page_to_phys(page) + size > dev->coherent_dma_mask) {
dma_release_from_contiguous(dev, page, count); dma_release_from_contiguous(dev, page, count);
......
...@@ -112,7 +112,7 @@ static inline int dma_declare_contiguous(struct device *dev, phys_addr_t size, ...@@ -112,7 +112,7 @@ static inline int dma_declare_contiguous(struct device *dev, phys_addr_t size,
} }
struct page *dma_alloc_from_contiguous(struct device *dev, size_t count, struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
unsigned int order, gfp_t gfp_mask); unsigned int order, bool no_warn);
bool dma_release_from_contiguous(struct device *dev, struct page *pages, bool dma_release_from_contiguous(struct device *dev, struct page *pages,
int count); int count);
...@@ -145,7 +145,7 @@ int dma_declare_contiguous(struct device *dev, phys_addr_t size, ...@@ -145,7 +145,7 @@ int dma_declare_contiguous(struct device *dev, phys_addr_t size,
static inline static inline
struct page *dma_alloc_from_contiguous(struct device *dev, size_t count, struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
unsigned int order, gfp_t gfp_mask) unsigned int order, bool no_warn)
{ {
return NULL; return NULL;
} }
......
...@@ -178,7 +178,7 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, ...@@ -178,7 +178,7 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
* @dev: Pointer to device for which the allocation is performed. * @dev: Pointer to device for which the allocation is performed.
* @count: Requested number of pages. * @count: Requested number of pages.
* @align: Requested alignment of pages (in PAGE_SIZE order). * @align: Requested alignment of pages (in PAGE_SIZE order).
* @gfp_mask: GFP flags to use for this allocation. * @no_warn: Avoid printing message about failed allocation.
* *
* This function allocates memory buffer for specified device. It uses * This function allocates memory buffer for specified device. It uses
* device specific contiguous memory area if available or the default * device specific contiguous memory area if available or the default
...@@ -186,13 +186,12 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, ...@@ -186,13 +186,12 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
* function. * function.
*/ */
struct page *dma_alloc_from_contiguous(struct device *dev, size_t count, struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
unsigned int align, gfp_t gfp_mask) unsigned int align, bool no_warn)
{ {
if (align > CONFIG_CMA_ALIGNMENT) if (align > CONFIG_CMA_ALIGNMENT)
align = CONFIG_CMA_ALIGNMENT; align = CONFIG_CMA_ALIGNMENT;
return cma_alloc(dev_get_cma_area(dev), count, align, return cma_alloc(dev_get_cma_area(dev), count, align, no_warn);
gfp_mask & __GFP_NOWARN);
} }
/** /**
......
...@@ -78,7 +78,8 @@ void *dma_direct_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle, ...@@ -78,7 +78,8 @@ void *dma_direct_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
again: again:
/* CMA can be used only in the context which permits sleeping */ /* CMA can be used only in the context which permits sleeping */
if (gfpflags_allow_blocking(gfp)) { if (gfpflags_allow_blocking(gfp)) {
page = dma_alloc_from_contiguous(dev, count, page_order, gfp); page = dma_alloc_from_contiguous(dev, count, page_order,
gfp & __GFP_NOWARN);
if (page && !dma_coherent_ok(dev, page_to_phys(page), size)) { if (page && !dma_coherent_ok(dev, page_to_phys(page), size)) {
dma_release_from_contiguous(dev, page, count); dma_release_from_contiguous(dev, page, count);
page = NULL; page = NULL;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment