Commit 6829e274 authored by Marek Szyprowski's avatar Marek Szyprowski Committed by Will Deacon

arm64: dma-mapping: always clear allocated buffers

Buffers allocated by dma_alloc_coherent() are always zeroed on Alpha,
ARM (32bit), MIPS, PowerPC, x86/x86_64 and probably other architectures.
It turned out that some drivers rely on this 'feature'. Allocated buffer
might be also exposed to userspace with dma_mmap() call, so clearing it
is desired from security point of view to avoid exposing random memory
to userspace. This patch unifies dma_alloc_coherent() behavior on ARM64
architecture with other implementations by unconditionally zeroing
allocated buffer.

Cc: <stable@vger.kernel.org> # v3.14+
Signed-off-by: default avatarMarek Szyprowski <m.szyprowski@samsung.com>
Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
parent 6544e67b
...@@ -67,7 +67,6 @@ static void *__alloc_from_pool(size_t size, struct page **ret_page, gfp_t flags) ...@@ -67,7 +67,6 @@ static void *__alloc_from_pool(size_t size, struct page **ret_page, gfp_t flags)
*ret_page = phys_to_page(phys); *ret_page = phys_to_page(phys);
ptr = (void *)val; ptr = (void *)val;
if (flags & __GFP_ZERO)
memset(ptr, 0, size); memset(ptr, 0, size);
} }
...@@ -113,7 +112,6 @@ static void *__dma_alloc_coherent(struct device *dev, size_t size, ...@@ -113,7 +112,6 @@ static void *__dma_alloc_coherent(struct device *dev, size_t size,
*dma_handle = phys_to_dma(dev, page_to_phys(page)); *dma_handle = phys_to_dma(dev, page_to_phys(page));
addr = page_address(page); addr = page_address(page);
if (flags & __GFP_ZERO)
memset(addr, 0, size); memset(addr, 0, size);
return addr; return addr;
} else { } else {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment