Commit 02f7a189 authored by FUJITA Tomonori's avatar FUJITA Tomonori Committed by Ingo Molnar

sparc: Use asm-generic/dma-mapping-common.h

Signed-off-by: default avatarFUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Tested-by: default avatarRobert Reif <reif@earthlink.net>
Acked-by: default avatarDavid S. Miller <davem@davemloft.net>
Cc: tony.luck@intel.com
Cc: fenghua.yu@intel.com
LKML-Reference: <1249872797-1314-5-git-send-email-fujita.tomonori@lab.ntt.co.jp>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent bc0a14f1
...@@ -25,6 +25,7 @@ config SPARC ...@@ -25,6 +25,7 @@ config SPARC
select ARCH_WANT_OPTIONAL_GPIOLIB select ARCH_WANT_OPTIONAL_GPIOLIB
select RTC_CLASS select RTC_CLASS
select RTC_DRV_M48T59 select RTC_DRV_M48T59
select HAVE_DMA_ATTRS
config SPARC32 config SPARC32
def_bool !64BIT def_bool !64BIT
......
...@@ -3,6 +3,7 @@ ...@@ -3,6 +3,7 @@
#include <linux/scatterlist.h> #include <linux/scatterlist.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/dma-debug.h>
#define DMA_ERROR_CODE (~(dma_addr_t)0x0) #define DMA_ERROR_CODE (~(dma_addr_t)0x0)
...@@ -13,113 +14,31 @@ extern int dma_set_mask(struct device *dev, u64 dma_mask); ...@@ -13,113 +14,31 @@ extern int dma_set_mask(struct device *dev, u64 dma_mask);
#define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h)
#define dma_is_consistent(d, h) (1) #define dma_is_consistent(d, h) (1)
extern const struct dma_map_ops *dma_ops; extern struct dma_map_ops *dma_ops;
static inline void *dma_alloc_coherent(struct device *dev, size_t size, static inline struct dma_map_ops *get_dma_ops(struct device *dev)
dma_addr_t *dma_handle, gfp_t flag)
{
return dma_ops->alloc_coherent(dev, size, dma_handle, flag);
}
static inline void dma_free_coherent(struct device *dev, size_t size,
void *cpu_addr, dma_addr_t dma_handle)
{
dma_ops->free_coherent(dev, size, cpu_addr, dma_handle);
}
static inline dma_addr_t dma_map_single(struct device *dev, void *cpu_addr,
size_t size,
enum dma_data_direction direction)
{ {
return dma_ops->map_page(dev, virt_to_page(cpu_addr), return dma_ops;
(unsigned long)cpu_addr & ~PAGE_MASK, size,
direction, NULL);
} }
static inline void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, #include <asm-generic/dma-mapping-common.h>
size_t size,
enum dma_data_direction direction)
{
dma_ops->unmap_page(dev, dma_addr, size, direction, NULL);
}
static inline dma_addr_t dma_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t size,
enum dma_data_direction direction)
{
return dma_ops->map_page(dev, page, offset, size, direction, NULL);
}
static inline void dma_unmap_page(struct device *dev, dma_addr_t dma_address, static inline void *dma_alloc_coherent(struct device *dev, size_t size,
size_t size, dma_addr_t *dma_handle, gfp_t flag)
enum dma_data_direction direction)
{
dma_ops->unmap_page(dev, dma_address, size, direction, NULL);
}
static inline int dma_map_sg(struct device *dev, struct scatterlist *sg,
int nents, enum dma_data_direction direction)
{
return dma_ops->map_sg(dev, sg, nents, direction, NULL);
}
static inline void dma_unmap_sg(struct device *dev, struct scatterlist *sg,
int nents, enum dma_data_direction direction)
{
dma_ops->unmap_sg(dev, sg, nents, direction, NULL);
}
static inline void dma_sync_single_for_cpu(struct device *dev,
dma_addr_t dma_handle, size_t size,
enum dma_data_direction direction)
{
dma_ops->sync_single_for_cpu(dev, dma_handle, size, direction);
}
static inline void dma_sync_single_for_device(struct device *dev,
dma_addr_t dma_handle,
size_t size,
enum dma_data_direction direction)
{
if (dma_ops->sync_single_for_device)
dma_ops->sync_single_for_device(dev, dma_handle, size,
direction);
}
static inline void dma_sync_sg_for_cpu(struct device *dev,
struct scatterlist *sg, int nelems,
enum dma_data_direction direction)
{ {
dma_ops->sync_sg_for_cpu(dev, sg, nelems, direction); struct dma_map_ops *ops = get_dma_ops(dev);
}
static inline void dma_sync_sg_for_device(struct device *dev, return ops->alloc_coherent(dev, size, dma_handle, flag);
struct scatterlist *sg, int nelems,
enum dma_data_direction direction)
{
if (dma_ops->sync_sg_for_device)
dma_ops->sync_sg_for_device(dev, sg, nelems, direction);
} }
static inline void dma_sync_single_range_for_cpu(struct device *dev, static inline void dma_free_coherent(struct device *dev, size_t size,
dma_addr_t dma_handle, void *cpu_addr, dma_addr_t dma_handle)
unsigned long offset,
size_t size,
enum dma_data_direction dir)
{ {
dma_sync_single_for_cpu(dev, dma_handle+offset, size, dir); struct dma_map_ops *ops = get_dma_ops(dev);
}
static inline void dma_sync_single_range_for_device(struct device *dev, ops->free_coherent(dev, size, cpu_addr, dma_handle);
dma_addr_t dma_handle,
unsigned long offset,
size_t size,
enum dma_data_direction dir)
{
dma_sync_single_for_device(dev, dma_handle+offset, size, dir);
} }
static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
{ {
return (dma_addr == DMA_ERROR_CODE); return (dma_addr == DMA_ERROR_CODE);
......
...@@ -165,7 +165,7 @@ static void dma32_sync_sg_for_device(struct device *dev, ...@@ -165,7 +165,7 @@ static void dma32_sync_sg_for_device(struct device *dev,
BUG(); BUG();
} }
static const struct dma_map_ops dma32_dma_ops = { static struct dma_map_ops dma32_dma_ops = {
.alloc_coherent = dma32_alloc_coherent, .alloc_coherent = dma32_alloc_coherent,
.free_coherent = dma32_free_coherent, .free_coherent = dma32_free_coherent,
.map_page = dma32_map_page, .map_page = dma32_map_page,
...@@ -178,5 +178,5 @@ static const struct dma_map_ops dma32_dma_ops = { ...@@ -178,5 +178,5 @@ static const struct dma_map_ops dma32_dma_ops = {
.sync_sg_for_device = dma32_sync_sg_for_device, .sync_sg_for_device = dma32_sync_sg_for_device,
}; };
const struct dma_map_ops *dma_ops = &dma32_dma_ops; struct dma_map_ops *dma_ops = &dma32_dma_ops;
EXPORT_SYMBOL(dma_ops); EXPORT_SYMBOL(dma_ops);
...@@ -826,7 +826,7 @@ static void dma_4u_sync_sg_for_cpu(struct device *dev, ...@@ -826,7 +826,7 @@ static void dma_4u_sync_sg_for_cpu(struct device *dev,
spin_unlock_irqrestore(&iommu->lock, flags); spin_unlock_irqrestore(&iommu->lock, flags);
} }
static const struct dma_map_ops sun4u_dma_ops = { static struct dma_map_ops sun4u_dma_ops = {
.alloc_coherent = dma_4u_alloc_coherent, .alloc_coherent = dma_4u_alloc_coherent,
.free_coherent = dma_4u_free_coherent, .free_coherent = dma_4u_free_coherent,
.map_page = dma_4u_map_page, .map_page = dma_4u_map_page,
...@@ -837,7 +837,7 @@ static const struct dma_map_ops sun4u_dma_ops = { ...@@ -837,7 +837,7 @@ static const struct dma_map_ops sun4u_dma_ops = {
.sync_sg_for_cpu = dma_4u_sync_sg_for_cpu, .sync_sg_for_cpu = dma_4u_sync_sg_for_cpu,
}; };
const struct dma_map_ops *dma_ops = &sun4u_dma_ops; struct dma_map_ops *dma_ops = &sun4u_dma_ops;
EXPORT_SYMBOL(dma_ops); EXPORT_SYMBOL(dma_ops);
int dma_supported(struct device *dev, u64 device_mask) int dma_supported(struct device *dev, u64 device_mask)
......
...@@ -539,7 +539,7 @@ static void dma_4v_sync_sg_for_cpu(struct device *dev, ...@@ -539,7 +539,7 @@ static void dma_4v_sync_sg_for_cpu(struct device *dev,
/* Nothing to do... */ /* Nothing to do... */
} }
static const struct dma_map_ops sun4v_dma_ops = { static struct dma_map_ops sun4v_dma_ops = {
.alloc_coherent = dma_4v_alloc_coherent, .alloc_coherent = dma_4v_alloc_coherent,
.free_coherent = dma_4v_free_coherent, .free_coherent = dma_4v_free_coherent,
.map_page = dma_4v_map_page, .map_page = dma_4v_map_page,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment