static inline int dma_is_consistent(dma_addr_t handle)
{
- return !!arch_is_coherent();
+ return 0;
}
/*
dma_map_single(struct device *dev, void *cpu_addr, size_t size,
enum dma_data_direction dir)
{
- if (!arch_is_coherent())
- consistent_sync(cpu_addr, size, dir);
-
+ consistent_sync(cpu_addr, size, dir);
return virt_to_dma(dev, (unsigned long)cpu_addr);
}
#else
sg->dma_address = page_to_dma(dev, sg->page) + sg->offset;
virt = page_address(sg->page) + sg->offset;
-
- if (!arch_is_coherent())
- consistent_sync(virt, sg->length, dir);
+ consistent_sync(virt, sg->length, dir);
}
return nents;
dma_sync_single_for_cpu(struct device *dev, dma_addr_t handle, size_t size,
enum dma_data_direction dir)
{
- if (!arch_is_coherent())
- consistent_sync((void *)dma_to_virt(dev, handle), size, dir);
+ consistent_sync((void *)dma_to_virt(dev, handle), size, dir);
}
static inline void
dma_sync_single_for_device(struct device *dev, dma_addr_t handle, size_t size,
enum dma_data_direction dir)
{
- if (!arch_is_coherent())
- consistent_sync((void *)dma_to_virt(dev, handle), size, dir);
+ consistent_sync((void *)dma_to_virt(dev, handle), size, dir);
}
#else
extern void dma_sync_single_for_cpu(struct device*, dma_addr_t, size_t, enum dma_data_direction);
for (i = 0; i < nents; i++, sg++) {
char *virt = page_address(sg->page) + sg->offset;
- if (!arch_is_coherent())
- consistent_sync(virt, sg->length, dir);
+ consistent_sync(virt, sg->length, dir);
}
}
for (i = 0; i < nents; i++, sg++) {
char *virt = page_address(sg->page) + sg->offset;
- if (!arch_is_coherent())
- consistent_sync(virt, sg->length, dir);
+ consistent_sync(virt, sg->length, dir);
}
}
#else