}
}
- dma_addr = virt_to_bus(ptr);
+ dma_addr = virt_to_dma(dev, ptr);
if (device_info && dma_needs_bounce(dev, dma_addr, size)) {
struct safe_buffer *buf;
dev_dbg(dev,
"%s: unsafe buffer %p (phy=%p) mapped to %p (phy=%p)\n",
- __func__, buf->ptr, (void *) virt_to_bus(buf->ptr),
+ __func__, buf->ptr, (void *) virt_to_dma(dev, buf->ptr),
buf->safe, (void *) buf->safe_dma_addr);
if ((dir == DMA_TO_DEVICE) ||
dev_dbg(dev,
"%s: unsafe buffer %p (phy=%p) mapped to %p (phy=%p)\n",
- __func__, buf->ptr, (void *) virt_to_bus(buf->ptr),
+ __func__, buf->ptr, (void *) virt_to_dma(dev, buf->ptr),
buf->safe, (void *) buf->safe_dma_addr);
dev_dbg(dev,
"%s: unsafe buffer %p (phy=%p) mapped to %p (phy=%p)\n",
- __func__, buf->ptr, (void *) virt_to_bus(buf->ptr),
+ __func__, buf->ptr, (void *) virt_to_dma(dev, buf->ptr),
buf->safe, (void *) buf->safe_dma_addr);
DO_STATS ( device_info->bounce_count++ );
}
consistent_sync(buf->safe, size, dir);
} else {
- consistent_sync(bus_to_virt(dma_addr), size, dir);
+ consistent_sync(dma_to_virt(dev, dma_addr), size, dir);
}
}