git://git.onelab.eu
/
linux-2.6.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
vserver 1.9.3
[linux-2.6.git]
/
arch
/
x86_64
/
kernel
/
pci-nommu.c
diff --git
a/arch/x86_64/kernel/pci-nommu.c
b/arch/x86_64/kernel/pci-nommu.c
index
00fb264
..
67d90b8
100644
(file)
--- a/
arch/x86_64/kernel/pci-nommu.c
+++ b/
arch/x86_64/kernel/pci-nommu.c
@@
-1,8
+1,11
@@
+/* Fallback functions when the main IOMMU code is not compiled in. This
+ code is roughly equivalent to i386. */
#include <linux/mm.h>
#include <linux/init.h>
#include <linux/pci.h>
#include <linux/string.h>
#include <asm/proto.h>
#include <linux/mm.h>
#include <linux/init.h>
#include <linux/pci.h>
#include <linux/string.h>
#include <asm/proto.h>
+#include <asm/processor.h>
int iommu_merge = 0;
EXPORT_SYMBOL(iommu_merge);
int iommu_merge = 0;
EXPORT_SYMBOL(iommu_merge);
@@
-10,57
+13,80
@@
EXPORT_SYMBOL(iommu_merge);
dma_addr_t bad_dma_address;
EXPORT_SYMBOL(bad_dma_address);
dma_addr_t bad_dma_address;
EXPORT_SYMBOL(bad_dma_address);
+int iommu_bio_merge = 0;
+EXPORT_SYMBOL(iommu_bio_merge);
+
+int iommu_sac_force = 0;
+EXPORT_SYMBOL(iommu_sac_force);
+
/*
* Dummy IO MMU functions
*/
/*
* Dummy IO MMU functions
*/
-void *
pci_alloc_consistent(struct pci_dev
*hwdev, size_t size,
-
dma_addr_t *dma_handle
)
+void *
dma_alloc_coherent(struct device
*hwdev, size_t size,
+
dma_addr_t *dma_handle, unsigned gfp
)
{
void *ret;
{
void *ret;
- int gfp = GFP_ATOMIC;
-
- if (hwdev == NULL ||
- end_pfn > (hwdev->dma_mask>>PAGE_SHIFT) || /* XXX */
- (u32)hwdev->dma_mask < 0xffffffff)
- gfp |= GFP_DMA;
- ret = (void *)__get_free_pages(gfp, get_order(size));
+ u64 mask;
+ int order = get_order(size);
- if (ret != NULL) {
- memset(ret, 0, size);
+ if (hwdev)
+ mask = hwdev->coherent_dma_mask & *hwdev->dma_mask;
+ else
+ mask = 0xffffffff;
+ for (;;) {
+ ret = (void *)__get_free_pages(gfp, order);
+ if (ret == NULL)
+ return NULL;
*dma_handle = virt_to_bus(ret);
*dma_handle = virt_to_bus(ret);
+ if ((*dma_handle & ~mask) == 0)
+ break;
+ free_pages((unsigned long)ret, order);
+ if (gfp & GFP_DMA)
+ return NULL;
+ gfp |= GFP_DMA;
}
}
+
+ memset(ret, 0, size);
return ret;
}
return ret;
}
+EXPORT_SYMBOL(dma_alloc_coherent);
-void
pci_free_consistent(struct pci_dev
*hwdev, size_t size,
+void
dma_free_coherent(struct device
*hwdev, size_t size,
void *vaddr, dma_addr_t dma_handle)
{
free_pages((unsigned long)vaddr, get_order(size));
}
void *vaddr, dma_addr_t dma_handle)
{
free_pages((unsigned long)vaddr, get_order(size));
}
+EXPORT_SYMBOL(dma_free_coherent);
-int
pci_dma_supported(struct pci_dev
*hwdev, u64 mask)
+int
dma_supported(struct device
*hwdev, u64 mask)
{
/*
* we fall back to GFP_DMA when the mask isn't all 1s,
* so we can't guarantee allocations that must be
* within a tighter range than GFP_DMA..
{
/*
* we fall back to GFP_DMA when the mask isn't all 1s,
* so we can't guarantee allocations that must be
* within a tighter range than GFP_DMA..
-
* RED-PEN this won't work for pci_map_single. Caller has to
-
* use GFP_DMA in the first place.
+ * RED-PEN this won't work for pci_map_single. Caller has to
+ * use GFP_DMA in the first place.
*/
if (mask < 0x00ffffff)
return 0;
return 1;
}
*/
if (mask < 0x00ffffff)
return 0;
return 1;
}
+EXPORT_SYMBOL(dma_supported);
-EXPORT_SYMBOL(pci_dma_supported);
+int dma_get_cache_alignment(void)
+{
+ return boot_cpu_data.x86_clflush_size;
+}
+EXPORT_SYMBOL(dma_get_cache_alignment);
static int __init check_ram(void)
{
if (end_pfn >= 0xffffffff>>PAGE_SHIFT) {
static int __init check_ram(void)
{
if (end_pfn >= 0xffffffff>>PAGE_SHIFT) {
- printk(KERN_ERR "WARNING more than 4GB of memory but no IOMMU.\n"
- KERN_ERR "WARNING 32bit PCI may malfunction.\n");
+ printk(
+ KERN_ERR "WARNING more than 4GB of memory but IOMMU not compiled in.\n"
+ KERN_ERR "WARNING 32bit PCI may malfunction.\n");
}
return 0;
}
}
return 0;
}