This commit was manufactured by cvs2svn to create tag
[linux-2.6.git] / drivers / ide / ide-dma.c
index aeaed74..0fa3f0b 100644 (file)
@@ -85,7 +85,6 @@
 #include <linux/init.h>
 #include <linux/ide.h>
 #include <linux/delay.h>
-#include <linux/scatterlist.h>
 
 #include <asm/io.h>
 #include <asm/irq.h>
@@ -131,9 +130,11 @@ static const struct drive_list_entry drive_blacklist [] = {
        { "CD-ROM Drive/F5A",   "ALL"           },
        { "WPI CDD-820",                "ALL"           },
        { "SAMSUNG CD-ROM SC-148C",     "ALL"           },
+       { "SAMSUNG CD-ROM SC-148F",     "ALL"           },
        { "SAMSUNG CD-ROM SC",  "ALL"           },
        { "SanDisk SDP3B-64"    ,       "ALL"           },
        { "SAMSUNG CD-ROM SN-124",      "ALL"           },
+       { "PLEXTOR CD-R PX-W8432T",     "ALL"           },
        { "ATAPI CD-ROM DRIVE 40X MAXIMUM",     "ALL"           },
        { "_NEC DV5800A",               "ALL"           },  
        { NULL                  ,       NULL            }
@@ -204,22 +205,72 @@ int ide_build_sglist(ide_drive_t *drive, struct request *rq)
 {
        ide_hwif_t *hwif = HWIF(drive);
        struct scatterlist *sg = hwif->sg_table;
+       int nents;
 
-       if ((rq->flags & REQ_DRIVE_TASKFILE) && rq->nr_sectors > 256)
-               BUG();
-
-       ide_map_sg(drive, rq);
-
+       nents = blk_rq_map_sg(drive->queue, rq, hwif->sg_table);
+               
        if (rq_data_dir(rq) == READ)
                hwif->sg_dma_direction = PCI_DMA_FROMDEVICE;
        else
                hwif->sg_dma_direction = PCI_DMA_TODEVICE;
 
-       return pci_map_sg(hwif->pci_dev, sg, hwif->sg_nents, hwif->sg_dma_direction);
+       return pci_map_sg(hwif->pci_dev, sg, nents, hwif->sg_dma_direction);
 }
 
 EXPORT_SYMBOL_GPL(ide_build_sglist);
 
+/**
+ *     ide_raw_build_sglist    -       map IDE scatter gather for DMA
+ *     @drive: the drive to build the DMA table for
+ *     @rq: the request holding the sg list
+ *
+ *     Perform the PCI mapping magic necessary to access the source or
+ *     target buffers of a taskfile request via PCI DMA. The lower layers 
+ *     of the  kernel provide the necessary cache management so that we can
+ *     operate in a portable fashion
+ */
+
+int ide_raw_build_sglist(ide_drive_t *drive, struct request *rq)
+{
+       ide_hwif_t *hwif = HWIF(drive);
+       struct scatterlist *sg = hwif->sg_table;
+       int nents = 0;
+       ide_task_t *args = rq->special;
+       u8 *virt_addr = rq->buffer;
+       int sector_count = rq->nr_sectors;
+
+       if (args->command_type == IDE_DRIVE_TASK_RAW_WRITE)
+               hwif->sg_dma_direction = PCI_DMA_TODEVICE;
+       else
+               hwif->sg_dma_direction = PCI_DMA_FROMDEVICE;
+
+#if 1
+       if (sector_count > 256)
+               BUG();
+
+       if (sector_count > 128) {
+#else
+       while (sector_count > 128) {
+#endif
+               memset(&sg[nents], 0, sizeof(*sg));
+               sg[nents].page = virt_to_page(virt_addr);
+               sg[nents].offset = offset_in_page(virt_addr);
+               sg[nents].length = 128  * SECTOR_SIZE;
+               nents++;
+               virt_addr = virt_addr + (128 * SECTOR_SIZE);
+               sector_count -= 128;
+       }
+       memset(&sg[nents], 0, sizeof(*sg));
+       sg[nents].page = virt_to_page(virt_addr);
+       sg[nents].offset = offset_in_page(virt_addr);
+       sg[nents].length =  sector_count  * SECTOR_SIZE;
+       nents++;
+
+       return pci_map_sg(hwif->pci_dev, sg, nents, hwif->sg_dma_direction);
+}
+
+EXPORT_SYMBOL_GPL(ide_raw_build_sglist);
+
 /**
  *     ide_build_dmatable      -       build IDE DMA table
  *
@@ -241,7 +292,10 @@ int ide_build_dmatable (ide_drive_t *drive, struct request *rq)
        int i;
        struct scatterlist *sg;
 
-       hwif->sg_nents = i = ide_build_sglist(drive, rq);
+       if (HWGROUP(drive)->rq->flags & REQ_DRIVE_TASKFILE)
+               hwif->sg_nents = i = ide_raw_build_sglist(drive, rq);
+       else
+               hwif->sg_nents = i = ide_build_sglist(drive, rq);
 
        if (!i)
                return 0;
@@ -535,8 +589,10 @@ int __ide_dma_check (ide_drive_t *drive)
 EXPORT_SYMBOL(__ide_dma_check);
 
 /**
- *     ide_dma_setup   -       begin a DMA phase
+ *     ide_start_dma   -       begin a DMA phase
+ *     @hwif: interface
  *     @drive: target device
+ *     @reading: set if reading, clear if writing
  *
  *     Build an IDE DMA PRD (IDE speak for scatter gather table)
  *     and then set up the DMA transfer registers for a device
@@ -546,24 +602,15 @@ EXPORT_SYMBOL(__ide_dma_check);
  *     Returns 0 on success. If a PIO fallback is required then 1
  *     is returned. 
  */
-
-int ide_dma_setup(ide_drive_t *drive)
+int ide_start_dma(ide_hwif_t *hwif, ide_drive_t *drive, int reading)
 {
-       ide_hwif_t *hwif = drive->hwif;
        struct request *rq = HWGROUP(drive)->rq;
-       unsigned int reading;
        u8 dma_stat;
 
-       if (rq_data_dir(rq))
-               reading = 0;
-       else
-               reading = 1 << 3;
-
        /* fall back to pio! */
-       if (!ide_build_dmatable(drive, rq)) {
-               ide_map_sg(drive, rq);
+       if (!ide_build_dmatable(drive, rq))
                return 1;
-       }
 
        /* PRD table */
        hwif->OUTL(hwif->dmatable_dma, hwif->dma_prdtable);
@@ -580,15 +627,73 @@ int ide_dma_setup(ide_drive_t *drive)
        return 0;
 }
 
-EXPORT_SYMBOL_GPL(ide_dma_setup);
+EXPORT_SYMBOL(ide_start_dma);
+
+int __ide_dma_read (ide_drive_t *drive /*, struct request *rq */)
+{
+       ide_hwif_t *hwif        = HWIF(drive);
+       struct request *rq      = HWGROUP(drive)->rq;
+       unsigned int reading    = 1 << 3;
+       u8 lba48                = (drive->addressing == 1) ? 1 : 0;
+       task_ioreg_t command    = WIN_NOP;
+
+       /* try pio */
+       if (ide_start_dma(hwif, drive, reading))
+               return 1;
+
+       if (drive->media != ide_disk)
+               return 0;
+
+       command = (lba48) ? WIN_READDMA_EXT : WIN_READDMA;
+       
+       if (drive->vdma)
+               command = (lba48) ? WIN_READ_EXT: WIN_READ;
+               
+       if (rq->flags & REQ_DRIVE_TASKFILE) {
+               ide_task_t *args = rq->special;
+               command = args->tfRegister[IDE_COMMAND_OFFSET];
+       }
+
+       /* issue cmd to drive */
+       ide_execute_command(drive, command, &ide_dma_intr, 2*WAIT_CMD, dma_timer_expiry);
+       return hwif->ide_dma_begin(drive);
+}
+
+EXPORT_SYMBOL(__ide_dma_read);
 
-static void ide_dma_exec_cmd(ide_drive_t *drive, u8 command)
+int __ide_dma_write (ide_drive_t *drive /*, struct request *rq */)
 {
+       ide_hwif_t *hwif        = HWIF(drive);
+       struct request *rq      = HWGROUP(drive)->rq;
+       unsigned int reading    = 0;
+       u8 lba48                = (drive->addressing == 1) ? 1 : 0;
+       task_ioreg_t command    = WIN_NOP;
+
+       /* try PIO instead of DMA */
+       if (ide_start_dma(hwif, drive, reading))
+               return 1;
+
+       if (drive->media != ide_disk)
+               return 0;
+
+       command = (lba48) ? WIN_WRITEDMA_EXT : WIN_WRITEDMA;
+       if (drive->vdma)
+               command = (lba48) ? WIN_WRITE_EXT: WIN_WRITE;
+               
+       if (rq->flags & REQ_DRIVE_TASKFILE) {
+               ide_task_t *args = rq->special;
+               command = args->tfRegister[IDE_COMMAND_OFFSET];
+       }
+
        /* issue cmd to drive */
        ide_execute_command(drive, command, &ide_dma_intr, 2*WAIT_CMD, dma_timer_expiry);
+
+       return hwif->ide_dma_begin(drive);
 }
 
-void ide_dma_start(ide_drive_t *drive)
+EXPORT_SYMBOL(__ide_dma_write);
+
+int __ide_dma_begin (ide_drive_t *drive)
 {
        ide_hwif_t *hwif        = HWIF(drive);
        u8 dma_cmd              = hwif->INB(hwif->dma_command);
@@ -602,9 +707,10 @@ void ide_dma_start(ide_drive_t *drive)
        hwif->OUTB(dma_cmd|1, hwif->dma_command);
        hwif->dma = 1;
        wmb();
+       return 0;
 }
 
-EXPORT_SYMBOL_GPL(ide_dma_start);
+EXPORT_SYMBOL(__ide_dma_begin);
 
 /* returns 1 on error, 0 otherwise */
 int __ide_dma_end (ide_drive_t *drive)
@@ -678,46 +784,17 @@ int __ide_dma_good_drive (ide_drive_t *drive)
 
 EXPORT_SYMBOL(__ide_dma_good_drive);
 
-int ide_use_dma(ide_drive_t *drive)
-{
-       struct hd_driveid *id = drive->id;
-       ide_hwif_t *hwif = drive->hwif;
-
-       /* consult the list of known "bad" drives */
-       if (__ide_dma_bad_drive(drive))
-               return 0;
-
-       /* capable of UltraDMA modes */
-       if (id->field_valid & 4) {
-               if (hwif->ultra_mask & id->dma_ultra)
-                       return 1;
-       }
-
-       /* capable of regular DMA modes */
-       if (id->field_valid & 2) {
-               if (hwif->mwdma_mask & id->dma_mword)
-                       return 1;
-               if (hwif->swdma_mask & id->dma_1word)
-                       return 1;
-       }
-
-       /* consult the list of known "good" drives */
-       if (__ide_dma_good_drive(drive) && id->eide_dma_time < 150)
-               return 1;
-
-       return 0;
-}
-
-EXPORT_SYMBOL_GPL(ide_use_dma);
-
-void ide_dma_verbose(ide_drive_t *drive)
+#ifdef CONFIG_BLK_DEV_IDEDMA_PCI
+int __ide_dma_verbose (ide_drive_t *drive)
 {
        struct hd_driveid *id   = drive->id;
        ide_hwif_t *hwif        = HWIF(drive);
 
        if (id->field_valid & 4) {
-               if ((id->dma_ultra >> 8) && (id->dma_mword >> 8))
-                       goto bug_dma_off;
+               if ((id->dma_ultra >> 8) && (id->dma_mword >> 8)) {
+                       printk(", BUG DMA OFF");
+                       return hwif->ide_dma_off_quietly(drive);
+               }
                if (id->dma_ultra & ((id->dma_ultra >> 8) & hwif->ultra_mask)) {
                        if (((id->dma_ultra >> 11) & 0x1F) &&
                            eighty_ninty_three(drive)) {
@@ -747,22 +824,19 @@ void ide_dma_verbose(ide_drive_t *drive)
                        printk(", (U)DMA");     /* Can be BIOS-enabled! */
                }
        } else if (id->field_valid & 2) {
-               if ((id->dma_mword >> 8) && (id->dma_1word >> 8))
-                       goto bug_dma_off;
+               if ((id->dma_mword >> 8) && (id->dma_1word >> 8)) {
+                       printk(", BUG DMA OFF");
+                       return hwif->ide_dma_off_quietly(drive);
+               }
                printk(", DMA");
        } else if (id->field_valid & 1) {
                printk(", BUG");
        }
-       return;
-bug_dma_off:
-       printk(", BUG DMA OFF");
-       hwif->ide_dma_off_quietly(drive);
-       return;
+       return 1;
 }
 
-EXPORT_SYMBOL(ide_dma_verbose);
+EXPORT_SYMBOL(__ide_dma_verbose);
 
-#ifdef CONFIG_BLK_DEV_IDEDMA_PCI
 int __ide_dma_lostirq (ide_drive_t *drive)
 {
        printk("%s: DMA interrupt recovery\n", drive->name);
@@ -794,6 +868,10 @@ int ide_release_dma_engine (ide_hwif_t *hwif)
                                    hwif->dmatable_dma);
                hwif->dmatable_cpu = NULL;
        }
+       if (hwif->sg_table) {
+               kfree(hwif->sg_table);
+               hwif->sg_table = NULL;
+       }
        return 1;
 }
 
@@ -826,12 +904,15 @@ int ide_allocate_dma_engine (ide_hwif_t *hwif)
        hwif->dmatable_cpu = pci_alloc_consistent(hwif->pci_dev,
                                                  PRD_ENTRIES * PRD_BYTES,
                                                  &hwif->dmatable_dma);
+       hwif->sg_table = kmalloc(sizeof(struct scatterlist) * PRD_ENTRIES,
+                               GFP_KERNEL);
 
-       if (hwif->dmatable_cpu)
+       if ((hwif->dmatable_cpu) && (hwif->sg_table))
                return 0;
 
-       printk(KERN_ERR "%s: -- Error, unable to allocate%s DMA table(s).\n",
+       printk(KERN_ERR "%s: -- Error, unable to allocate%s%s table(s).\n",
                (hwif->dmatable_cpu == NULL) ? " CPU" : "",
+               (hwif->sg_table == NULL) ?  " SG DMA" : " DMA",
                hwif->cds->name);
 
        ide_release_dma_engine(hwif);
@@ -927,16 +1008,18 @@ void ide_setup_dma (ide_hwif_t *hwif, unsigned long dma_base, unsigned int num_p
                hwif->ide_dma_host_on = &__ide_dma_host_on;
        if (!hwif->ide_dma_check)
                hwif->ide_dma_check = &__ide_dma_check;
-       if (!hwif->dma_setup)
-               hwif->dma_setup = &ide_dma_setup;
-       if (!hwif->dma_exec_cmd)
-               hwif->dma_exec_cmd = &ide_dma_exec_cmd;
-       if (!hwif->dma_start)
-               hwif->dma_start = &ide_dma_start;
+       if (!hwif->ide_dma_read)
+               hwif->ide_dma_read = &__ide_dma_read;
+       if (!hwif->ide_dma_write)
+               hwif->ide_dma_write = &__ide_dma_write;
+       if (!hwif->ide_dma_begin)
+               hwif->ide_dma_begin = &__ide_dma_begin;
        if (!hwif->ide_dma_end)
                hwif->ide_dma_end = &__ide_dma_end;
        if (!hwif->ide_dma_test_irq)
                hwif->ide_dma_test_irq = &__ide_dma_test_irq;
+       if (!hwif->ide_dma_verbose)
+               hwif->ide_dma_verbose = &__ide_dma_verbose;
        if (!hwif->ide_dma_timeout)
                hwif->ide_dma_timeout = &__ide_dma_timeout;
        if (!hwif->ide_dma_lostirq)