This commit was manufactured by cvs2svn to create branch 'vserver'.
[linux-2.6.git] / arch / ppc / platforms / hdpu.c
diff --git a/arch/ppc/platforms/hdpu.c b/arch/ppc/platforms/hdpu.c
new file mode 100644 (file)
index 0000000..b659d7b
--- /dev/null
@@ -0,0 +1,1062 @@
+
+/*
+ * arch/ppc/platforms/hdpu_setup.c
+ *
+ * Board setup routines for the Sky Computers HDPU Compute Blade.
+ *
+ * Written by Brian Waite <waite@skycomputers.com>
+ *
+ * Based on code done by - Mark A. Greer <mgreer@mvista.com>
+ *                         Rabeeh Khoury - rabeeh@galileo.co.il
+ *
+ * This program is free software; you can redistribute  it and/or modify it
+ * under  the terms of  the GNU General  Public License as published by the
+ * Free Software Foundation;  either version 2 of the  License, or (at your
+ * option) any later version.
+ */
+
+#include <linux/config.h>
+
+#include <linux/pci.h>
+#include <linux/delay.h>
+#include <linux/irq.h>
+#include <linux/ide.h>
+#include <linux/seq_file.h>
+
+#include <linux/initrd.h>
+#include <linux/root_dev.h>
+#include <linux/smp.h>
+
+#include <asm/time.h>
+#include <asm/machdep.h>
+#include <asm/todc.h>
+#include <asm/mv64x60.h>
+#include <asm/ppcboot.h>
+#include <platforms/hdpu.h>
+#include <linux/mv643xx.h>
+#include <linux/hdpu_features.h>
+#include <linux/device.h>
+#include <linux/mtd/physmap.h>
+
+#define BOARD_VENDOR   "Sky Computers"
+#define BOARD_MACHINE  "HDPU-CB-A"
+
+bd_t ppcboot_bd;
+int ppcboot_bd_valid = 0;
+
+static mv64x60_handle_t bh;
+
+extern char cmd_line[];
+
+unsigned long hdpu_find_end_of_memory(void);
+void hdpu_mpsc_progress(char *s, unsigned short hex);
+void hdpu_heartbeat(void);
+
+static void parse_bootinfo(unsigned long r3,
+                          unsigned long r4, unsigned long r5,
+                          unsigned long r6, unsigned long r7);
+static void hdpu_set_l1pe(void);
+static void hdpu_cpustate_set(unsigned char new_state);
+#ifdef CONFIG_SMP
+static spinlock_t timebase_lock = SPIN_LOCK_UNLOCKED;
+static unsigned int timebase_upper = 0, timebase_lower = 0;
+extern int smp_tb_synchronized;
+
+void __devinit hdpu_tben_give(void);
+void __devinit hdpu_tben_take(void);
+#endif
+
+static int __init
+hdpu_map_irq(struct pci_dev *dev, unsigned char idsel, unsigned char pin)
+{
+       struct pci_controller *hose = pci_bus_to_hose(dev->bus->number);
+
+       if (hose->index == 0) {
+               static char pci_irq_table[][4] = {
+                       {HDPU_PCI_0_IRQ, 0, 0, 0},
+                       {HDPU_PCI_0_IRQ, 0, 0, 0},
+               };
+
+               const long min_idsel = 1, max_idsel = 2, irqs_per_slot = 4;
+               return PCI_IRQ_TABLE_LOOKUP;
+       } else {
+               static char pci_irq_table[][4] = {
+                       {HDPU_PCI_1_IRQ, 0, 0, 0},
+               };
+
+               const long min_idsel = 1, max_idsel = 1, irqs_per_slot = 4;
+               return PCI_IRQ_TABLE_LOOKUP;
+       }
+}
+
+static void __init hdpu_intr_setup(void)
+{
+       mv64x60_write(&bh, MV64x60_GPP_IO_CNTL,
+                     (1 | (1 << 2) | (1 << 3) | (1 << 4) | (1 << 5) |
+                      (1 << 6) | (1 << 7) | (1 << 12) | (1 << 16) |
+                      (1 << 18) | (1 << 19) | (1 << 20) | (1 << 21) |
+                      (1 << 22) | (1 << 23) | (1 << 24) | (1 << 25) |
+                      (1 << 26) | (1 << 27) | (1 << 28) | (1 << 29)));
+
+       /* XXXX Erranum FEr PCI-#8 */
+       mv64x60_clr_bits(&bh, MV64x60_PCI0_CMD, (1 << 5) | (1 << 9));
+       mv64x60_clr_bits(&bh, MV64x60_PCI1_CMD, (1 << 5) | (1 << 9));
+
+       /*
+        * Dismiss and then enable interrupt on GPP interrupt cause
+        * for CPU #0
+        */
+       mv64x60_write(&bh, MV64x60_GPP_INTR_CAUSE, ~((1 << 8) | (1 << 13)));
+       mv64x60_set_bits(&bh, MV64x60_GPP_INTR_MASK, (1 << 8) | (1 << 13));
+
+       /*
+        * Dismiss and then enable interrupt on CPU #0 high cause reg
+        * BIT25 summarizes GPP interrupts 8-15
+        */
+       mv64x60_set_bits(&bh, MV64360_IC_CPU0_INTR_MASK_HI, (1 << 25));
+}
+
+static void __init hdpu_setup_peripherals(void)
+{
+       unsigned int val;
+
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2BOOT_WIN,
+                                HDPU_EMB_FLASH_BASE, HDPU_EMB_FLASH_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2BOOT_WIN);
+
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2DEV_0_WIN,
+                                HDPU_TBEN_BASE, HDPU_TBEN_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2DEV_0_WIN);
+
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2DEV_1_WIN,
+                                HDPU_NEXUS_ID_BASE, HDPU_NEXUS_ID_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2DEV_1_WIN);
+
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2SRAM_WIN,
+                                HDPU_INTERNAL_SRAM_BASE,
+                                HDPU_INTERNAL_SRAM_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2SRAM_WIN);
+
+       bh.ci->disable_window_32bit(&bh, MV64x60_ENET2MEM_4_WIN);
+       mv64x60_set_32bit_window(&bh, MV64x60_ENET2MEM_4_WIN, 0, 0, 0);
+
+       mv64x60_clr_bits(&bh, MV64x60_PCI0_PCI_DECODE_CNTL, (1 << 3));
+       mv64x60_clr_bits(&bh, MV64x60_PCI1_PCI_DECODE_CNTL, (1 << 3));
+       mv64x60_clr_bits(&bh, MV64x60_TIMR_CNTR_0_3_CNTL,
+                        ((1 << 0) | (1 << 8) | (1 << 16) | (1 << 24)));
+
+       /* Enable pipelining */
+       mv64x60_set_bits(&bh, MV64x60_CPU_CONFIG, (1 << 13));
+       /* Enable Snoop Pipelineing */
+       mv64x60_set_bits(&bh, MV64360_D_UNIT_CONTROL_HIGH, (1 << 24));
+
+       /*
+        * Change DRAM read buffer assignment.
+        * Assign read buffer 0 dedicated only for CPU,
+        * and the rest read buffer 1.
+        */
+       val = mv64x60_read(&bh, MV64360_SDRAM_CONFIG);
+       val = val & 0x03ffffff;
+       val = val | 0xf8000000;
+       mv64x60_write(&bh, MV64360_SDRAM_CONFIG, val);
+
+       /*
+        * Configure internal SRAM -
+        * Cache coherent write back, if CONFIG_MV64360_SRAM_CACHE_COHERENT set
+        * Parity enabled.
+        * Parity error propagation
+        * Arbitration not parked for CPU only
+        * Other bits are reserved.
+        */
+#ifdef CONFIG_MV64360_SRAM_CACHE_COHERENT
+       mv64x60_write(&bh, MV64360_SRAM_CONFIG, 0x001600b2);
+#else
+       mv64x60_write(&bh, MV64360_SRAM_CONFIG, 0x001600b0);
+#endif
+
+       hdpu_intr_setup();
+}
+
+static void __init hdpu_setup_bridge(void)
+{
+       struct mv64x60_setup_info si;
+       int i;
+
+       memset(&si, 0, sizeof(si));
+
+       si.phys_reg_base = HDPU_BRIDGE_REG_BASE;
+       si.pci_0.enable_bus = 1;
+       si.pci_0.pci_io.cpu_base = HDPU_PCI0_IO_START_PROC_ADDR;
+       si.pci_0.pci_io.pci_base_hi = 0;
+       si.pci_0.pci_io.pci_base_lo = HDPU_PCI0_IO_START_PCI_ADDR;
+       si.pci_0.pci_io.size = HDPU_PCI0_IO_SIZE;
+       si.pci_0.pci_io.swap = MV64x60_CPU2PCI_SWAP_NONE;
+       si.pci_0.pci_mem[0].cpu_base = HDPU_PCI0_MEM_START_PROC_ADDR;
+       si.pci_0.pci_mem[0].pci_base_hi = HDPU_PCI0_MEM_START_PCI_HI_ADDR;
+       si.pci_0.pci_mem[0].pci_base_lo = HDPU_PCI0_MEM_START_PCI_LO_ADDR;
+       si.pci_0.pci_mem[0].size = HDPU_PCI0_MEM_SIZE;
+       si.pci_0.pci_mem[0].swap = MV64x60_CPU2PCI_SWAP_NONE;
+       si.pci_0.pci_cmd_bits = 0;
+       si.pci_0.latency_timer = 0x80;
+
+       si.pci_1.enable_bus = 1;
+       si.pci_1.pci_io.cpu_base = HDPU_PCI1_IO_START_PROC_ADDR;
+       si.pci_1.pci_io.pci_base_hi = 0;
+       si.pci_1.pci_io.pci_base_lo = HDPU_PCI1_IO_START_PCI_ADDR;
+       si.pci_1.pci_io.size = HDPU_PCI1_IO_SIZE;
+       si.pci_1.pci_io.swap = MV64x60_CPU2PCI_SWAP_NONE;
+       si.pci_1.pci_mem[0].cpu_base = HDPU_PCI1_MEM_START_PROC_ADDR;
+       si.pci_1.pci_mem[0].pci_base_hi = HDPU_PCI1_MEM_START_PCI_HI_ADDR;
+       si.pci_1.pci_mem[0].pci_base_lo = HDPU_PCI1_MEM_START_PCI_LO_ADDR;
+       si.pci_1.pci_mem[0].size = HDPU_PCI1_MEM_SIZE;
+       si.pci_1.pci_mem[0].swap = MV64x60_CPU2PCI_SWAP_NONE;
+       si.pci_1.pci_cmd_bits = 0;
+       si.pci_1.latency_timer = 0x80;
+
+       for (i = 0; i < MV64x60_CPU2MEM_WINDOWS; i++) {
+#if defined(CONFIG_NOT_COHERENT_CACHE)
+               si.cpu_prot_options[i] = 0;
+               si.enet_options[i] = MV64360_ENET2MEM_SNOOP_NONE;
+               si.mpsc_options[i] = MV64360_MPSC2MEM_SNOOP_NONE;
+               si.idma_options[i] = MV64360_IDMA2MEM_SNOOP_NONE;
+
+               si.pci_1.acc_cntl_options[i] =
+                   MV64360_PCI_ACC_CNTL_SNOOP_NONE |
+                   MV64360_PCI_ACC_CNTL_SWAP_NONE |
+                   MV64360_PCI_ACC_CNTL_MBURST_128_BYTES |
+                   MV64360_PCI_ACC_CNTL_RDSIZE_256_BYTES;
+
+               si.pci_0.acc_cntl_options[i] =
+                   MV64360_PCI_ACC_CNTL_SNOOP_NONE |
+                   MV64360_PCI_ACC_CNTL_SWAP_NONE |
+                   MV64360_PCI_ACC_CNTL_MBURST_128_BYTES |
+                   MV64360_PCI_ACC_CNTL_RDSIZE_256_BYTES;
+
+#else
+               si.cpu_prot_options[i] = 0;
+               si.enet_options[i] = MV64360_ENET2MEM_SNOOP_WB; /* errata */
+               si.mpsc_options[i] = MV64360_MPSC2MEM_SNOOP_WB; /* errata */
+               si.idma_options[i] = MV64360_IDMA2MEM_SNOOP_WB; /* errata */
+
+               si.pci_0.acc_cntl_options[i] =
+                   MV64360_PCI_ACC_CNTL_SNOOP_WB |
+                   MV64360_PCI_ACC_CNTL_SWAP_NONE |
+                   MV64360_PCI_ACC_CNTL_MBURST_32_BYTES |
+                   MV64360_PCI_ACC_CNTL_RDSIZE_256_BYTES;
+
+               si.pci_1.acc_cntl_options[i] =
+                   MV64360_PCI_ACC_CNTL_SNOOP_WB |
+                   MV64360_PCI_ACC_CNTL_SWAP_NONE |
+                   MV64360_PCI_ACC_CNTL_MBURST_32_BYTES |
+                   MV64360_PCI_ACC_CNTL_RDSIZE_256_BYTES;
+#endif
+       }
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_INIT_PCI);
+
+       /* Lookup PCI host bridges */
+       mv64x60_init(&bh, &si);
+       pci_dram_offset = 0;    /* System mem at same addr on PCI & cpu bus */
+       ppc_md.pci_swizzle = common_swizzle;
+       ppc_md.pci_map_irq = hdpu_map_irq;
+
+       mv64x60_set_bus(&bh, 0, 0);
+       bh.hose_a->first_busno = 0;
+       bh.hose_a->last_busno = 0xff;
+       bh.hose_a->last_busno = pciauto_bus_scan(bh.hose_a, 0);
+
+       bh.hose_b->first_busno = bh.hose_a->last_busno + 1;
+       mv64x60_set_bus(&bh, 1, bh.hose_b->first_busno);
+       bh.hose_b->last_busno = 0xff;
+       bh.hose_b->last_busno = pciauto_bus_scan(bh.hose_b,
+               bh.hose_b->first_busno);
+
+       ppc_md.pci_exclude_device = mv64x60_pci_exclude_device;
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_INIT_REG);
+       /*
+        * Enabling of PCI internal-vs-external arbitration
+        * is a platform- and errata-dependent decision.
+        */
+       return;
+}
+
+#if defined(CONFIG_SERIAL_MPSC_CONSOLE)
+static void __init hdpu_early_serial_map(void)
+{
+#ifdef CONFIG_KGDB
+       static char first_time = 1;
+
+#if defined(CONFIG_KGDB_TTYS0)
+#define KGDB_PORT 0
+#elif defined(CONFIG_KGDB_TTYS1)
+#define KGDB_PORT 1
+#else
+#error "Invalid kgdb_tty port"
+#endif
+
+       if (first_time) {
+               gt_early_mpsc_init(KGDB_PORT,
+                                  B9600 | CS8 | CREAD | HUPCL | CLOCAL);
+               first_time = 0;
+       }
+
+       return;
+#endif
+}
+#endif
+
+static void hdpu_init2(void)
+{
+       return;
+}
+
+#if defined(CONFIG_MV643XX_ETH)
+static void __init hdpu_fixup_eth_pdata(struct platform_device *pd)
+{
+
+       struct mv643xx_eth_platform_data *eth_pd;
+       eth_pd = pd->dev.platform_data;
+
+       eth_pd->port_serial_control =
+           mv64x60_read(&bh, MV643XX_ETH_PORT_SERIAL_CONTROL_REG(pd->id) & ~1);
+
+       eth_pd->force_phy_addr = 1;
+       eth_pd->phy_addr = pd->id;
+       eth_pd->tx_queue_size = 400;
+       eth_pd->rx_queue_size = 800;
+}
+#endif
+
+static void __init hdpu_fixup_mpsc_pdata(struct platform_device *pd)
+{
+
+       struct mpsc_pdata *pdata;
+
+       pdata = (struct mpsc_pdata *)pd->dev.platform_data;
+
+       pdata->max_idle = 40;
+       if (ppcboot_bd_valid)
+               pdata->default_baud = ppcboot_bd.bi_baudrate;
+       else
+               pdata->default_baud = HDPU_DEFAULT_BAUD;
+       pdata->brg_clk_src = HDPU_MPSC_CLK_SRC;
+       pdata->brg_clk_freq = HDPU_MPSC_CLK_FREQ;
+}
+
+#if defined(CONFIG_HDPU_FEATURES)
+static void __init hdpu_fixup_cpustate_pdata(struct platform_device *pd)
+{
+       struct platform_device *pds[1];
+       pds[0] = pd;
+       mv64x60_pd_fixup(&bh, pds, 1);
+}
+#endif
+
+static int __init hdpu_platform_notify(struct device *dev)
+{
+       static struct {
+               char *bus_id;
+               void ((*rtn) (struct platform_device * pdev));
+       } dev_map[] = {
+               {
+               MPSC_CTLR_NAME ".0", hdpu_fixup_mpsc_pdata},
+#if defined(CONFIG_MV643XX_ETH)
+               {
+               MV643XX_ETH_NAME ".0", hdpu_fixup_eth_pdata},
+#endif
+#if defined(CONFIG_HDPU_FEATURES)
+               {
+               HDPU_CPUSTATE_NAME ".0", hdpu_fixup_cpustate_pdata},
+#endif
+       };
+       struct platform_device *pdev;
+       int i;
+
+       if (dev && dev->bus_id)
+               for (i = 0; i < ARRAY_SIZE(dev_map); i++)
+                       if (!strncmp(dev->bus_id, dev_map[i].bus_id,
+                                    BUS_ID_SIZE)) {
+
+                               pdev = container_of(dev,
+                                                   struct platform_device,
+                                                   dev);
+                               dev_map[i].rtn(pdev);
+                       }
+
+       return 0;
+}
+
+static void __init hdpu_setup_arch(void)
+{
+       if (ppc_md.progress)
+               ppc_md.progress("hdpu_setup_arch: enter", 0);
+#ifdef CONFIG_BLK_DEV_INITRD
+       if (initrd_start)
+               ROOT_DEV = Root_RAM0;
+       else
+#endif
+#ifdef CONFIG_ROOT_NFS
+               ROOT_DEV = Root_NFS;
+#else
+               ROOT_DEV = Root_SDA2;
+#endif
+
+       ppc_md.heartbeat = hdpu_heartbeat;
+
+       ppc_md.heartbeat_reset = HZ;
+       ppc_md.heartbeat_count = 1;
+
+       if (ppc_md.progress)
+               ppc_md.progress("hdpu_setup_arch: Enabling L2 cache", 0);
+
+       /* Enable L1 Parity Bits */
+       hdpu_set_l1pe();
+
+       /* Enable L2 and L3 caches (if 745x) */
+       _set_L2CR(0x80080000);
+
+       if (ppc_md.progress)
+               ppc_md.progress("hdpu_setup_arch: enter", 0);
+
+       hdpu_setup_bridge();
+
+       hdpu_setup_peripherals();
+
+#ifdef CONFIG_SERIAL_MPSC_CONSOLE
+       hdpu_early_serial_map();
+#endif
+
+       printk("SKY HDPU Compute Blade \n");
+
+       if (ppc_md.progress)
+               ppc_md.progress("hdpu_setup_arch: exit", 0);
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_OK);
+       return;
+}
+static void __init hdpu_init_irq(void)
+{
+       mv64360_init_irq();
+}
+
+static void __init hdpu_set_l1pe()
+{
+       unsigned long ictrl;
+       asm volatile ("mfspr %0, 1011":"=r" (ictrl):);
+       ictrl |= ICTRL_EICE | ICTRL_EDC | ICTRL_EICP;
+       asm volatile ("mtspr 1011, %0"::"r" (ictrl));
+}
+
+/*
+ * Set BAT 1 to map 0xf1000000 to end of physical memory space.
+ */
+static __inline__ void hdpu_set_bat(void)
+{
+       mb();
+       mtspr(SPRN_DBAT1U, 0xf10001fe);
+       mtspr(SPRN_DBAT1L, 0xf100002a);
+       mb();
+
+       return;
+}
+
+unsigned long __init hdpu_find_end_of_memory(void)
+{
+       return mv64x60_get_mem_size(CONFIG_MV64X60_NEW_BASE,
+                                   MV64x60_TYPE_MV64360);
+}
+
+static void hdpu_reset_board(void)
+{
+       volatile int infinite = 1;
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_RESET);
+
+       local_irq_disable();
+
+       /* Clear all the LEDs */
+       mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, ((1 << 4) |
+                                                  (1 << 5) | (1 << 6)));
+
+       /* disable and invalidate the L2 cache */
+       _set_L2CR(0);
+       _set_L2CR(0x200000);
+
+       /* flush and disable L1 I/D cache */
+       __asm__ __volatile__
+           ("\n"
+            "mfspr   3,1008\n"
+            "ori       5,5,0xcc00\n"
+            "ori       4,3,0xc00\n"
+            "andc      5,3,5\n"
+            "sync\n"
+            "mtspr     1008,4\n"
+            "isync\n" "sync\n" "mtspr  1008,5\n" "isync\n" "sync\n");
+
+       /* Hit the reset bit */
+       mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, (1 << 3));
+
+       while (infinite)
+               infinite = infinite;
+
+       return;
+}
+
+static void hdpu_restart(char *cmd)
+{
+       volatile ulong i = 10000000;
+
+       hdpu_reset_board();
+
+       while (i-- > 0) ;
+       panic("restart failed\n");
+}
+
+static void hdpu_halt(void)
+{
+       local_irq_disable();
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_HALT);
+
+       /* Clear all the LEDs */
+       mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, ((1 << 4) | (1 << 5) |
+                                                  (1 << 6)));
+       while (1) ;
+       /* NOTREACHED */
+}
+
+static void hdpu_power_off(void)
+{
+       hdpu_halt();
+       /* NOTREACHED */
+}
+
+static int hdpu_show_cpuinfo(struct seq_file *m)
+{
+       uint pvid;
+
+       pvid = mfspr(SPRN_PVR);
+       seq_printf(m, "vendor\t\t: Sky Computers\n");
+       seq_printf(m, "machine\t\t: HDPU Compute Blade\n");
+       seq_printf(m, "PVID\t\t: 0x%x, vendor: %s\n",
+                  pvid, (pvid & (1 << 15) ? "IBM" : "Motorola"));
+
+       return 0;
+}
+
+static void __init hdpu_calibrate_decr(void)
+{
+       ulong freq;
+
+       if (ppcboot_bd_valid)
+               freq = ppcboot_bd.bi_busfreq / 4;
+       else
+               freq = 133000000;
+
+       printk("time_init: decrementer frequency = %lu.%.6lu MHz\n",
+              freq / 1000000, freq % 1000000);
+
+       tb_ticks_per_jiffy = freq / HZ;
+       tb_to_us = mulhwu_scale_factor(freq, 1000000);
+
+       return;
+}
+
+static void parse_bootinfo(unsigned long r3,
+                          unsigned long r4, unsigned long r5,
+                          unsigned long r6, unsigned long r7)
+{
+       bd_t *bd = NULL;
+       char *cmdline_start = NULL;
+       int cmdline_len = 0;
+
+       if (r3) {
+               if ((r3 & 0xf0000000) == 0)
+                       r3 += KERNELBASE;
+               if ((r3 & 0xf0000000) == KERNELBASE) {
+                       bd = (void *)r3;
+
+                       memcpy(&ppcboot_bd, bd, sizeof(ppcboot_bd));
+                       ppcboot_bd_valid = 1;
+               }
+       }
+#ifdef CONFIG_BLK_DEV_INITRD
+       if (r4 && r5 && r5 > r4) {
+               if ((r4 & 0xf0000000) == 0)
+                       r4 += KERNELBASE;
+               if ((r5 & 0xf0000000) == 0)
+                       r5 += KERNELBASE;
+               if ((r4 & 0xf0000000) == KERNELBASE) {
+                       initrd_start = r4;
+                       initrd_end = r5;
+                       initrd_below_start_ok = 1;
+               }
+       }
+#endif                         /* CONFIG_BLK_DEV_INITRD */
+
+       if (r6 && r7 && r7 > r6) {
+               if ((r6 & 0xf0000000) == 0)
+                       r6 += KERNELBASE;
+               if ((r7 & 0xf0000000) == 0)
+                       r7 += KERNELBASE;
+               if ((r6 & 0xf0000000) == KERNELBASE) {
+                       cmdline_start = (void *)r6;
+                       cmdline_len = (r7 - r6);
+                       strncpy(cmd_line, cmdline_start, cmdline_len);
+               }
+       }
+}
+
+#if defined(CONFIG_BLK_DEV_IDE) || defined(CONFIG_BLK_DEV_IDE_MODULE)
+static int hdpu_ide_check_region(ide_ioreg_t from, unsigned int extent)
+{
+       return check_region(from, extent);
+}
+
+static void
+hdpu_ide_request_region(ide_ioreg_t from, unsigned int extent, const char *name)
+{
+       request_region(from, extent, name);
+       return;
+}
+
+static void hdpu_ide_release_region(ide_ioreg_t from, unsigned int extent)
+{
+       release_region(from, extent);
+       return;
+}
+
+static void __init
+hdpu_ide_pci_init_hwif_ports(hw_regs_t * hw, ide_ioreg_t data_port,
+                            ide_ioreg_t ctrl_port, int *irq)
+{
+       struct pci_dev *dev;
+
+       pci_for_each_dev(dev) {
+               if (((dev->class >> 8) == PCI_CLASS_STORAGE_IDE) ||
+                   ((dev->class >> 8) == PCI_CLASS_STORAGE_RAID)) {
+                       hw->irq = dev->irq;
+
+                       if (irq != NULL) {
+                               *irq = dev->irq;
+                       }
+               }
+       }
+
+       return;
+}
+#endif
+
+void hdpu_heartbeat(void)
+{
+       if (mv64x60_read(&bh, MV64x60_GPP_VALUE) & (1 << 5))
+               mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, (1 << 5));
+       else
+               mv64x60_write(&bh, MV64x60_GPP_VALUE_SET, (1 << 5));
+
+       ppc_md.heartbeat_count = ppc_md.heartbeat_reset;
+
+}
+
+static void __init hdpu_map_io(void)
+{
+       io_block_mapping(0xf1000000, 0xf1000000, 0x20000, _PAGE_IO);
+}
+
+#ifdef CONFIG_SMP
+char hdpu_smp0[] = "SMP Cpu #0";
+char hdpu_smp1[] = "SMP Cpu #1";
+
+static irqreturn_t hdpu_smp_cpu0_int_handler(int irq, void *dev_id,
+                                            struct pt_regs *regs)
+{
+       volatile unsigned int doorbell;
+
+       doorbell = mv64x60_read(&bh, MV64360_CPU0_DOORBELL);
+
+       /* Ack the doorbell interrupts */
+       mv64x60_write(&bh, MV64360_CPU0_DOORBELL_CLR, doorbell);
+
+       if (doorbell & 1) {
+               smp_message_recv(0, regs);
+       }
+       if (doorbell & 2) {
+               smp_message_recv(1, regs);
+       }
+       if (doorbell & 4) {
+               smp_message_recv(2, regs);
+       }
+       if (doorbell & 8) {
+               smp_message_recv(3, regs);
+       }
+       return IRQ_HANDLED;
+}
+
+static irqreturn_t hdpu_smp_cpu1_int_handler(int irq, void *dev_id,
+                                            struct pt_regs *regs)
+{
+       volatile unsigned int doorbell;
+
+       doorbell = mv64x60_read(&bh, MV64360_CPU1_DOORBELL);
+
+       /* Ack the doorbell interrupts */
+       mv64x60_write(&bh, MV64360_CPU1_DOORBELL_CLR, doorbell);
+
+       if (doorbell & 1) {
+               smp_message_recv(0, regs);
+       }
+       if (doorbell & 2) {
+               smp_message_recv(1, regs);
+       }
+       if (doorbell & 4) {
+               smp_message_recv(2, regs);
+       }
+       if (doorbell & 8) {
+               smp_message_recv(3, regs);
+       }
+       return IRQ_HANDLED;
+}
+
+static void smp_hdpu_CPU_two(void)
+{
+       __asm__ __volatile__
+           ("\n"
+            "lis     3,0x0000\n"
+            "ori     3,3,0x00c0\n"
+            "mtspr   26, 3\n" "li      4,0\n" "mtspr   27,4\n" "rfi");
+
+}
+
+static int smp_hdpu_probe(void)
+{
+       int *cpu_count_reg;
+       int num_cpus = 0;
+
+       cpu_count_reg = ioremap(HDPU_NEXUS_ID_BASE, HDPU_NEXUS_ID_SIZE);
+       if (cpu_count_reg) {
+               num_cpus = (*cpu_count_reg >> 20) & 0x3;
+               iounmap(cpu_count_reg);
+       }
+
+       /* Validate the bits in the CPLD. If we could not map the reg, return 2.
+        * If the register reported 0 or 3, return 2.
+        * Older CPLD revisions set these bits to all ones (val = 3).
+        */
+       if ((num_cpus < 1) || (num_cpus > 2)) {
+               printk
+                   ("Unable to determine the number of processors %d . deafulting to 2.\n",
+                    num_cpus);
+               num_cpus = 2;
+       }
+       return num_cpus;
+}
+
+static void
+smp_hdpu_message_pass(int target, int msg, unsigned long data, int wait)
+{
+       if (msg > 0x3) {
+               printk("SMP %d: smp_message_pass: unknown msg %d\n",
+                      smp_processor_id(), msg);
+               return;
+       }
+       switch (target) {
+       case MSG_ALL:
+               mv64x60_write(&bh, MV64360_CPU0_DOORBELL, 1 << msg);
+               mv64x60_write(&bh, MV64360_CPU1_DOORBELL, 1 << msg);
+               break;
+       case MSG_ALL_BUT_SELF:
+               if (smp_processor_id())
+                       mv64x60_write(&bh, MV64360_CPU0_DOORBELL, 1 << msg);
+               else
+                       mv64x60_write(&bh, MV64360_CPU1_DOORBELL, 1 << msg);
+               break;
+       default:
+               if (target == 0)
+                       mv64x60_write(&bh, MV64360_CPU0_DOORBELL, 1 << msg);
+               else
+                       mv64x60_write(&bh, MV64360_CPU1_DOORBELL, 1 << msg);
+               break;
+       }
+}
+
+static void smp_hdpu_kick_cpu(int nr)
+{
+       volatile unsigned int *bootaddr;
+
+       if (ppc_md.progress)
+               ppc_md.progress("smp_hdpu_kick_cpu", 0);
+
+       hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR | CPUSTATE_KERNEL_CPU1_KICK);
+
+       /* Disable BootCS. Must also reduce the windows size to zero. */
+       bh.ci->disable_window_32bit(&bh, MV64x60_CPU2BOOT_WIN);
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2BOOT_WIN, 0, 0, 0);
+
+       bootaddr = ioremap(HDPU_INTERNAL_SRAM_BASE, HDPU_INTERNAL_SRAM_SIZE);
+       if (!bootaddr) {
+               if (ppc_md.progress)
+                       ppc_md.progress("smp_hdpu_kick_cpu: ioremap failed", 0);
+               return;
+       }
+
+       memcpy((void *)(bootaddr + 0x40), (void *)&smp_hdpu_CPU_two, 0x20);
+
+       /* map SRAM to 0xfff00000 */
+       bh.ci->disable_window_32bit(&bh, MV64x60_CPU2SRAM_WIN);
+
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2SRAM_WIN,
+                                0xfff00000, HDPU_INTERNAL_SRAM_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2SRAM_WIN);
+
+       /* Enable CPU1 arbitration */
+       mv64x60_clr_bits(&bh, MV64x60_CPU_MASTER_CNTL, (1 << 9));
+
+       /*
+        * Wait 100mSecond until other CPU has reached __secondary_start.
+        * When it reaches, it is permittable to rever the SRAM mapping etc...
+        */
+       mdelay(100);
+       *(unsigned long *)KERNELBASE = nr;
+       asm volatile ("dcbf 0,%0"::"r" (KERNELBASE):"memory");
+
+       iounmap(bootaddr);
+
+       /* Set up window for internal sram (256KByte insize) */
+       bh.ci->disable_window_32bit(&bh, MV64x60_CPU2SRAM_WIN);
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2SRAM_WIN,
+                                HDPU_INTERNAL_SRAM_BASE,
+                                HDPU_INTERNAL_SRAM_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2SRAM_WIN);
+       /*
+        * Set up windows for embedded FLASH (using boot CS window).
+        */
+
+       bh.ci->disable_window_32bit(&bh, MV64x60_CPU2BOOT_WIN);
+       mv64x60_set_32bit_window(&bh, MV64x60_CPU2BOOT_WIN,
+                                HDPU_EMB_FLASH_BASE, HDPU_EMB_FLASH_SIZE, 0);
+       bh.ci->enable_window_32bit(&bh, MV64x60_CPU2BOOT_WIN);
+}
+
+static void smp_hdpu_setup_cpu(int cpu_nr)
+{
+       if (cpu_nr == 0) {
+               if (ppc_md.progress)
+                       ppc_md.progress("smp_hdpu_setup_cpu 0", 0);
+               mv64x60_write(&bh, MV64360_CPU0_DOORBELL_CLR, 0xff);
+               mv64x60_write(&bh, MV64360_CPU0_DOORBELL_MASK, 0xff);
+               request_irq(60, hdpu_smp_cpu0_int_handler,
+                           SA_INTERRUPT, hdpu_smp0, 0);
+       }
+
+       if (cpu_nr == 1) {
+               if (ppc_md.progress)
+                       ppc_md.progress("smp_hdpu_setup_cpu 1", 0);
+
+               hdpu_cpustate_set(CPUSTATE_KERNEL_MAJOR |
+                                 CPUSTATE_KERNEL_CPU1_OK);
+
+               /* Enable L1 Parity Bits */
+               hdpu_set_l1pe();
+
+               /* Enable L2 cache */
+               _set_L2CR(0);
+               _set_L2CR(0x80080000);
+
+               mv64x60_write(&bh, MV64360_CPU1_DOORBELL_CLR, 0x0);
+               mv64x60_write(&bh, MV64360_CPU1_DOORBELL_MASK, 0xff);
+               request_irq(28, hdpu_smp_cpu1_int_handler,
+                           SA_INTERRUPT, hdpu_smp1, 0);
+       }
+
+}
+
+void __devinit hdpu_tben_give()
+{
+       volatile unsigned long *val = 0;
+
+       /* By writing 0 to the TBEN_BASE, the timebases is frozen */
+       val = ioremap(HDPU_TBEN_BASE, 4);
+       *val = 0;
+       mb();
+
+       spin_lock(&timebase_lock);
+       timebase_upper = get_tbu();
+       timebase_lower = get_tbl();
+       spin_unlock(&timebase_lock);
+
+       while (timebase_upper || timebase_lower)
+               barrier();
+
+       /* By writing 1 to the TBEN_BASE, the timebases is thawed */
+       *val = 1;
+       mb();
+
+       iounmap(val);
+
+}
+
+void __devinit hdpu_tben_take()
+{
+       while (!(timebase_upper || timebase_lower))
+               barrier();
+
+       spin_lock(&timebase_lock);
+       set_tb(timebase_upper, timebase_lower);
+       timebase_upper = 0;
+       timebase_lower = 0;
+       spin_unlock(&timebase_lock);
+}
+
+static struct smp_ops_t hdpu_smp_ops = {
+       .message_pass = smp_hdpu_message_pass,
+       .probe = smp_hdpu_probe,
+       .kick_cpu = smp_hdpu_kick_cpu,
+       .setup_cpu = smp_hdpu_setup_cpu,
+       .give_timebase = hdpu_tben_give,
+       .take_timebase = hdpu_tben_take,
+};
+#endif                         /* CONFIG_SMP */
+
+void __init
+platform_init(unsigned long r3, unsigned long r4, unsigned long r5,
+             unsigned long r6, unsigned long r7)
+{
+       parse_bootinfo(r3, r4, r5, r6, r7);
+
+       isa_mem_base = 0;
+
+       ppc_md.setup_arch = hdpu_setup_arch;
+       ppc_md.init = hdpu_init2;
+       ppc_md.show_cpuinfo = hdpu_show_cpuinfo;
+       ppc_md.init_IRQ = hdpu_init_irq;
+       ppc_md.get_irq = mv64360_get_irq;
+       ppc_md.restart = hdpu_restart;
+       ppc_md.power_off = hdpu_power_off;
+       ppc_md.halt = hdpu_halt;
+       ppc_md.find_end_of_memory = hdpu_find_end_of_memory;
+       ppc_md.calibrate_decr = hdpu_calibrate_decr;
+       ppc_md.setup_io_mappings = hdpu_map_io;
+
+       bh.p_base = CONFIG_MV64X60_NEW_BASE;
+       bh.v_base = (unsigned long *)bh.p_base;
+
+       hdpu_set_bat();
+
+#if defined(CONFIG_SERIAL_TEXT_DEBUG)
+       ppc_md.progress = hdpu_mpsc_progress;   /* embedded UART */
+       mv64x60_progress_init(bh.p_base);
+#endif                         /* CONFIG_SERIAL_TEXT_DEBUG */
+
+#ifdef CONFIG_SMP
+       ppc_md.smp_ops = &hdpu_smp_ops;
+#endif                         /* CONFIG_SMP */
+
+#if defined(CONFIG_SERIAL_MPSC) || defined(CONFIG_MV643XX_ETH)
+       platform_notify = hdpu_platform_notify;
+#endif
+       return;
+}
+
+#if defined(CONFIG_SERIAL_TEXT_DEBUG) && defined(CONFIG_SERIAL_MPSC_CONSOLE)
+/* SMP safe version of the serial text debug routine. Uses Semaphore 0 */
+void hdpu_mpsc_progress(char *s, unsigned short hex)
+{
+       while (mv64x60_read(&bh, MV64360_WHO_AM_I) !=
+              mv64x60_read(&bh, MV64360_SEMAPHORE_0)) {
+       }
+       mv64x60_mpsc_progress(s, hex);
+       mv64x60_write(&bh, MV64360_SEMAPHORE_0, 0xff);
+}
+#endif
+
+static void hdpu_cpustate_set(unsigned char new_state)
+{
+       unsigned int state = (new_state << 21);
+       mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, (0xff << 21));
+       mv64x60_write(&bh, MV64x60_GPP_VALUE_CLR, state);
+}
+
+#ifdef CONFIG_MTD_PHYSMAP
+static struct mtd_partition hdpu_partitions[] = {
+       {
+        .name = "Root FS",
+        .size = 0x03400000,
+        .offset = 0,
+        .mask_flags = 0,
+        },{
+        .name = "User FS",
+        .size = 0x00800000,
+        .offset = 0x03400000,
+        .mask_flags = 0,
+        },{
+        .name = "Kernel Image",
+        .size = 0x002C0000,
+        .offset = 0x03C00000,
+        .mask_flags = 0,
+        },{
+        .name = "bootEnv",
+        .size = 0x00040000,
+        .offset = 0x03EC0000,
+        .mask_flags = 0,
+        },{
+        .name = "bootROM",
+        .size = 0x00100000,
+        .offset = 0x03F00000,
+        .mask_flags = 0,
+        }
+};
+
+static int __init hdpu_setup_mtd(void)
+{
+
+       physmap_set_partitions(hdpu_partitions, 5);
+       return 0;
+}
+
+arch_initcall(hdpu_setup_mtd);
+#endif
+
+#ifdef CONFIG_HDPU_FEATURES
+
+static struct resource hdpu_cpustate_resources[] = {
+       [0] = {
+              .name = "addr base",
+              .start = MV64x60_GPP_VALUE_SET,
+              .end = MV64x60_GPP_VALUE_CLR + 1,
+              .flags = IORESOURCE_MEM,
+              },
+};
+
+static struct resource hdpu_nexus_resources[] = {
+       [0] = {
+              .name = "nexus register",
+              .start = HDPU_NEXUS_ID_BASE,
+              .end = HDPU_NEXUS_ID_BASE + HDPU_NEXUS_ID_SIZE,
+              .flags = IORESOURCE_MEM,
+              },
+};
+
+static struct platform_device hdpu_cpustate_device = {
+       .name = HDPU_CPUSTATE_NAME,
+       .id = 0,
+       .num_resources = ARRAY_SIZE(hdpu_cpustate_resources),
+       .resource = hdpu_cpustate_resources,
+};
+
+static struct platform_device hdpu_nexus_device = {
+       .name = HDPU_NEXUS_NAME,
+       .id = 0,
+       .num_resources = ARRAY_SIZE(hdpu_nexus_resources),
+       .resource = hdpu_nexus_resources,
+};
+
+static int __init hdpu_add_pds(void)
+{
+       platform_device_register(&hdpu_cpustate_device);
+       platform_device_register(&hdpu_nexus_device);
+       return 0;
+}
+
+arch_initcall(hdpu_add_pds);
+#endif