This commit was manufactured by cvs2svn to create branch 'vserver'.
[linux-2.6.git] / drivers / infiniband / core / agent.c
diff --git a/drivers/infiniband/core/agent.c b/drivers/infiniband/core/agent.c
new file mode 100644 (file)
index 0000000..49f4f9f
--- /dev/null
@@ -0,0 +1,399 @@
+/*
+ * Copyright (c) 2004 Mellanox Technologies Ltd.  All rights reserved.
+ * Copyright (c) 2004 Infinicon Corporation.  All rights reserved.
+ * Copyright (c) 2004 Intel Corporation.  All rights reserved.
+ * Copyright (c) 2004 Topspin Corporation.  All rights reserved.
+ * Copyright (c) 2004 Voltaire Corporation.  All rights reserved.
+ *
+ * This software is available to you under a choice of one of two
+ * licenses.  You may choose to be licensed under the terms of the GNU
+ * General Public License (GPL) Version 2, available from the file
+ * COPYING in the main directory of this source tree, or the
+ * OpenIB.org BSD license below:
+ *
+ *     Redistribution and use in source and binary forms, with or
+ *     without modification, are permitted provided that the following
+ *     conditions are met:
+ *
+ *      - Redistributions of source code must retain the above
+ *        copyright notice, this list of conditions and the following
+ *        disclaimer.
+ *
+ *      - Redistributions in binary form must reproduce the above
+ *        copyright notice, this list of conditions and the following
+ *        disclaimer in the documentation and/or other materials
+ *        provided with the distribution.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+ * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
+ * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
+ * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
+ * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+ * SOFTWARE.
+ *
+ * $Id: agent.c 1389 2004-12-27 22:56:47Z roland $
+ */
+
+#include <linux/dma-mapping.h>
+
+#include <asm/bug.h>
+
+#include <ib_smi.h>
+
+#include "smi.h"
+#include "agent_priv.h"
+#include "mad_priv.h"
+
+
+spinlock_t ib_agent_port_list_lock;
+static LIST_HEAD(ib_agent_port_list);
+
+extern kmem_cache_t *ib_mad_cache;
+
+
+/*
+ * Caller must hold ib_agent_port_list_lock
+ */
+static inline struct ib_agent_port_private *
+__ib_get_agent_port(struct ib_device *device, int port_num,
+                   struct ib_mad_agent *mad_agent)
+{
+       struct ib_agent_port_private *entry;
+
+       BUG_ON(!(!!device ^ !!mad_agent));  /* Exactly one MUST be (!NULL) */
+
+       if (device) {
+               list_for_each_entry(entry, &ib_agent_port_list, port_list) {
+                       if (entry->dr_smp_agent->device == device &&
+                           entry->port_num == port_num)
+                               return entry;
+               }
+       } else {
+               list_for_each_entry(entry, &ib_agent_port_list, port_list) {
+                       if ((entry->dr_smp_agent == mad_agent) ||
+                           (entry->lr_smp_agent == mad_agent) ||
+                           (entry->perf_mgmt_agent == mad_agent))
+                               return entry;
+               }
+       }
+       return NULL;
+}
+
+static inline struct ib_agent_port_private *
+ib_get_agent_port(struct ib_device *device, int port_num,
+                 struct ib_mad_agent *mad_agent)
+{
+       struct ib_agent_port_private *entry;
+       unsigned long flags;
+
+       spin_lock_irqsave(&ib_agent_port_list_lock, flags);
+       entry = __ib_get_agent_port(device, port_num, mad_agent);
+       spin_unlock_irqrestore(&ib_agent_port_list_lock, flags);
+
+       return entry;
+}
+
+int smi_check_local_dr_smp(struct ib_smp *smp,
+                          struct ib_device *device,
+                          int port_num)
+{
+       struct ib_agent_port_private *port_priv;
+
+       if (smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE)
+               return 1;
+       port_priv = ib_get_agent_port(device, port_num, NULL);
+       if (!port_priv) {
+               printk(KERN_DEBUG SPFX "smi_check_local_dr_smp %s port %d "
+                      "not open\n",
+                      device->name, port_num);
+               return 1;
+       }
+
+       return smi_check_local_smp(port_priv->dr_smp_agent, smp);
+}
+
+static int agent_mad_send(struct ib_mad_agent *mad_agent,
+                         struct ib_agent_port_private *port_priv,
+                         struct ib_mad_private *mad_priv,
+                         struct ib_grh *grh,
+                         struct ib_wc *wc)
+{
+       struct ib_agent_send_wr *agent_send_wr;
+       struct ib_sge gather_list;
+       struct ib_send_wr send_wr;
+       struct ib_send_wr *bad_send_wr;
+       struct ib_ah_attr ah_attr;
+       unsigned long flags;
+       int ret = 1;
+
+       agent_send_wr = kmalloc(sizeof(*agent_send_wr), GFP_KERNEL);
+       if (!agent_send_wr)
+               goto out;
+       agent_send_wr->mad = mad_priv;
+
+       /* PCI mapping */
+       gather_list.addr = dma_map_single(mad_agent->device->dma_device,
+                                         &mad_priv->mad,
+                                         sizeof(mad_priv->mad),
+                                         DMA_TO_DEVICE);
+       gather_list.length = sizeof(mad_priv->mad);
+       gather_list.lkey = (*port_priv->mr).lkey;
+
+       send_wr.next = NULL;
+       send_wr.opcode = IB_WR_SEND;
+       send_wr.sg_list = &gather_list;
+       send_wr.num_sge = 1;
+       send_wr.wr.ud.remote_qpn = wc->src_qp; /* DQPN */
+       send_wr.wr.ud.timeout_ms = 0;
+       send_wr.send_flags = IB_SEND_SIGNALED | IB_SEND_SOLICITED;
+
+       ah_attr.dlid = wc->slid;
+       ah_attr.port_num = mad_agent->port_num;
+       ah_attr.src_path_bits = wc->dlid_path_bits;
+       ah_attr.sl = wc->sl;
+       ah_attr.static_rate = 0;
+       ah_attr.ah_flags = 0; /* No GRH */
+       if (mad_priv->mad.mad.mad_hdr.mgmt_class == IB_MGMT_CLASS_PERF_MGMT) {
+               if (wc->wc_flags & IB_WC_GRH) {
+                       ah_attr.ah_flags = IB_AH_GRH;
+                       /* Should sgid be looked up ? */
+                       ah_attr.grh.sgid_index = 0;
+                       ah_attr.grh.hop_limit = grh->hop_limit;
+                       ah_attr.grh.flow_label = be32_to_cpup(
+                               &grh->version_tclass_flow)  & 0xfffff;
+                       ah_attr.grh.traffic_class = (be32_to_cpup(
+                               &grh->version_tclass_flow) >> 20) & 0xff;
+                       memcpy(ah_attr.grh.dgid.raw,
+                              grh->sgid.raw,
+                              sizeof(ah_attr.grh.dgid));
+               }
+       }
+
+       agent_send_wr->ah = ib_create_ah(mad_agent->qp->pd, &ah_attr);
+       if (IS_ERR(agent_send_wr->ah)) {
+               printk(KERN_ERR SPFX "No memory for address handle\n");
+               kfree(agent_send_wr);
+               goto out;
+       }
+
+       send_wr.wr.ud.ah = agent_send_wr->ah;
+       if (mad_priv->mad.mad.mad_hdr.mgmt_class == IB_MGMT_CLASS_PERF_MGMT) {
+               send_wr.wr.ud.pkey_index = wc->pkey_index;
+               send_wr.wr.ud.remote_qkey = IB_QP1_QKEY;
+       } else {        /* for SMPs */
+               send_wr.wr.ud.pkey_index = 0;
+               send_wr.wr.ud.remote_qkey = 0;
+       }
+       send_wr.wr.ud.mad_hdr = &mad_priv->mad.mad.mad_hdr;
+       send_wr.wr_id = (unsigned long)agent_send_wr;
+
+       pci_unmap_addr_set(agent_send_wr, mapping, gather_list.addr);
+
+       /* Send */
+       spin_lock_irqsave(&port_priv->send_list_lock, flags);
+       if (ib_post_send_mad(mad_agent, &send_wr, &bad_send_wr)) {
+               spin_unlock_irqrestore(&port_priv->send_list_lock, flags);
+               dma_unmap_single(mad_agent->device->dma_device,
+                                pci_unmap_addr(agent_send_wr, mapping),
+                                sizeof(mad_priv->mad),
+                                DMA_TO_DEVICE);
+               ib_destroy_ah(agent_send_wr->ah);
+               kfree(agent_send_wr);
+       } else {
+               list_add_tail(&agent_send_wr->send_list,
+                             &port_priv->send_posted_list);
+               spin_unlock_irqrestore(&port_priv->send_list_lock, flags);
+               ret = 0;
+       }
+
+out:
+       return ret;
+}
+
+int agent_send(struct ib_mad_private *mad,
+              struct ib_grh *grh,
+              struct ib_wc *wc,
+              struct ib_device *device,
+              int port_num)
+{
+       struct ib_agent_port_private *port_priv;
+       struct ib_mad_agent *mad_agent;
+
+       port_priv = ib_get_agent_port(device, port_num, NULL);
+       if (!port_priv) {
+               printk(KERN_DEBUG SPFX "agent_send %s port %d not open\n",
+                      device->name, port_num);
+               return 1;
+       }
+
+       /* Get mad agent based on mgmt_class in MAD */
+       switch (mad->mad.mad.mad_hdr.mgmt_class) {
+               case IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE:
+                       mad_agent = port_priv->dr_smp_agent;
+                       break;
+               case IB_MGMT_CLASS_SUBN_LID_ROUTED:
+                       mad_agent = port_priv->lr_smp_agent;
+                       break;
+               case IB_MGMT_CLASS_PERF_MGMT:
+                       mad_agent = port_priv->perf_mgmt_agent;
+                       break;
+               default:
+                       return 1;
+       }
+
+       return agent_mad_send(mad_agent, port_priv, mad, grh, wc);
+}
+
+static void agent_send_handler(struct ib_mad_agent *mad_agent,
+                              struct ib_mad_send_wc *mad_send_wc)
+{
+       struct ib_agent_port_private    *port_priv;
+       struct ib_agent_send_wr         *agent_send_wr;
+       unsigned long                   flags;
+
+       /* Find matching MAD agent */
+       port_priv = ib_get_agent_port(NULL, 0, mad_agent);
+       if (!port_priv) {
+               printk(KERN_ERR SPFX "agent_send_handler: no matching MAD "
+                      "agent %p\n", mad_agent);
+               return;
+       }
+
+       agent_send_wr = (struct ib_agent_send_wr *)(unsigned long)mad_send_wc->wr_id;
+       spin_lock_irqsave(&port_priv->send_list_lock, flags);
+       /* Remove completed send from posted send MAD list */
+       list_del(&agent_send_wr->send_list);
+       spin_unlock_irqrestore(&port_priv->send_list_lock, flags);
+
+       /* Unmap PCI */
+       dma_unmap_single(mad_agent->device->dma_device,
+                        pci_unmap_addr(agent_send_wr, mapping),
+                        sizeof(agent_send_wr->mad->mad),
+                        DMA_TO_DEVICE);
+
+       ib_destroy_ah(agent_send_wr->ah);
+
+       /* Release allocated memory */
+       kmem_cache_free(ib_mad_cache, agent_send_wr->mad);
+       kfree(agent_send_wr);
+}
+
+int ib_agent_port_open(struct ib_device *device, int port_num)
+{
+       int ret;
+       struct ib_agent_port_private *port_priv;
+       struct ib_mad_reg_req reg_req;
+       unsigned long flags;
+
+       /* First, check if port already open for SMI */
+       port_priv = ib_get_agent_port(device, port_num, NULL);
+       if (port_priv) {
+               printk(KERN_DEBUG SPFX "%s port %d already open\n",
+                      device->name, port_num);
+               return 0;
+       }
+
+       /* Create new device info */
+       port_priv = kmalloc(sizeof *port_priv, GFP_KERNEL);
+       if (!port_priv) {
+               printk(KERN_ERR SPFX "No memory for ib_agent_port_private\n");
+               ret = -ENOMEM;
+               goto error1;
+       }
+
+       memset(port_priv, 0, sizeof *port_priv);
+       port_priv->port_num = port_num;
+       spin_lock_init(&port_priv->send_list_lock);
+       INIT_LIST_HEAD(&port_priv->send_posted_list);
+
+       /* Obtain MAD agent for directed route SM class */
+       reg_req.mgmt_class = IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE;
+       reg_req.mgmt_class_version = 1;
+
+       port_priv->dr_smp_agent = ib_register_mad_agent(device, port_num,
+                                                       IB_QPT_SMI,
+                                                       NULL, 0,
+                                                      &agent_send_handler,
+                                                       NULL, NULL);
+
+       if (IS_ERR(port_priv->dr_smp_agent)) {
+               ret = PTR_ERR(port_priv->dr_smp_agent);
+               goto error2;
+       }
+
+       /* Obtain MAD agent for LID routed SM class */
+       reg_req.mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED;
+       port_priv->lr_smp_agent = ib_register_mad_agent(device, port_num,
+                                                       IB_QPT_SMI,
+                                                       NULL, 0,
+                                                      &agent_send_handler,
+                                                       NULL, NULL);
+       if (IS_ERR(port_priv->lr_smp_agent)) {
+               ret = PTR_ERR(port_priv->lr_smp_agent);
+               goto error3;
+       }
+
+       /* Obtain MAD agent for PerfMgmt class */
+       reg_req.mgmt_class = IB_MGMT_CLASS_PERF_MGMT;
+       port_priv->perf_mgmt_agent = ib_register_mad_agent(device, port_num,
+                                                          IB_QPT_GSI,
+                                                          NULL, 0,
+                                                         &agent_send_handler,
+                                                          NULL, NULL);
+       if (IS_ERR(port_priv->perf_mgmt_agent)) {
+               ret = PTR_ERR(port_priv->perf_mgmt_agent);
+               goto error4;
+       }
+
+       port_priv->mr = ib_get_dma_mr(port_priv->dr_smp_agent->qp->pd,
+                                     IB_ACCESS_LOCAL_WRITE);
+       if (IS_ERR(port_priv->mr)) {
+               printk(KERN_ERR SPFX "Couldn't get DMA MR\n");
+               ret = PTR_ERR(port_priv->mr);
+               goto error5;
+       }
+
+       spin_lock_irqsave(&ib_agent_port_list_lock, flags);
+       list_add_tail(&port_priv->port_list, &ib_agent_port_list);
+       spin_unlock_irqrestore(&ib_agent_port_list_lock, flags);
+
+       return 0;
+
+error5:
+       ib_unregister_mad_agent(port_priv->perf_mgmt_agent);
+error4:
+       ib_unregister_mad_agent(port_priv->lr_smp_agent);
+error3:
+       ib_unregister_mad_agent(port_priv->dr_smp_agent);
+error2:
+       kfree(port_priv);
+error1:
+       return ret;
+}
+
+int ib_agent_port_close(struct ib_device *device, int port_num)
+{
+       struct ib_agent_port_private *port_priv;
+       unsigned long flags;
+
+       spin_lock_irqsave(&ib_agent_port_list_lock, flags);
+       port_priv = __ib_get_agent_port(device, port_num, NULL);
+       if (port_priv == NULL) {
+               spin_unlock_irqrestore(&ib_agent_port_list_lock, flags);
+               printk(KERN_ERR SPFX "Port %d not found\n", port_num);
+               return -ENODEV;
+       }
+       list_del(&port_priv->port_list);
+       spin_unlock_irqrestore(&ib_agent_port_list_lock, flags);
+
+       ib_dereg_mr(port_priv->mr);
+
+       ib_unregister_mad_agent(port_priv->perf_mgmt_agent);
+       ib_unregister_mad_agent(port_priv->lr_smp_agent);
+       ib_unregister_mad_agent(port_priv->dr_smp_agent);
+       kfree(port_priv);
+
+       return 0;
+}