/* * Copyright (c) 2006 Chelsio, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include #include #include #include "cxgb3_offload.h" #include "iwch_provider.h" #include #include "iwch.h" #include "iwch_cm.h" #define DRV_VERSION "1.1" MODULE_AUTHOR("Boyd Faulkner, Steve Wise"); MODULE_DESCRIPTION("Chelsio T3 RDMA Driver"); MODULE_LICENSE("Dual BSD/GPL"); static void open_rnic_dev(struct t3cdev *); static void close_rnic_dev(struct t3cdev *); static void iwch_event_handler(struct t3cdev *, u32, u32); struct cxgb3_client t3c_client = { .name = "iw_cxgb3", .add = open_rnic_dev, .remove = close_rnic_dev, .handlers = t3c_handlers, .redirect = iwch_ep_redirect, .event_handler = iwch_event_handler }; static LIST_HEAD(dev_list); static DEFINE_MUTEX(dev_mutex); static void disable_dbs(struct iwch_dev *rnicp) { unsigned long index; struct iwch_qp *qhp; xa_lock_irq(&rnicp->qps); xa_for_each(&rnicp->qps, index, qhp) cxio_disable_wq_db(&qhp->wq); xa_unlock_irq(&rnicp->qps); } static void enable_dbs(struct iwch_dev *rnicp, int ring_db) { unsigned long index; struct iwch_qp *qhp; xa_lock_irq(&rnicp->qps); xa_for_each(&rnicp->qps, index, qhp) { if (ring_db) ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid); cxio_enable_wq_db(&qhp->wq); } xa_unlock_irq(&rnicp->qps); } static void iwch_db_drop_task(struct work_struct *work) { struct iwch_dev *rnicp = container_of(work, struct iwch_dev, db_drop_task.work); enable_dbs(rnicp, 1); } static void rnic_init(struct iwch_dev *rnicp) { pr_debug("%s iwch_dev %p\n", __func__, rnicp); xa_init_flags(&rnicp->cqs, XA_FLAGS_LOCK_IRQ); xa_init_flags(&rnicp->qps, XA_FLAGS_LOCK_IRQ); xa_init_flags(&rnicp->mrs, XA_FLAGS_LOCK_IRQ); INIT_DELAYED_WORK(&rnicp->db_drop_task, iwch_db_drop_task); rnicp->attr.max_qps = T3_MAX_NUM_QP - 32; rnicp->attr.max_wrs = T3_MAX_QP_DEPTH; rnicp->attr.max_sge_per_wr = T3_MAX_SGE; rnicp->attr.max_sge_per_rdma_write_wr = T3_MAX_SGE; rnicp->attr.max_cqs = T3_MAX_NUM_CQ - 1; rnicp->attr.max_cqes_per_cq = T3_MAX_CQ_DEPTH; rnicp->attr.max_mem_regs = cxio_num_stags(&rnicp->rdev); rnicp->attr.max_phys_buf_entries = T3_MAX_PBL_SIZE; rnicp->attr.max_pds = T3_MAX_NUM_PD - 1; rnicp->attr.mem_pgsizes_bitmask = T3_PAGESIZE_MASK; rnicp->attr.max_mr_size = T3_MAX_MR_SIZE; rnicp->attr.can_resize_wq = 0; rnicp->attr.max_rdma_reads_per_qp = 8; rnicp->attr.max_rdma_read_resources = rnicp->attr.max_rdma_reads_per_qp * rnicp->attr.max_qps; rnicp->attr.max_rdma_read_qp_depth = 8; /* IRD */ rnicp->attr.max_rdma_read_depth = rnicp->attr.max_rdma_read_qp_depth * rnicp->attr.max_qps; rnicp->attr.rq_overflow_handled = 0; rnicp->attr.can_modify_ird = 0; rnicp->attr.can_modify_ord = 0; rnicp->attr.max_mem_windows = rnicp->attr.max_mem_regs - 1; rnicp->attr.stag0_value = 1; rnicp->attr.zbva_support = 1; rnicp->attr.local_invalidate_fence = 1; rnicp->attr.cq_overflow_detection = 1; return; } static void open_rnic_dev(struct t3cdev *tdev) { struct iwch_dev *rnicp; pr_debug("%s t3cdev %p\n", __func__, tdev); pr_info_once("Chelsio T3 RDMA Driver - version %s\n", DRV_VERSION); rnicp = ib_alloc_device(iwch_dev, ibdev); if (!rnicp) { pr_err("Cannot allocate ib device\n"); return; } rnicp->rdev.ulp = rnicp; rnicp->rdev.t3cdev_p = tdev; mutex_lock(&dev_mutex); if (cxio_rdev_open(&rnicp->rdev)) { mutex_unlock(&dev_mutex); pr_err("Unable to open CXIO rdev\n"); ib_dealloc_device(&rnicp->ibdev); return; } rnic_init(rnicp); list_add_tail(&rnicp->entry, &dev_list); mutex_unlock(&dev_mutex); if (iwch_register_device(rnicp)) { pr_err("Unable to register device\n"); close_rnic_dev(tdev); } pr_info("Initialized device %s\n", pci_name(rnicp->rdev.rnic_info.pdev)); return; } static void close_rnic_dev(struct t3cdev *tdev) { struct iwch_dev *dev, *tmp; pr_debug("%s t3cdev %p\n", __func__, tdev); mutex_lock(&dev_mutex); list_for_each_entry_safe(dev, tmp, &dev_list, entry) { if (dev->rdev.t3cdev_p == tdev) { dev->rdev.flags = CXIO_ERROR_FATAL; synchronize_net(); cancel_delayed_work_sync(&dev->db_drop_task); list_del(&dev->entry); iwch_unregister_device(dev); cxio_rdev_close(&dev->rdev); WARN_ON(!xa_empty(&dev->cqs)); WARN_ON(!xa_empty(&dev->qps)); WARN_ON(!xa_empty(&dev->mrs)); ib_dealloc_device(&dev->ibdev); break; } } mutex_unlock(&dev_mutex); } static void iwch_event_handler(struct t3cdev *tdev, u32 evt, u32 port_id) { struct cxio_rdev *rdev = tdev->ulp; struct iwch_dev *rnicp; struct ib_event event; u32 portnum = port_id + 1; int dispatch = 0; if (!rdev) return; rnicp = rdev_to_iwch_dev(rdev); switch (evt) { case OFFLOAD_STATUS_DOWN: { rdev->flags = CXIO_ERROR_FATAL; synchronize_net(); event.event = IB_EVENT_DEVICE_FATAL; dispatch = 1; break; } case OFFLOAD_PORT_DOWN: { event.event = IB_EVENT_PORT_ERR; dispatch = 1; break; } case OFFLOAD_PORT_UP: { event.event = IB_EVENT_PORT_ACTIVE; dispatch = 1; break; } case OFFLOAD_DB_FULL: { disable_dbs(rnicp); break; } case OFFLOAD_DB_EMPTY: { enable_dbs(rnicp, 1); break; } case OFFLOAD_DB_DROP: { unsigned long delay = 1000; unsigned short r; disable_dbs(rnicp); get_random_bytes(&r, 2); delay += r & 1023; /* * delay is between 1000-2023 usecs. */ schedule_delayed_work(&rnicp->db_drop_task, usecs_to_jiffies(delay)); break; } } if (dispatch) { event.device = &rnicp->ibdev; event.element.port_num = portnum; ib_dispatch_event(&event); } return; } static int __init iwch_init_module(void) { int err; err = cxio_hal_init(); if (err) return err; err = iwch_cm_init(); if (err) return err; cxio_register_ev_cb(iwch_ev_dispatch); cxgb3_register_client(&t3c_client); return 0; } static void __exit iwch_exit_module(void) { cxgb3_unregister_client(&t3c_client); cxio_unregister_ev_cb(iwch_ev_dispatch); iwch_cm_term(); cxio_hal_exit(); } module_init(iwch_init_module); module_exit(iwch_exit_module);