/* * Engenio/LSI RDAC DM HW handler * * Copyright (C) 2005 Mike Christie. All rights reserved. * Copyright (C) Chandra Seetharaman, IBM Corp. 2007 * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. * */ #include #include #include #define DM_MSG_PREFIX "multipath rdac" #include "dm.h" #include "dm-hw-handler.h" #define RDAC_DM_HWH_NAME "rdac" #define RDAC_DM_HWH_VER "0.4" /* * LSI mode page stuff * * These struct definitions and the forming of the * mode page were taken from the LSI RDAC 2.4 GPL'd * driver, and then converted to Linux conventions. */ #define RDAC_QUIESCENCE_TIME 20; /* * Page Codes */ #define RDAC_PAGE_CODE_REDUNDANT_CONTROLLER 0x2c /* * Controller modes definitions */ #define RDAC_MODE_TRANSFER_ALL_LUNS 0x01 #define RDAC_MODE_TRANSFER_SPECIFIED_LUNS 0x02 /* * RDAC Options field */ #define RDAC_FORCED_QUIESENCE 0x02 #define RDAC_FAILOVER_TIMEOUT (60 * HZ) struct rdac_mode_6_hdr { u8 data_len; u8 medium_type; u8 device_params; u8 block_desc_len; }; struct rdac_mode_10_hdr { u16 data_len; u8 medium_type; u8 device_params; u16 reserved; u16 block_desc_len; }; struct rdac_mode_common { u8 controller_serial[16]; u8 alt_controller_serial[16]; u8 rdac_mode[2]; u8 alt_rdac_mode[2]; u8 quiescence_timeout; u8 rdac_options; }; struct rdac_pg_legacy { struct rdac_mode_6_hdr hdr; u8 page_code; u8 page_len; struct rdac_mode_common common; #define MODE6_MAX_LUN 32 u8 lun_table[MODE6_MAX_LUN]; u8 reserved2[32]; u8 reserved3; u8 reserved4; }; struct rdac_pg_expanded { struct rdac_mode_10_hdr hdr; u8 page_code; u8 subpage_code; u8 page_len[2]; struct rdac_mode_common common; u8 lun_table[256]; u8 reserved3; u8 reserved4; }; struct c9_inquiry { u8 peripheral_info; u8 page_code; /* 0xC9 */ u8 reserved1; u8 page_len; u8 page_id[4]; /* "vace" */ u8 avte_cvp; u8 path_prio; u8 reserved2[38]; }; #define SUBSYS_ID_LEN 16 #define SLOT_ID_LEN 2 struct c4_inquiry { u8 peripheral_info; u8 page_code; /* 0xC4 */ u8 reserved1; u8 page_len; u8 page_id[4]; /* "subs" */ u8 subsys_id[SUBSYS_ID_LEN]; u8 revision[4]; u8 slot_id[SLOT_ID_LEN]; u8 reserved[2]; }; struct rdac_controller { u8 subsys_id[SUBSYS_ID_LEN]; u8 slot_id[SLOT_ID_LEN]; int use_10_ms; struct kref kref; struct list_head node; /* list of all controllers */ spinlock_t lock; int submitted; struct list_head cmd_list; /* list of commands to be submitted */ union { struct rdac_pg_legacy legacy; struct rdac_pg_expanded expanded; } mode_select; }; struct c8_inquiry { u8 peripheral_info; u8 page_code; /* 0xC8 */ u8 reserved1; u8 page_len; u8 page_id[4]; /* "edid" */ u8 reserved2[3]; u8 vol_uniq_id_len; u8 vol_uniq_id[16]; u8 vol_user_label_len; u8 vol_user_label[60]; u8 array_uniq_id_len; u8 array_unique_id[16]; u8 array_user_label_len; u8 array_user_label[60]; u8 lun[8]; }; struct c2_inquiry { u8 peripheral_info; u8 page_code; /* 0xC2 */ u8 reserved1; u8 page_len; u8 page_id[4]; /* "swr4" */ u8 sw_version[3]; u8 sw_date[3]; u8 features_enabled; u8 max_lun_supported; u8 partitions[239]; /* Total allocation length should be 0xFF */ }; struct rdac_handler { struct list_head entry; /* list waiting to submit MODE SELECT */ unsigned timeout; struct rdac_controller *ctlr; #define UNINITIALIZED_LUN (1 << 8) unsigned lun; unsigned char sense[SCSI_SENSE_BUFFERSIZE]; struct dm_path *path; struct work_struct work; #define SEND_C2_INQUIRY 1 #define SEND_C4_INQUIRY 2 #define SEND_C8_INQUIRY 3 #define SEND_C9_INQUIRY 4 #define SEND_MODE_SELECT 5 int cmd_to_send; union { struct c2_inquiry c2; struct c4_inquiry c4; struct c8_inquiry c8; struct c9_inquiry c9; } inq; }; static LIST_HEAD(ctlr_list); static DEFINE_SPINLOCK(list_lock); static struct workqueue_struct *rdac_wkqd; static inline int had_failures(struct request *req, int error) { return (error || host_byte(req->errors) != DID_OK || msg_byte(req->errors) != COMMAND_COMPLETE); } static void rdac_resubmit_all(struct rdac_handler *h) { struct rdac_controller *ctlr = h->ctlr; struct rdac_handler *tmp, *h1; spin_lock(&ctlr->lock); list_for_each_entry_safe(h1, tmp, &ctlr->cmd_list, entry) { h1->cmd_to_send = SEND_C9_INQUIRY; queue_work(rdac_wkqd, &h1->work); list_del(&h1->entry); } ctlr->submitted = 0; spin_unlock(&ctlr->lock); } static void mode_select_endio(struct request *req, int error) { struct rdac_handler *h = req->end_io_data; struct scsi_sense_hdr sense_hdr; int sense = 0, fail = 0; if (had_failures(req, error)) { fail = 1; goto failed; } if (status_byte(req->errors) == CHECK_CONDITION) { scsi_normalize_sense(req->sense, SCSI_SENSE_BUFFERSIZE, &sense_hdr); sense = (sense_hdr.sense_key << 16) | (sense_hdr.asc << 8) | sense_hdr.ascq; /* If it is retryable failure, submit the c9 inquiry again */ if (sense == 0x59136 || sense == 0x68b02 || sense == 0xb8b02 || sense == 0x62900) { /* 0x59136 - Command lock contention * 0x[6b]8b02 - Quiesense in progress or achieved * 0x62900 - Power On, Reset, or Bus Device Reset */ h->cmd_to_send = SEND_C9_INQUIRY; queue_work(rdac_wkqd, &h->work); goto done; } if (sense) DMINFO("MODE_SELECT failed on %s with sense 0x%x", h->path->dev->name, sense); } failed: if (fail || sense) dm_pg_init_complete(h->path, MP_FAIL_PATH); else dm_pg_init_complete(h->path, 0); done: rdac_resubmit_all(h); __blk_put_request(req->q, req); } static struct request *get_rdac_req(struct rdac_handler *h, void *buffer, unsigned buflen, int rw) { struct request *rq; struct request_queue *q = bdev_get_queue(h->path->dev->bdev); rq = blk_get_request(q, rw, GFP_KERNEL); if (!rq) { DMINFO("get_rdac_req: blk_get_request failed"); return NULL; } if (buflen && blk_rq_map_kern(q, rq, buffer, buflen, GFP_KERNEL)) { blk_put_request(rq); DMINFO("get_rdac_req: blk_rq_map_kern failed"); return NULL; } memset(&rq->cmd, 0, BLK_MAX_CDB); rq->sense = h->sense; memset(rq->sense, 0, SCSI_SENSE_BUFFERSIZE); rq->sense_len = 0; rq->end_io_data = h; rq->timeout = h->timeout; rq->cmd_type = REQ_TYPE_BLOCK_PC; rq->cmd_flags |= REQ_FAILFAST | REQ_NOMERGE; return rq; } static struct request *rdac_failover_get(struct rdac_handler *h) { struct request *rq; struct rdac_mode_common *common; unsigned data_size; if (h->ctlr->use_10_ms) { struct rdac_pg_expanded *rdac_pg; data_size = sizeof(struct rdac_pg_expanded); rdac_pg = &h->ctlr->mode_select.expanded; memset(rdac_pg, 0, data_size); common = &rdac_pg->common; rdac_pg->page_code = RDAC_PAGE_CODE_REDUNDANT_CONTROLLER + 0x40; rdac_pg->subpage_code = 0x1; rdac_pg->page_len[0] = 0x01; rdac_pg->page_len[1] = 0x28; rdac_pg->lun_table[h->lun] = 0x81; } else { struct rdac_pg_legacy *rdac_pg; data_size = sizeof(struct rdac_pg_legacy); rdac_pg = &h->ctlr->mode_select.legacy; memset(rdac_pg, 0, data_size); common = &rdac_pg->common; rdac_pg->page_code = RDAC_PAGE_CODE_REDUNDANT_CONTROLLER; rdac_pg->page_len = 0x68; rdac_pg->lun_table[h->lun] = 0x81; } common->rdac_mode[1] = RDAC_MODE_TRANSFER_SPECIFIED_LUNS; common->quiescence_timeout = RDAC_QUIESCENCE_TIME; common->rdac_options = RDAC_FORCED_QUIESENCE; /* get request for block layer packet command */ rq = get_rdac_req(h, &h->ctlr->mode_select, data_size, WRITE); if (!rq) { DMERR("rdac_failover_get: no rq"); return NULL; } /* Prepare the command. */ if (h->ctlr->use_10_ms) { rq->cmd[0] = MODE_SELECT_10; rq->cmd[7] = data_size >> 8; rq->cmd[8] = data_size & 0xff; } else { rq->cmd[0] = MODE_SELECT; rq->cmd[4] = data_size; } rq->cmd_len = COMMAND_SIZE(rq->cmd[0]); return rq; } /* Acquires h->ctlr->lock */ static void submit_mode_select(struct rdac_handler *h) { struct request *rq; struct request_queue *q = bdev_get_queue(h->path->dev->bdev); spin_lock(&h->ctlr->lock); if (h->ctlr->submitted) { list_add(&h->entry, &h->ctlr->cmd_list); goto drop_lock; } if (!q) { DMINFO("submit_mode_select: no queue"); goto fail_path; } rq = rdac_failover_get(h); if (!rq) { DMERR("submit_mode_select: no rq"); goto fail_path; } DMINFO("queueing MODE_SELECT command on %s", h->path->dev->name); blk_execute_rq_nowait(q, NULL, rq, 1, mode_select_endio); h->ctlr->submitted = 1; goto drop_lock; fail_path: dm_pg_init_complete(h->path, MP_FAIL_PATH); drop_lock: spin_unlock(&h->ctlr->lock); } static void release_ctlr(struct kref *kref) { struct rdac_controller *ctlr; ctlr = container_of(kref, struct rdac_controller, kref); spin_lock(&list_lock); list_del(&ctlr->node); spin_unlock(&list_lock); kfree(ctlr); } static struct rdac_controller *get_controller(u8 *subsys_id, u8 *slot_id) { struct rdac_controller *ctlr, *tmp; spin_lock(&list_lock); list_for_each_entry(tmp, &ctlr_list, node) { if ((memcmp(tmp->subsys_id, subsys_id, SUBSYS_ID_LEN) == 0) && (memcmp(tmp->slot_id, slot_id, SLOT_ID_LEN) == 0)) { kref_get(&tmp->kref); spin_unlock(&list_lock); return tmp; } } ctlr = kmalloc(sizeof(*ctlr), GFP_ATOMIC); if (!ctlr) goto done; /* initialize fields of controller */ memcpy(ctlr->subsys_id, subsys_id, SUBSYS_ID_LEN); memcpy(ctlr->slot_id, slot_id, SLOT_ID_LEN); kref_init(&ctlr->kref); spin_lock_init(&ctlr->lock); ctlr->submitted = 0; ctlr->use_10_ms = -1; INIT_LIST_HEAD(&ctlr->cmd_list); list_add(&ctlr->node, &ctlr_list); done: spin_unlock(&list_lock); return ctlr; } static void c4_endio(struct request *req, int error) { struct rdac_handler *h = req->end_io_data; struct c4_inquiry *sp; if (had_failures(req, error)) { dm_pg_init_complete(h->path, MP_FAIL_PATH); goto done; } sp = &h->inq.c4; h->ctlr = get_controller(sp->subsys_id, sp->slot_id); if (h->ctlr) { h->cmd_to_send = SEND_C9_INQUIRY; queue_work(rdac_wkqd, &h->work); } else dm_pg_init_complete(h->path, MP_FAIL_PATH); done: __blk_put_request(req->q, req); } static void c2_endio(struct request *req, int error) { struct rdac_handler *h = req->end_io_data; struct c2_inquiry *sp; if (had_failures(req, error)) { dm_pg_init_complete(h->path, MP_FAIL_PATH); goto done; } sp = &h->inq.c2; /* If more than MODE6_MAX_LUN luns are supported, use mode select 10 */ if (sp->max_lun_supported >= MODE6_MAX_LUN) h->ctlr->use_10_ms = 1; else h->ctlr->use_10_ms = 0; h->cmd_to_send = SEND_MODE_SELECT; queue_work(rdac_wkqd, &h->work); done: __blk_put_request(req->q, req); } static void c9_endio(struct request *req, int error) { struct rdac_handler *h = req->end_io_data; struct c9_inquiry *sp; if (had_failures(req, error)) { dm_pg_init_complete(h->path, MP_FAIL_PATH); goto done; } /* We need to look at the sense keys here to take clear action. * For now simple logic: If the host is in AVT mode or if controller * owns the lun, return dm_pg_init_complete(), otherwise submit * MODE SELECT. */ sp = &h->inq.c9; /* If in AVT mode, return success */ if ((sp->avte_cvp >> 7) == 0x1) { dm_pg_init_complete(h->path, 0); goto done; } /* If the controller on this path owns the LUN, return success */ if (sp->avte_cvp & 0x1) { dm_pg_init_complete(h->path, 0); goto done; } if (h->ctlr) { if (h->ctlr->use_10_ms == -1) h->cmd_to_send = SEND_C2_INQUIRY; else h->cmd_to_send = SEND_MODE_SELECT; } else h->cmd_to_send = SEND_C4_INQUIRY; queue_work(rdac_wkqd, &h->work); done: __blk_put_request(req->q, req); } static void c8_endio(struct request *req, int error) { struct rdac_handler *h = req->end_io_data; struct c8_inquiry *sp; if (had_failures(req, error)) { dm_pg_init_complete(h->path, MP_FAIL_PATH); goto done; } /* We need to look at the sense keys here to take clear action. * For now simple logic: Get the lun from the inquiry page. */ sp = &h->inq.c8; h->lun = sp->lun[7]; /* currently it uses only one byte */ h->cmd_to_send = SEND_C9_INQUIRY; queue_work(rdac_wkqd, &h->work); done: __blk_put_request(req->q, req); } static void submit_inquiry(struct rdac_handler *h, int page_code, unsigned int len, rq_end_io_fn endio) { struct request *rq; struct request_queue *q = bdev_get_queue(h->path->dev->bdev); if (!q) goto fail_path; rq = get_rdac_req(h, &h->inq, len, READ); if (!rq) goto fail_path; /* Prepare the command. */ rq->cmd[0] = INQUIRY; rq->cmd[1] = 1; rq->cmd[2] = page_code; rq->cmd[4] = len; rq->cmd_len = COMMAND_SIZE(INQUIRY); blk_execute_rq_nowait(q, NULL, rq, 1, endio); return; fail_path: dm_pg_init_complete(h->path, MP_FAIL_PATH); } static void service_wkq(struct work_struct *work) { struct rdac_handler *h = container_of(work, struct rdac_handler, work); switch (h->cmd_to_send) { case SEND_C2_INQUIRY: submit_inquiry(h, 0xC2, sizeof(struct c2_inquiry), c2_endio); break; case SEND_C4_INQUIRY: submit_inquiry(h, 0xC4, sizeof(struct c4_inquiry), c4_endio); break; case SEND_C8_INQUIRY: submit_inquiry(h, 0xC8, sizeof(struct c8_inquiry), c8_endio); break; case SEND_C9_INQUIRY: submit_inquiry(h, 0xC9, sizeof(struct c9_inquiry), c9_endio); break; case SEND_MODE_SELECT: submit_mode_select(h); break; default: BUG(); } } /* * only support subpage2c until we confirm that this is just a matter of * of updating firmware or not, and RDAC (basic AVT works already) for now * but we can add these in in when we get time and testers */ static int rdac_create(struct hw_handler *hwh, unsigned argc, char **argv) { struct rdac_handler *h; unsigned timeout; if (argc == 0) { /* No arguments: use defaults */ timeout = RDAC_FAILOVER_TIMEOUT; } else if (argc != 1) { DMWARN("incorrect number of arguments"); return -EINVAL; } else { if (sscanf(argv[1], "%u", &timeout) != 1) { DMWARN("invalid timeout value"); return -EINVAL; } } h = kzalloc(sizeof(*h), GFP_KERNEL); if (!h) return -ENOMEM; hwh->context = h; h->timeout = timeout; h->lun = UNINITIALIZED_LUN; INIT_WORK(&h->work, service_wkq); DMWARN("using RDAC command with timeout %u", h->timeout); return 0; } static void rdac_destroy(struct hw_handler *hwh) { struct rdac_handler *h = hwh->context; if (h->ctlr) kref_put(&h->ctlr->kref, release_ctlr); kfree(h); hwh->context = NULL; } static unsigned rdac_error(struct hw_handler *hwh, struct bio *bio) { /* Try default handler */ return dm_scsi_err_handler(hwh, bio); } static void rdac_pg_init(struct hw_handler *hwh, unsigned bypassed, struct dm_path *path) { struct rdac_handler *h = hwh->context; h->path = path; switch (h->lun) { case UNINITIALIZED_LUN: submit_inquiry(h, 0xC8, sizeof(struct c8_inquiry), c8_endio); break; default: submit_inquiry(h, 0xC9, sizeof(struct c9_inquiry), c9_endio); } } static struct hw_handler_type rdac_handler = { .name = RDAC_DM_HWH_NAME, .module = THIS_MODULE, .create = rdac_create, .destroy = rdac_destroy, .pg_init = rdac_pg_init, .error = rdac_error, }; static int __init rdac_init(void) { int r; rdac_wkqd = create_singlethread_workqueue("rdac_wkqd"); if (!rdac_wkqd) { DMERR("Failed to create workqueue rdac_wkqd."); return -ENOMEM; } r = dm_register_hw_handler(&rdac_handler); if (r < 0) { DMERR("%s: register failed %d", RDAC_DM_HWH_NAME, r); destroy_workqueue(rdac_wkqd); return r; } DMINFO("%s: version %s loaded", RDAC_DM_HWH_NAME, RDAC_DM_HWH_VER); return 0; } static void __exit rdac_exit(void) { int r = dm_unregister_hw_handler(&rdac_handler); destroy_workqueue(rdac_wkqd); if (r < 0) DMERR("%s: unregister failed %d", RDAC_DM_HWH_NAME, r); } module_init(rdac_init); module_exit(rdac_exit); MODULE_DESCRIPTION("DM Multipath LSI/Engenio RDAC support"); MODULE_AUTHOR("Mike Christie, Chandra Seetharaman"); MODULE_LICENSE("GPL"); MODULE_VERSION(RDAC_DM_HWH_VER);