// SPDX-License-Identifier: GPL-2.0+ /* * GPIO driver for virtio-based virtual GPIO controllers * * Copyright (C) 2021 metux IT consult * Enrico Weigelt, metux IT consult * * Copyright (C) 2021 Linaro. * Viresh Kumar */ #include #include #include #include #include #include #include #include #include #include #include struct virtio_gpio_line { struct mutex lock; /* Protects line operation */ struct completion completion; struct virtio_gpio_request req ____cacheline_aligned; struct virtio_gpio_response res ____cacheline_aligned; unsigned int rxlen; }; struct vgpio_irq_line { u8 type; bool disabled; bool masked; bool queued; bool update_pending; bool queue_pending; struct virtio_gpio_irq_request ireq ____cacheline_aligned; struct virtio_gpio_irq_response ires ____cacheline_aligned; }; struct virtio_gpio { struct virtio_device *vdev; struct mutex lock; /* Protects virtqueue operation */ struct gpio_chip gc; struct virtio_gpio_line *lines; struct virtqueue *request_vq; /* irq support */ struct virtqueue *event_vq; struct mutex irq_lock; /* Protects irq operation */ raw_spinlock_t eventq_lock; /* Protects queuing of the buffer */ struct vgpio_irq_line *irq_lines; }; static int _virtio_gpio_req(struct virtio_gpio *vgpio, u16 type, u16 gpio, u8 txvalue, u8 *rxvalue, void *response, u32 rxlen) { struct virtio_gpio_line *line = &vgpio->lines[gpio]; struct virtio_gpio_request *req = &line->req; struct virtio_gpio_response *res = response; struct scatterlist *sgs[2], req_sg, res_sg; struct device *dev = &vgpio->vdev->dev; int ret; /* * Prevent concurrent requests for the same line since we have * pre-allocated request/response buffers for each GPIO line. Moreover * Linux always accesses a GPIO line sequentially, so this locking shall * always go through without any delays. */ mutex_lock(&line->lock); req->type = cpu_to_le16(type); req->gpio = cpu_to_le16(gpio); req->value = cpu_to_le32(txvalue); sg_init_one(&req_sg, req, sizeof(*req)); sg_init_one(&res_sg, res, rxlen); sgs[0] = &req_sg; sgs[1] = &res_sg; line->rxlen = 0; reinit_completion(&line->completion); /* * Virtqueue callers need to ensure they don't call its APIs with other * virtqueue operations at the same time. */ mutex_lock(&vgpio->lock); ret = virtqueue_add_sgs(vgpio->request_vq, sgs, 1, 1, line, GFP_KERNEL); if (ret) { dev_err(dev, "failed to add request to vq\n"); mutex_unlock(&vgpio->lock); goto out; } virtqueue_kick(vgpio->request_vq); mutex_unlock(&vgpio->lock); wait_for_completion(&line->completion); if (unlikely(res->status != VIRTIO_GPIO_STATUS_OK)) { dev_err(dev, "GPIO request failed: %d\n", gpio); ret = -EINVAL; goto out; } if (unlikely(line->rxlen != rxlen)) { dev_err(dev, "GPIO operation returned incorrect len (%u : %u)\n", rxlen, line->rxlen); ret = -EINVAL; goto out; } if (rxvalue) *rxvalue = res->value; out: mutex_unlock(&line->lock); return ret; } static int virtio_gpio_req(struct virtio_gpio *vgpio, u16 type, u16 gpio, u8 txvalue, u8 *rxvalue) { struct virtio_gpio_line *line = &vgpio->lines[gpio]; struct virtio_gpio_response *res = &line->res; return _virtio_gpio_req(vgpio, type, gpio, txvalue, rxvalue, res, sizeof(*res)); } static void virtio_gpio_free(struct gpio_chip *gc, unsigned int gpio) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_SET_DIRECTION, gpio, VIRTIO_GPIO_DIRECTION_NONE, NULL); } static int virtio_gpio_get_direction(struct gpio_chip *gc, unsigned int gpio) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); u8 direction; int ret; ret = virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_GET_DIRECTION, gpio, 0, &direction); if (ret) return ret; switch (direction) { case VIRTIO_GPIO_DIRECTION_IN: return GPIO_LINE_DIRECTION_IN; case VIRTIO_GPIO_DIRECTION_OUT: return GPIO_LINE_DIRECTION_OUT; default: return -EINVAL; } } static int virtio_gpio_direction_input(struct gpio_chip *gc, unsigned int gpio) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); return virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_SET_DIRECTION, gpio, VIRTIO_GPIO_DIRECTION_IN, NULL); } static int virtio_gpio_direction_output(struct gpio_chip *gc, unsigned int gpio, int value) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); int ret; ret = virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_SET_VALUE, gpio, value, NULL); if (ret) return ret; return virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_SET_DIRECTION, gpio, VIRTIO_GPIO_DIRECTION_OUT, NULL); } static int virtio_gpio_get(struct gpio_chip *gc, unsigned int gpio) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); u8 value; int ret; ret = virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_GET_VALUE, gpio, 0, &value); return ret ? ret : value; } static void virtio_gpio_set(struct gpio_chip *gc, unsigned int gpio, int value) { struct virtio_gpio *vgpio = gpiochip_get_data(gc); virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_SET_VALUE, gpio, value, NULL); } /* Interrupt handling */ static void virtio_gpio_irq_prepare(struct virtio_gpio *vgpio, u16 gpio) { struct vgpio_irq_line *irq_line = &vgpio->irq_lines[gpio]; struct virtio_gpio_irq_request *ireq = &irq_line->ireq; struct virtio_gpio_irq_response *ires = &irq_line->ires; struct scatterlist *sgs[2], req_sg, res_sg; int ret; if (WARN_ON(irq_line->queued || irq_line->masked || irq_line->disabled)) return; ireq->gpio = cpu_to_le16(gpio); sg_init_one(&req_sg, ireq, sizeof(*ireq)); sg_init_one(&res_sg, ires, sizeof(*ires)); sgs[0] = &req_sg; sgs[1] = &res_sg; ret = virtqueue_add_sgs(vgpio->event_vq, sgs, 1, 1, irq_line, GFP_ATOMIC); if (ret) { dev_err(&vgpio->vdev->dev, "failed to add request to eventq\n"); return; } irq_line->queued = true; virtqueue_kick(vgpio->event_vq); } static void virtio_gpio_irq_enable(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; raw_spin_lock(&vgpio->eventq_lock); irq_line->disabled = false; irq_line->masked = false; irq_line->queue_pending = true; raw_spin_unlock(&vgpio->eventq_lock); irq_line->update_pending = true; } static void virtio_gpio_irq_disable(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; raw_spin_lock(&vgpio->eventq_lock); irq_line->disabled = true; irq_line->masked = true; irq_line->queue_pending = false; raw_spin_unlock(&vgpio->eventq_lock); irq_line->update_pending = true; } static void virtio_gpio_irq_mask(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; raw_spin_lock(&vgpio->eventq_lock); irq_line->masked = true; raw_spin_unlock(&vgpio->eventq_lock); } static void virtio_gpio_irq_unmask(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; raw_spin_lock(&vgpio->eventq_lock); irq_line->masked = false; /* Queue the buffer unconditionally on unmask */ virtio_gpio_irq_prepare(vgpio, d->hwirq); raw_spin_unlock(&vgpio->eventq_lock); } static int virtio_gpio_irq_set_type(struct irq_data *d, unsigned int type) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; switch (type) { case IRQ_TYPE_EDGE_RISING: type = VIRTIO_GPIO_IRQ_TYPE_EDGE_RISING; break; case IRQ_TYPE_EDGE_FALLING: type = VIRTIO_GPIO_IRQ_TYPE_EDGE_FALLING; break; case IRQ_TYPE_EDGE_BOTH: type = VIRTIO_GPIO_IRQ_TYPE_EDGE_BOTH; break; case IRQ_TYPE_LEVEL_LOW: type = VIRTIO_GPIO_IRQ_TYPE_LEVEL_LOW; break; case IRQ_TYPE_LEVEL_HIGH: type = VIRTIO_GPIO_IRQ_TYPE_LEVEL_HIGH; break; default: dev_err(&vgpio->vdev->dev, "unsupported irq type: %u\n", type); return -EINVAL; } irq_line->type = type; irq_line->update_pending = true; return 0; } static void virtio_gpio_irq_bus_lock(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); mutex_lock(&vgpio->irq_lock); } static void virtio_gpio_irq_bus_sync_unlock(struct irq_data *d) { struct gpio_chip *gc = irq_data_get_irq_chip_data(d); struct virtio_gpio *vgpio = gpiochip_get_data(gc); struct vgpio_irq_line *irq_line = &vgpio->irq_lines[d->hwirq]; u8 type = irq_line->disabled ? VIRTIO_GPIO_IRQ_TYPE_NONE : irq_line->type; unsigned long flags; if (irq_line->update_pending) { irq_line->update_pending = false; virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_IRQ_TYPE, d->hwirq, type, NULL); /* Queue the buffer only after interrupt is enabled */ raw_spin_lock_irqsave(&vgpio->eventq_lock, flags); if (irq_line->queue_pending) { irq_line->queue_pending = false; virtio_gpio_irq_prepare(vgpio, d->hwirq); } raw_spin_unlock_irqrestore(&vgpio->eventq_lock, flags); } mutex_unlock(&vgpio->irq_lock); } static struct irq_chip vgpio_irq_chip = { .name = "virtio-gpio", .irq_enable = virtio_gpio_irq_enable, .irq_disable = virtio_gpio_irq_disable, .irq_mask = virtio_gpio_irq_mask, .irq_unmask = virtio_gpio_irq_unmask, .irq_set_type = virtio_gpio_irq_set_type, /* These are required to implement irqchip for slow busses */ .irq_bus_lock = virtio_gpio_irq_bus_lock, .irq_bus_sync_unlock = virtio_gpio_irq_bus_sync_unlock, }; static bool ignore_irq(struct virtio_gpio *vgpio, int gpio, struct vgpio_irq_line *irq_line) { bool ignore = false; raw_spin_lock(&vgpio->eventq_lock); irq_line->queued = false; /* Interrupt is disabled currently */ if (irq_line->masked || irq_line->disabled) { ignore = true; goto unlock; } /* * Buffer is returned as the interrupt was disabled earlier, but is * enabled again now. Requeue the buffers. */ if (irq_line->ires.status == VIRTIO_GPIO_IRQ_STATUS_INVALID) { virtio_gpio_irq_prepare(vgpio, gpio); ignore = true; goto unlock; } if (WARN_ON(irq_line->ires.status != VIRTIO_GPIO_IRQ_STATUS_VALID)) ignore = true; unlock: raw_spin_unlock(&vgpio->eventq_lock); return ignore; } static void virtio_gpio_event_vq(struct virtqueue *vq) { struct virtio_gpio *vgpio = vq->vdev->priv; struct device *dev = &vgpio->vdev->dev; struct vgpio_irq_line *irq_line; int gpio, ret; unsigned int len; while (true) { irq_line = virtqueue_get_buf(vgpio->event_vq, &len); if (!irq_line) break; if (len != sizeof(irq_line->ires)) { dev_err(dev, "irq with incorrect length (%u : %u)\n", len, (unsigned int)sizeof(irq_line->ires)); continue; } /* * Find GPIO line number from the offset of irq_line within the * irq_lines block. We can also get GPIO number from * irq-request, but better not to rely on a buffer returned by * remote. */ gpio = irq_line - vgpio->irq_lines; WARN_ON(gpio >= vgpio->gc.ngpio); if (unlikely(ignore_irq(vgpio, gpio, irq_line))) continue; ret = generic_handle_domain_irq(vgpio->gc.irq.domain, gpio); if (ret) dev_err(dev, "failed to handle interrupt: %d\n", ret); } } static void virtio_gpio_request_vq(struct virtqueue *vq) { struct virtio_gpio_line *line; unsigned int len; do { line = virtqueue_get_buf(vq, &len); if (!line) return; line->rxlen = len; complete(&line->completion); } while (1); } static void virtio_gpio_free_vqs(struct virtio_device *vdev) { vdev->config->reset(vdev); vdev->config->del_vqs(vdev); } static int virtio_gpio_alloc_vqs(struct virtio_gpio *vgpio, struct virtio_device *vdev) { const char * const names[] = { "requestq", "eventq" }; vq_callback_t *cbs[] = { virtio_gpio_request_vq, virtio_gpio_event_vq, }; struct virtqueue *vqs[2] = { NULL, NULL }; int ret; ret = virtio_find_vqs(vdev, vgpio->irq_lines ? 2 : 1, vqs, cbs, names, NULL); if (ret) { dev_err(&vdev->dev, "failed to find vqs: %d\n", ret); return ret; } if (!vqs[0]) { dev_err(&vdev->dev, "failed to find requestq vq\n"); goto out; } vgpio->request_vq = vqs[0]; if (vgpio->irq_lines && !vqs[1]) { dev_err(&vdev->dev, "failed to find eventq vq\n"); goto out; } vgpio->event_vq = vqs[1]; return 0; out: if (vqs[0] || vqs[1]) virtio_gpio_free_vqs(vdev); return -ENODEV; } static const char **virtio_gpio_get_names(struct virtio_gpio *vgpio, u32 gpio_names_size, u16 ngpio) { struct virtio_gpio_response_get_names *res; struct device *dev = &vgpio->vdev->dev; u8 *gpio_names, *str; const char **names; int i, ret, len; if (!gpio_names_size) return NULL; len = sizeof(*res) + gpio_names_size; res = devm_kzalloc(dev, len, GFP_KERNEL); if (!res) return NULL; gpio_names = res->value; ret = _virtio_gpio_req(vgpio, VIRTIO_GPIO_MSG_GET_NAMES, 0, 0, NULL, res, len); if (ret) { dev_err(dev, "Failed to get GPIO names: %d\n", ret); return NULL; } names = devm_kcalloc(dev, ngpio, sizeof(*names), GFP_KERNEL); if (!names) return NULL; /* NULL terminate the string instead of checking it */ gpio_names[gpio_names_size - 1] = '\0'; for (i = 0, str = gpio_names; i < ngpio; i++) { names[i] = str; str += strlen(str) + 1; /* zero-length strings are allowed */ if (str > gpio_names + gpio_names_size) { dev_err(dev, "gpio_names block is too short (%d)\n", i); return NULL; } } return names; } static int virtio_gpio_probe(struct virtio_device *vdev) { struct virtio_gpio_config config; struct device *dev = &vdev->dev; struct virtio_gpio *vgpio; u32 gpio_names_size; u16 ngpio; int ret, i; vgpio = devm_kzalloc(dev, sizeof(*vgpio), GFP_KERNEL); if (!vgpio) return -ENOMEM; /* Read configuration */ virtio_cread_bytes(vdev, 0, &config, sizeof(config)); gpio_names_size = le32_to_cpu(config.gpio_names_size); ngpio = le16_to_cpu(config.ngpio); if (!ngpio) { dev_err(dev, "Number of GPIOs can't be zero\n"); return -EINVAL; } vgpio->lines = devm_kcalloc(dev, ngpio, sizeof(*vgpio->lines), GFP_KERNEL); if (!vgpio->lines) return -ENOMEM; for (i = 0; i < ngpio; i++) { mutex_init(&vgpio->lines[i].lock); init_completion(&vgpio->lines[i].completion); } mutex_init(&vgpio->lock); vdev->priv = vgpio; vgpio->vdev = vdev; vgpio->gc.free = virtio_gpio_free; vgpio->gc.get_direction = virtio_gpio_get_direction; vgpio->gc.direction_input = virtio_gpio_direction_input; vgpio->gc.direction_output = virtio_gpio_direction_output; vgpio->gc.get = virtio_gpio_get; vgpio->gc.set = virtio_gpio_set; vgpio->gc.ngpio = ngpio; vgpio->gc.base = -1; /* Allocate base dynamically */ vgpio->gc.label = dev_name(dev); vgpio->gc.parent = dev; vgpio->gc.owner = THIS_MODULE; vgpio->gc.can_sleep = true; /* Interrupt support */ if (virtio_has_feature(vdev, VIRTIO_GPIO_F_IRQ)) { vgpio->irq_lines = devm_kcalloc(dev, ngpio, sizeof(*vgpio->irq_lines), GFP_KERNEL); if (!vgpio->irq_lines) return -ENOMEM; /* The event comes from the outside so no parent handler */ vgpio->gc.irq.parent_handler = NULL; vgpio->gc.irq.num_parents = 0; vgpio->gc.irq.parents = NULL; vgpio->gc.irq.default_type = IRQ_TYPE_NONE; vgpio->gc.irq.handler = handle_level_irq; vgpio->gc.irq.chip = &vgpio_irq_chip; for (i = 0; i < ngpio; i++) { vgpio->irq_lines[i].type = VIRTIO_GPIO_IRQ_TYPE_NONE; vgpio->irq_lines[i].disabled = true; vgpio->irq_lines[i].masked = true; } mutex_init(&vgpio->irq_lock); raw_spin_lock_init(&vgpio->eventq_lock); } ret = virtio_gpio_alloc_vqs(vgpio, vdev); if (ret) return ret; /* Mark the device ready to perform operations from within probe() */ virtio_device_ready(vdev); vgpio->gc.names = virtio_gpio_get_names(vgpio, gpio_names_size, ngpio); ret = gpiochip_add_data(&vgpio->gc, vgpio); if (ret) { virtio_gpio_free_vqs(vdev); dev_err(dev, "Failed to add virtio-gpio controller\n"); } return ret; } static void virtio_gpio_remove(struct virtio_device *vdev) { struct virtio_gpio *vgpio = vdev->priv; gpiochip_remove(&vgpio->gc); virtio_gpio_free_vqs(vdev); } static const struct virtio_device_id id_table[] = { { VIRTIO_ID_GPIO, VIRTIO_DEV_ANY_ID }, {}, }; MODULE_DEVICE_TABLE(virtio, id_table); static const unsigned int features[] = { VIRTIO_GPIO_F_IRQ, }; static struct virtio_driver virtio_gpio_driver = { .feature_table = features, .feature_table_size = ARRAY_SIZE(features), .id_table = id_table, .probe = virtio_gpio_probe, .remove = virtio_gpio_remove, .driver = { .name = KBUILD_MODNAME, .owner = THIS_MODULE, }, }; module_virtio_driver(virtio_gpio_driver); MODULE_AUTHOR("Enrico Weigelt, metux IT consult "); MODULE_AUTHOR("Viresh Kumar "); MODULE_DESCRIPTION("VirtIO GPIO driver"); MODULE_LICENSE("GPL");