/* * Copyright (c) 2011 Samsung Electronics Co., Ltd. * Authors: * Inki Dae * Joonyoung Shim * Seung-Woo Kim * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License as published by the * Free Software Foundation; either version 2 of the License, or (at your * option) any later version. */ #include #include #include #include #include #include #include #include "exynos_drm_drv.h" #include "exynos_drm_crtc.h" #include "exynos_drm_fbdev.h" #include "exynos_drm_fb.h" #include "exynos_drm_gem.h" #include "exynos_drm_plane.h" #include "exynos_drm_vidi.h" #include "exynos_drm_g2d.h" #include "exynos_drm_ipp.h" #include "exynos_drm_iommu.h" #define DRIVER_NAME "exynos" #define DRIVER_DESC "Samsung SoC DRM" #define DRIVER_DATE "20110530" #define DRIVER_MAJOR 1 #define DRIVER_MINOR 0 struct exynos_atomic_commit { struct work_struct work; struct drm_device *dev; struct drm_atomic_state *state; u32 crtcs; }; static void exynos_atomic_wait_for_commit(struct drm_atomic_state *state) { struct drm_crtc_state *crtc_state; struct drm_crtc *crtc; int i, ret; for_each_crtc_in_state(state, crtc, crtc_state, i) { struct exynos_drm_crtc *exynos_crtc = to_exynos_crtc(crtc); if (!crtc->state->enable) continue; ret = drm_crtc_vblank_get(crtc); if (ret) continue; exynos_drm_crtc_wait_pending_update(exynos_crtc); drm_crtc_vblank_put(crtc); } } static void exynos_atomic_commit_complete(struct exynos_atomic_commit *commit) { struct drm_device *dev = commit->dev; struct exynos_drm_private *priv = dev->dev_private; struct drm_atomic_state *state = commit->state; struct drm_plane *plane; struct drm_crtc *crtc; struct drm_plane_state *plane_state; struct drm_crtc_state *crtc_state; int i; drm_atomic_helper_commit_modeset_disables(dev, state); drm_atomic_helper_commit_modeset_enables(dev, state); /* * Exynos can't update planes with CRTCs and encoders disabled, * its updates routines, specially for FIMD, requires the clocks * to be enabled. So it is necessary to handle the modeset operations * *before* the commit_planes() step, this way it will always * have the relevant clocks enabled to perform the update. */ for_each_crtc_in_state(state, crtc, crtc_state, i) { struct exynos_drm_crtc *exynos_crtc = to_exynos_crtc(crtc); atomic_set(&exynos_crtc->pending_update, 0); } for_each_plane_in_state(state, plane, plane_state, i) { struct exynos_drm_crtc *exynos_crtc = to_exynos_crtc(plane->crtc); if (!plane->crtc) continue; atomic_inc(&exynos_crtc->pending_update); } drm_atomic_helper_commit_planes(dev, state, false); exynos_atomic_wait_for_commit(state); drm_atomic_helper_cleanup_planes(dev, state); drm_atomic_state_free(state); spin_lock(&priv->lock); priv->pending &= ~commit->crtcs; spin_unlock(&priv->lock); wake_up_all(&priv->wait); kfree(commit); } static void exynos_drm_atomic_work(struct work_struct *work) { struct exynos_atomic_commit *commit = container_of(work, struct exynos_atomic_commit, work); exynos_atomic_commit_complete(commit); } static struct device *exynos_drm_get_dma_device(void); static int exynos_drm_load(struct drm_device *dev, unsigned long flags) { struct exynos_drm_private *private; struct drm_encoder *encoder; unsigned int clone_mask; int cnt, ret; private = kzalloc(sizeof(struct exynos_drm_private), GFP_KERNEL); if (!private) return -ENOMEM; init_waitqueue_head(&private->wait); spin_lock_init(&private->lock); dev_set_drvdata(dev->dev, dev); dev->dev_private = (void *)private; /* the first real CRTC device is used for all dma mapping operations */ private->dma_dev = exynos_drm_get_dma_device(); if (!private->dma_dev) { DRM_ERROR("no device found for DMA mapping operations.\n"); ret = -ENODEV; goto err_free_private; } DRM_INFO("Exynos DRM: using %s device for DMA mapping operations\n", dev_name(private->dma_dev)); /* create common IOMMU mapping for all devices attached to Exynos DRM */ ret = drm_create_iommu_mapping(dev); if (ret < 0) { DRM_ERROR("failed to create iommu mapping.\n"); goto err_free_private; } drm_mode_config_init(dev); exynos_drm_mode_config_init(dev); /* setup possible_clones. */ cnt = 0; clone_mask = 0; list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) clone_mask |= (1 << (cnt++)); list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) encoder->possible_clones = clone_mask; platform_set_drvdata(dev->platformdev, dev); /* Try to bind all sub drivers. */ ret = component_bind_all(dev->dev, dev); if (ret) goto err_mode_config_cleanup; ret = drm_vblank_init(dev, dev->mode_config.num_crtc); if (ret) goto err_unbind_all; /* Probe non kms sub drivers and virtual display driver. */ ret = exynos_drm_device_subdrv_probe(dev); if (ret) goto err_cleanup_vblank; drm_mode_config_reset(dev); /* * enable drm irq mode. * - with irq_enabled = true, we can use the vblank feature. * * P.S. note that we wouldn't use drm irq handler but * just specific driver own one instead because * drm framework supports only one irq handler. */ dev->irq_enabled = true; /* init kms poll for handling hpd */ drm_kms_helper_poll_init(dev); /* force connectors detection */ drm_helper_hpd_irq_event(dev); return 0; err_cleanup_vblank: drm_vblank_cleanup(dev); err_unbind_all: component_unbind_all(dev->dev, dev); err_mode_config_cleanup: drm_mode_config_cleanup(dev); drm_release_iommu_mapping(dev); err_free_private: kfree(private); return ret; } static int exynos_drm_unload(struct drm_device *dev) { exynos_drm_device_subdrv_remove(dev); exynos_drm_fbdev_fini(dev); drm_kms_helper_poll_fini(dev); drm_vblank_cleanup(dev); component_unbind_all(dev->dev, dev); drm_mode_config_cleanup(dev); drm_release_iommu_mapping(dev); kfree(dev->dev_private); dev->dev_private = NULL; return 0; } static int commit_is_pending(struct exynos_drm_private *priv, u32 crtcs) { bool pending; spin_lock(&priv->lock); pending = priv->pending & crtcs; spin_unlock(&priv->lock); return pending; } int exynos_atomic_commit(struct drm_device *dev, struct drm_atomic_state *state, bool nonblock) { struct exynos_drm_private *priv = dev->dev_private; struct exynos_atomic_commit *commit; struct drm_crtc *crtc; struct drm_crtc_state *crtc_state; int i, ret; commit = kzalloc(sizeof(*commit), GFP_KERNEL); if (!commit) return -ENOMEM; ret = drm_atomic_helper_prepare_planes(dev, state); if (ret) { kfree(commit); return ret; } /* This is the point of no return */ INIT_WORK(&commit->work, exynos_drm_atomic_work); commit->dev = dev; commit->state = state; /* Wait until all affected CRTCs have completed previous commits and * mark them as pending. */ for_each_crtc_in_state(state, crtc, crtc_state, i) commit->crtcs |= drm_crtc_mask(crtc); wait_event(priv->wait, !commit_is_pending(priv, commit->crtcs)); spin_lock(&priv->lock); priv->pending |= commit->crtcs; spin_unlock(&priv->lock); drm_atomic_helper_swap_state(state, true); if (nonblock) schedule_work(&commit->work); else exynos_atomic_commit_complete(commit); return 0; } static int exynos_drm_open(struct drm_device *dev, struct drm_file *file) { struct drm_exynos_file_private *file_priv; int ret; file_priv = kzalloc(sizeof(*file_priv), GFP_KERNEL); if (!file_priv) return -ENOMEM; file->driver_priv = file_priv; ret = exynos_drm_subdrv_open(dev, file); if (ret) goto err_file_priv_free; return ret; err_file_priv_free: kfree(file_priv); file->driver_priv = NULL; return ret; } static void exynos_drm_preclose(struct drm_device *dev, struct drm_file *file) { struct drm_crtc *crtc; exynos_drm_subdrv_close(dev, file); list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) exynos_drm_crtc_cancel_page_flip(crtc, file); } static void exynos_drm_postclose(struct drm_device *dev, struct drm_file *file) { kfree(file->driver_priv); file->driver_priv = NULL; } static void exynos_drm_lastclose(struct drm_device *dev) { exynos_drm_fbdev_restore_mode(dev); } static const struct vm_operations_struct exynos_drm_gem_vm_ops = { .fault = exynos_drm_gem_fault, .open = drm_gem_vm_open, .close = drm_gem_vm_close, }; static const struct drm_ioctl_desc exynos_ioctls[] = { DRM_IOCTL_DEF_DRV(EXYNOS_GEM_CREATE, exynos_drm_gem_create_ioctl, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_GEM_MAP, exynos_drm_gem_map_ioctl, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_GEM_GET, exynos_drm_gem_get_ioctl, DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_VIDI_CONNECTION, vidi_connection_ioctl, DRM_AUTH), DRM_IOCTL_DEF_DRV(EXYNOS_G2D_GET_VER, exynos_g2d_get_ver_ioctl, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_G2D_SET_CMDLIST, exynos_g2d_set_cmdlist_ioctl, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, exynos_g2d_exec_ioctl, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, exynos_drm_ipp_get_property, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, exynos_drm_ipp_set_property, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, exynos_drm_ipp_queue_buf, DRM_AUTH | DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, exynos_drm_ipp_cmd_ctrl, DRM_AUTH | DRM_RENDER_ALLOW), }; static const struct file_operations exynos_drm_driver_fops = { .owner = THIS_MODULE, .open = drm_open, .mmap = exynos_drm_gem_mmap, .poll = drm_poll, .read = drm_read, .unlocked_ioctl = drm_ioctl, #ifdef CONFIG_COMPAT .compat_ioctl = drm_compat_ioctl, #endif .release = drm_release, }; static struct drm_driver exynos_drm_driver = { .driver_features = DRIVER_MODESET | DRIVER_GEM | DRIVER_PRIME | DRIVER_ATOMIC | DRIVER_RENDER, .load = exynos_drm_load, .unload = exynos_drm_unload, .open = exynos_drm_open, .preclose = exynos_drm_preclose, .lastclose = exynos_drm_lastclose, .postclose = exynos_drm_postclose, .get_vblank_counter = drm_vblank_no_hw_counter, .enable_vblank = exynos_drm_crtc_enable_vblank, .disable_vblank = exynos_drm_crtc_disable_vblank, .gem_free_object_unlocked = exynos_drm_gem_free_object, .gem_vm_ops = &exynos_drm_gem_vm_ops, .dumb_create = exynos_drm_gem_dumb_create, .dumb_map_offset = exynos_drm_gem_dumb_map_offset, .dumb_destroy = drm_gem_dumb_destroy, .prime_handle_to_fd = drm_gem_prime_handle_to_fd, .prime_fd_to_handle = drm_gem_prime_fd_to_handle, .gem_prime_export = drm_gem_prime_export, .gem_prime_import = drm_gem_prime_import, .gem_prime_get_sg_table = exynos_drm_gem_prime_get_sg_table, .gem_prime_import_sg_table = exynos_drm_gem_prime_import_sg_table, .gem_prime_vmap = exynos_drm_gem_prime_vmap, .gem_prime_vunmap = exynos_drm_gem_prime_vunmap, .gem_prime_mmap = exynos_drm_gem_prime_mmap, .ioctls = exynos_ioctls, .num_ioctls = ARRAY_SIZE(exynos_ioctls), .fops = &exynos_drm_driver_fops, .name = DRIVER_NAME, .desc = DRIVER_DESC, .date = DRIVER_DATE, .major = DRIVER_MAJOR, .minor = DRIVER_MINOR, }; #ifdef CONFIG_PM_SLEEP static int exynos_drm_suspend(struct device *dev) { struct drm_device *drm_dev = dev_get_drvdata(dev); struct drm_connector *connector; if (pm_runtime_suspended(dev) || !drm_dev) return 0; drm_modeset_lock_all(drm_dev); drm_for_each_connector(connector, drm_dev) { int old_dpms = connector->dpms; if (connector->funcs->dpms) connector->funcs->dpms(connector, DRM_MODE_DPMS_OFF); /* Set the old mode back to the connector for resume */ connector->dpms = old_dpms; } drm_modeset_unlock_all(drm_dev); return 0; } static int exynos_drm_resume(struct device *dev) { struct drm_device *drm_dev = dev_get_drvdata(dev); struct drm_connector *connector; if (pm_runtime_suspended(dev) || !drm_dev) return 0; drm_modeset_lock_all(drm_dev); drm_for_each_connector(connector, drm_dev) { if (connector->funcs->dpms) { int dpms = connector->dpms; connector->dpms = DRM_MODE_DPMS_OFF; connector->funcs->dpms(connector, dpms); } } drm_modeset_unlock_all(drm_dev); return 0; } #endif static const struct dev_pm_ops exynos_drm_pm_ops = { SET_SYSTEM_SLEEP_PM_OPS(exynos_drm_suspend, exynos_drm_resume) }; /* forward declaration */ static struct platform_driver exynos_drm_platform_driver; struct exynos_drm_driver_info { struct platform_driver *driver; unsigned int flags; }; #define DRM_COMPONENT_DRIVER BIT(0) /* supports component framework */ #define DRM_VIRTUAL_DEVICE BIT(1) /* create virtual platform device */ #define DRM_DMA_DEVICE BIT(2) /* can be used for dma allocations */ #define DRV_PTR(drv, cond) (IS_ENABLED(cond) ? &drv : NULL) /* * Connector drivers should not be placed before associated crtc drivers, * because connector requires pipe number of its crtc during initialization. */ static struct exynos_drm_driver_info exynos_drm_drivers[] = { { DRV_PTR(fimd_driver, CONFIG_DRM_EXYNOS_FIMD), DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE }, { DRV_PTR(exynos5433_decon_driver, CONFIG_DRM_EXYNOS5433_DECON), DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE }, { DRV_PTR(decon_driver, CONFIG_DRM_EXYNOS7_DECON), DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE }, { DRV_PTR(mixer_driver, CONFIG_DRM_EXYNOS_MIXER), DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE }, { DRV_PTR(mic_driver, CONFIG_DRM_EXYNOS_MIC), DRM_COMPONENT_DRIVER }, { DRV_PTR(dp_driver, CONFIG_DRM_EXYNOS_DP), DRM_COMPONENT_DRIVER }, { DRV_PTR(dsi_driver, CONFIG_DRM_EXYNOS_DSI), DRM_COMPONENT_DRIVER }, { DRV_PTR(hdmi_driver, CONFIG_DRM_EXYNOS_HDMI), DRM_COMPONENT_DRIVER }, { DRV_PTR(vidi_driver, CONFIG_DRM_EXYNOS_VIDI), DRM_COMPONENT_DRIVER | DRM_VIRTUAL_DEVICE }, { DRV_PTR(g2d_driver, CONFIG_DRM_EXYNOS_G2D), }, { DRV_PTR(fimc_driver, CONFIG_DRM_EXYNOS_FIMC), }, { DRV_PTR(rotator_driver, CONFIG_DRM_EXYNOS_ROTATOR), }, { DRV_PTR(gsc_driver, CONFIG_DRM_EXYNOS_GSC), }, { DRV_PTR(ipp_driver, CONFIG_DRM_EXYNOS_IPP), DRM_VIRTUAL_DEVICE }, { &exynos_drm_platform_driver, DRM_VIRTUAL_DEVICE } }; static int compare_dev(struct device *dev, void *data) { return dev == (struct device *)data; } static struct component_match *exynos_drm_match_add(struct device *dev) { struct component_match *match = NULL; int i; for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; struct device *p = NULL, *d; if (!info->driver || !(info->flags & DRM_COMPONENT_DRIVER)) continue; while ((d = bus_find_device(&platform_bus_type, p, &info->driver->driver, (void *)platform_bus_type.match))) { put_device(p); component_match_add(dev, &match, compare_dev, d); p = d; } put_device(p); } return match ?: ERR_PTR(-ENODEV); } static int exynos_drm_bind(struct device *dev) { return drm_platform_init(&exynos_drm_driver, to_platform_device(dev)); } static void exynos_drm_unbind(struct device *dev) { drm_put_dev(dev_get_drvdata(dev)); } static const struct component_master_ops exynos_drm_ops = { .bind = exynos_drm_bind, .unbind = exynos_drm_unbind, }; static int exynos_drm_platform_probe(struct platform_device *pdev) { struct component_match *match; pdev->dev.coherent_dma_mask = DMA_BIT_MASK(32); exynos_drm_driver.num_ioctls = ARRAY_SIZE(exynos_ioctls); match = exynos_drm_match_add(&pdev->dev); if (IS_ERR(match)) return PTR_ERR(match); return component_master_add_with_match(&pdev->dev, &exynos_drm_ops, match); } static int exynos_drm_platform_remove(struct platform_device *pdev) { component_master_del(&pdev->dev, &exynos_drm_ops); return 0; } static struct platform_driver exynos_drm_platform_driver = { .probe = exynos_drm_platform_probe, .remove = exynos_drm_platform_remove, .driver = { .name = "exynos-drm", .pm = &exynos_drm_pm_ops, }, }; static struct device *exynos_drm_get_dma_device(void) { int i; for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; struct device *dev; if (!info->driver || !(info->flags & DRM_DMA_DEVICE)) continue; while ((dev = bus_find_device(&platform_bus_type, NULL, &info->driver->driver, (void *)platform_bus_type.match))) { put_device(dev); return dev; } } return NULL; } static void exynos_drm_unregister_devices(void) { int i; for (i = ARRAY_SIZE(exynos_drm_drivers) - 1; i >= 0; --i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; struct device *dev; if (!info->driver || !(info->flags & DRM_VIRTUAL_DEVICE)) continue; while ((dev = bus_find_device(&platform_bus_type, NULL, &info->driver->driver, (void *)platform_bus_type.match))) { put_device(dev); platform_device_unregister(to_platform_device(dev)); } } } static int exynos_drm_register_devices(void) { struct platform_device *pdev; int i; for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; if (!info->driver || !(info->flags & DRM_VIRTUAL_DEVICE)) continue; pdev = platform_device_register_simple( info->driver->driver.name, -1, NULL, 0); if (IS_ERR(pdev)) goto fail; } return 0; fail: exynos_drm_unregister_devices(); return PTR_ERR(pdev); } static void exynos_drm_unregister_drivers(void) { int i; for (i = ARRAY_SIZE(exynos_drm_drivers) - 1; i >= 0; --i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; if (!info->driver) continue; platform_driver_unregister(info->driver); } } static int exynos_drm_register_drivers(void) { int i, ret; for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; if (!info->driver) continue; ret = platform_driver_register(info->driver); if (ret) goto fail; } return 0; fail: exynos_drm_unregister_drivers(); return ret; } static int exynos_drm_init(void) { int ret; ret = exynos_drm_register_devices(); if (ret) return ret; ret = exynos_drm_register_drivers(); if (ret) goto err_unregister_pdevs; return 0; err_unregister_pdevs: exynos_drm_unregister_devices(); return ret; } static void exynos_drm_exit(void) { exynos_drm_unregister_drivers(); exynos_drm_unregister_devices(); } module_init(exynos_drm_init); module_exit(exynos_drm_exit); MODULE_AUTHOR("Inki Dae "); MODULE_AUTHOR("Joonyoung Shim "); MODULE_AUTHOR("Seung-Woo Kim "); MODULE_DESCRIPTION("Samsung SoC DRM Driver"); MODULE_LICENSE("GPL");