Message ID | 1355297682-7990-2-git-send-email-chulspro.kim@samsung.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
> -----Original Message----- > From: Eunchul Kim [mailto:chulspro.kim@samsung.com] > Sent: Wednesday, December 12, 2012 4:35 PM > To: dri-devel@lists.freedesktop.org; inki.dae@samsung.com > Cc: jy0.jeon@samsung.com; yj44.cho@samsung.com; jmock.shin@samsung.com; > jaejoon.seo@samsung.com; kyungmin.park@samsung.com; > chulspro.kim@samsung.com > Subject: [RFC v3 1/5] drm/exynos: add ipp subsystem > > IPP stand for Image Post Processing and supports image scaler/rotator > /crop/flip/csc(color space conversion) and input/output DMA operations > using ipp drivers. > also supports writeback and display output operations. > ipp driver include FIMC, Rotator, GSC, SC, so on. > and ipp is integration device driver for each hardware. > > Signed-off-by: Eunchul Kim <chulspro.kim@samsung.com> > Signed-off-by: Jinyoung Jeon <jy0.jeon@samsung.com> > --- > drivers/gpu/drm/exynos/Kconfig | 6 + > drivers/gpu/drm/exynos/Makefile | 1 + > drivers/gpu/drm/exynos/exynos_drm_drv.c | 24 + > drivers/gpu/drm/exynos/exynos_drm_drv.h | 7 + > drivers/gpu/drm/exynos/exynos_drm_ipp.c | 2059 > +++++++++++++++++++++++++++++++ > drivers/gpu/drm/exynos/exynos_drm_ipp.h | 266 ++++ > include/uapi/drm/exynos_drm.h | 190 +++ > 7 files changed, 2553 insertions(+), 0 deletions(-) > create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.c > create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.h > > diff --git a/drivers/gpu/drm/exynos/Kconfig > b/drivers/gpu/drm/exynos/Kconfig > index 4ea8cdc..bcf1c9d 100644 > --- a/drivers/gpu/drm/exynos/Kconfig > +++ b/drivers/gpu/drm/exynos/Kconfig > @@ -45,3 +45,9 @@ config DRM_EXYNOS_G2D > depends on DRM_EXYNOS && !VIDEO_SAMSUNG_S5P_G2D > help > Choose this option if you want to use Exynos G2D for DRM. > + > +config DRM_EXYNOS_IPP > + bool "Exynos DRM IPP" > + depends on DRM_EXYNOS > + help > + Choose this option if you want to use IPP feature for DRM. > diff --git a/drivers/gpu/drm/exynos/Makefile > b/drivers/gpu/drm/exynos/Makefile > index 26813b8..6c536ce 100644 > --- a/drivers/gpu/drm/exynos/Makefile > +++ b/drivers/gpu/drm/exynos/Makefile > @@ -16,5 +16,6 @@ exynosdrm-$(CONFIG_DRM_EXYNOS_HDMI) += > exynos_hdmi.o exynos_mixer.o \ > exynos_drm_hdmi.o > exynosdrm-$(CONFIG_DRM_EXYNOS_VIDI) += exynos_drm_vidi.o > exynosdrm-$(CONFIG_DRM_EXYNOS_G2D) += exynos_drm_g2d.o > +exynosdrm-$(CONFIG_DRM_EXYNOS_IPP) += exynos_drm_ipp.o > > obj-$(CONFIG_DRM_EXYNOS) += exynosdrm.o > diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.c > b/drivers/gpu/drm/exynos/exynos_drm_drv.c > index 4a1168d..0eb8a97 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_drv.c > +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.c > @@ -40,6 +40,7 @@ > #include "exynos_drm_vidi.h" > #include "exynos_drm_dmabuf.h" > #include "exynos_drm_g2d.h" > +#include "exynos_drm_ipp.h" > #include "exynos_drm_iommu.h" > > #define DRIVER_NAME "exynos" > @@ -249,6 +250,14 @@ static struct drm_ioctl_desc exynos_ioctls[] = { > exynos_g2d_set_cmdlist_ioctl, DRM_UNLOCKED | DRM_AUTH), > DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, > exynos_g2d_exec_ioctl, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, > + exynos_drm_ipp_get_property, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, > + exynos_drm_ipp_set_property, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, > + exynos_drm_ipp_queue_buf, DRM_UNLOCKED | DRM_AUTH), > + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, > + exynos_drm_ipp_cmd_ctrl, DRM_UNLOCKED | DRM_AUTH), > }; > > static const struct file_operations exynos_drm_driver_fops = { > @@ -363,6 +372,12 @@ static int __init exynos_drm_init(void) > goto out_g2d; > #endif > > +#ifdef CONFIG_DRM_EXYNOS_IPP > + ret = platform_driver_register(&ipp_driver); > + if (ret < 0) > + goto out_ipp; > +#endif > + > ret = platform_driver_register(&exynos_drm_platform_driver); > if (ret < 0) > goto out_drm; > @@ -380,6 +395,11 @@ out: > platform_driver_unregister(&exynos_drm_platform_driver); > > out_drm: > +#ifdef CONFIG_DRM_EXYNOS_IPP > + platform_driver_unregister(&ipp_driver); > +out_ipp: > +#endif > + > #ifdef CONFIG_DRM_EXYNOS_G2D > platform_driver_unregister(&g2d_driver); > out_g2d: > @@ -416,6 +436,10 @@ static void __exit exynos_drm_exit(void) > > platform_driver_unregister(&exynos_drm_platform_driver); > > +#ifdef CONFIG_DRM_EXYNOS_IPP > + platform_driver_unregister(&ipp_driver); > +#endif > + > #ifdef CONFIG_DRM_EXYNOS_G2D > platform_driver_unregister(&g2d_driver); > #endif > diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.h > b/drivers/gpu/drm/exynos/exynos_drm_drv.h > index a9db025..a365788 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_drv.h > +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.h > @@ -235,8 +235,14 @@ struct exynos_drm_g2d_private { > unsigned int gem_nr; > }; > > +struct exynos_drm_ipp_private { > + struct device *dev; > + struct list_head event_list; > +}; > + > struct drm_exynos_file_private { > struct exynos_drm_g2d_private *g2d_priv; > + struct exynos_drm_ipp_private *ipp_priv; > }; > > /* > @@ -346,4 +352,5 @@ extern struct platform_driver mixer_driver; > extern struct platform_driver exynos_drm_common_hdmi_driver; > extern struct platform_driver vidi_driver; > extern struct platform_driver g2d_driver; > +extern struct platform_driver ipp_driver; > #endif > diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.c > b/drivers/gpu/drm/exynos/exynos_drm_ipp.c > new file mode 100644 > index 0000000..eb12cb5 > --- /dev/null > +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.c > @@ -0,0 +1,2059 @@ > +/* > + * Copyright (C) 2012 Samsung Electronics Co.Ltd > + * Authors: > + * Eunchul Kim <chulspro.kim@samsung.com> > + * Jinyoung Jeon <jy0.jeon@samsung.com> > + * Sangmin Lee <lsmin.lee@samsung.com> > + * > + * This program is free software; you can redistribute it and/or modify > it > + * under the terms of the GNU General Public License as published by > the > + * Free Software Foundation; either version 2 of the License, or (at > your > + * option) any later version. > + * > + */ > +#include <linux/kernel.h> > +#include <linux/module.h> > +#include <linux/platform_device.h> > +#include <linux/clk.h> > +#include <linux/pm_runtime.h> > +#include <plat/map-base.h> > + > +#include <drm/drmP.h> > +#include <drm/exynos_drm.h> > +#include "exynos_drm_drv.h" > +#include "exynos_drm_gem.h" > +#include "exynos_drm_ipp.h" > + > +/* > + * IPP is stand for Image Post Processing and > + * supports image scaler/rotator and input/output DMA operations. > + * using FIMC, GSC, Rotator, so on. > + * IPP is integration device driver of same attribute h/w > + */ > + > +#define get_ipp_context(dev) > platform_get_drvdata(to_platform_device(dev)) > +#define ipp_is_m2m_cmd(c) (c == IPP_CMD_M2M) > + > +/* > + * A structure of event. > + * > + * @base: base of event. > + * @event: ipp event. > + */ > +struct drm_exynos_ipp_send_event { > + struct drm_pending_event base; > + struct drm_exynos_ipp_event event; > +}; > + > +/* > + * A structure of memory node. > + * > + * @list: list head to memory queue information. > + * @ops_id: id of operations. > + * @prop_id: id of property. > + * @buf_id: id of buffer. > + * @buf_info: gem objects and dma address, size. > + * @filp: a pointer to drm_file. > + */ > +struct drm_exynos_ipp_mem_node { > + struct list_head list; > + enum drm_exynos_ops_id ops_id; > + u32 prop_id; > + u32 buf_id; > + struct drm_exynos_ipp_buf_info buf_info; > + struct drm_file *filp; > +}; > + > +/* > + * A structure of ipp context. > + * > + * @subdrv: prepare initialization using subdrv. > + * @ipp_lock: lock for synchronization of access to ipp_idr. > + * @prop_lock: lock for synchronization of access to prop_idr. > + * @ipp_idr: ipp driver idr. > + * @prop_idr: property idr. > + * @event_workq: event work queue. > + * @cmd_workq: command work queue. > + */ > +struct ipp_context { > + struct exynos_drm_subdrv subdrv; > + struct mutex ipp_lock; > + struct mutex prop_lock; > + struct idr ipp_idr; > + struct idr prop_idr; > + struct workqueue_struct *event_workq; > + struct workqueue_struct *cmd_workq; > +}; > + > +static LIST_HEAD(exynos_drm_ippdrv_list); > +static DEFINE_MUTEX(exynos_drm_ippdrv_lock); > +static BLOCKING_NOTIFIER_HEAD(exynos_drm_ippnb_list); > + > +int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ippdrv) > + return -EINVAL; > + > + mutex_lock(&exynos_drm_ippdrv_lock); > + list_add_tail(&ippdrv->drv_list, &exynos_drm_ippdrv_list); > + mutex_unlock(&exynos_drm_ippdrv_lock); > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_register); > + > +int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ippdrv) > + return -EINVAL; > + > + mutex_lock(&exynos_drm_ippdrv_lock); > + list_del(&ippdrv->drv_list); > + mutex_unlock(&exynos_drm_ippdrv_lock); > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_unregister); > + > +static int ipp_create_id(struct idr *id_idr, struct mutex *lock, void > *obj, > + u32 *idp) > +{ > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > +again: > + /* ensure there is space available to allocate a handle */ > + if (idr_pre_get(id_idr, GFP_KERNEL) == 0) { > + DRM_ERROR("failed to get idr.\n"); > + return -ENOMEM; > + } > + > + /* do the allocation under our mutexlock */ > + mutex_lock(lock); > + ret = idr_get_new_above(id_idr, obj, 1, (int *)idp); > + mutex_unlock(lock); > + if (ret == -EAGAIN) > + goto again; > + > + return ret; > +} > + > +static void *ipp_find_obj(struct idr *id_idr, struct mutex *lock, u32 id) > +{ > + void *obj; > + > + DRM_DEBUG_KMS("%s:id[%d]\n", __func__, id); > + > + mutex_lock(lock); > + > + /* find object using handle */ > + obj = idr_find(id_idr, id); > + if (!obj) { > + DRM_ERROR("failed to find object.\n"); > + mutex_unlock(lock); > + return ERR_PTR(-ENODEV); > + } > + > + mutex_unlock(lock); > + > + return obj; > +} > + > +static inline bool ipp_check_dedicated(struct exynos_drm_ippdrv *ippdrv, > + enum drm_exynos_ipp_cmd cmd) > +{ > + /* > + * check dedicated flag and WB, OUTPUT operation with > + * power on state. > + */ > + if (ippdrv->dedicated || (!ipp_is_m2m_cmd(cmd) && > + !pm_runtime_suspended(ippdrv->dev))) > + return true; > + > + return false; > +} > + > +static struct exynos_drm_ippdrv *ipp_find_driver(struct ipp_context *ctx, > + struct drm_exynos_ipp_property *property) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + u32 ipp_id = property->ipp_id; > + > + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, ipp_id); > + > + if (ipp_id) { > + /* find ipp driver using idr */ > + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, > + ipp_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("not found ipp%d driver.\n", ipp_id); > + goto err_null; Just return ippdrv; > + } > + > + /* > + * WB, OUTPUT opertion not supported multi-operation. > + * so, make dedicated state at set property ioctl. > + * when ipp driver finished operations, clear dedicated flags. > + */ > + if (ipp_check_dedicated(ippdrv, property->cmd)) { > + DRM_ERROR("already used choose device.\n"); > + goto err_null; Just return proper error; > + } > + > + /* > + * This is necessary to find correct device in ipp drivers. > + * ipp drivers have different abilities, > + * so need to check property. > + */ > + if (ippdrv->check_property && > + ippdrv->check_property(ippdrv->dev, property)) { > + DRM_ERROR("not support property.\n"); > + goto err_null; Just return proper error; > + } > + > + return ippdrv; > + } else { > + /* > + * This case is search all ipp driver for finding. > + * user application don't set ipp_id in this case, > + * so ipp subsystem search correct driver in driver list. > + */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, > drv_list) { > + if (ipp_check_dedicated(ippdrv, property->cmd)) { > + DRM_DEBUG_KMS("%s:used device.\n", __func__); > + continue; > + } > + > + if (ippdrv->check_property && > + ippdrv->check_property(ippdrv->dev, property)) { > + DRM_DEBUG_KMS("%s:not support property.\n", > + __func__); > + continue; > + } > + > + return ippdrv; > + } > + > + DRM_ERROR("not support ipp driver operations.\n"); > + } > + > + return ERR_PTR(-ENODEV); > + > +err_null: > + return NULL; And remove the above codes. > +} > + > +static struct exynos_drm_ippdrv *ipp_find_drv_by_handle(u32 prop_id) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); > + > + if (list_empty(&exynos_drm_ippdrv_list)) { > + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); > + return ERR_PTR(-ENODEV); > + } > + > + /* > + * This case is search ipp driver by prop_id handle. > + * sometimes, ipp subsystem find driver by prop_id. > + * e.g PAUSE state, queue buf, command contro. > + */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", __func__, > + count++, (int)ippdrv); > + > + if (!list_empty(&ippdrv->cmd_list)) { > + list_for_each_entry(c_node, &ippdrv->cmd_list, list) > + if (c_node->property.prop_id == prop_id) > + return ippdrv; > + } > + } > + > + return ERR_PTR(-ENODEV); > +} > + > +int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_prop_list *prop_list = data; > + struct exynos_drm_ippdrv *ippdrv; > + int count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!prop_list) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, prop_list->ipp_id); > + > + if (!prop_list->ipp_id) { > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, > drv_list) > + count++; > + /* > + * Supports ippdrv list count for user application. > + * First step user application getting ippdrv count. > + * and second step getting ippdrv capability using ipp_id. > + */ > + prop_list->count = count; > + } else { > + /* > + * Getting ippdrv capability by ipp_id. > + * some deivce not supported wb, output interface. > + * so, user application detect correct ipp driver > + * using this ioctl. > + */ > + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, > + prop_list->ipp_id); > + if (!ippdrv) { > + DRM_ERROR("not found ipp%d driver.\n", > + prop_list->ipp_id); > + return -EINVAL; > + } > + > + prop_list = ippdrv->prop_list; > + } > + > + return 0; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_get_property); Should this function really be exported? Which external module calls this function? If not so, remove it. > + > +static void ipp_print_property(struct drm_exynos_ipp_property *property, > + int idx) > +{ > + struct drm_exynos_ipp_config *config = &property->config[idx]; > + struct drm_exynos_pos *pos = &config->pos; > + struct drm_exynos_sz *sz = &config->sz; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]ops[%s]fmt[0x%x]\n", > + __func__, property->prop_id, idx ? "dst" : "src", config- > >fmt); > + > + DRM_DEBUG_KMS("%s:pos[%d %d %d %d]sz[%d %d]f[%d]r[%d]\n", > + __func__, pos->x, pos->y, pos->w, pos->h, > + sz->hsize, sz->vsize, config->flip, config->degree); > +} > + > +static int ipp_find_and_set_property(struct drm_exynos_ipp_property > *property) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + u32 prop_id = property->prop_id; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); > + > + ippdrv = ipp_find_drv_by_handle(prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; > + } > + > + /* > + * Find command node using command list in ippdrv. > + * when we find this command no using prop_id. > + * return property information set in this command node. > + */ > + list_for_each_entry(c_node, &ippdrv->cmd_list, list) { > + if ((c_node->property.prop_id == prop_id) && > + (c_node->state == IPP_STATE_STOP)) { > + DRM_DEBUG_KMS("%s:found cmd[%d]ippdrv[0x%x]\n", > + __func__, property->cmd, (int)ippdrv); > + > + c_node->property = *property; > + return 0; > + } > + } > + > + DRM_ERROR("failed to search property.\n"); > + > + return -EINVAL; > +} > + > +static struct drm_exynos_ipp_cmd_work *ipp_create_cmd_work(void) > +{ > + struct drm_exynos_ipp_cmd_work *cmd_work; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + cmd_work = kzalloc(sizeof(*cmd_work), GFP_KERNEL); > + if (!cmd_work) { > + DRM_ERROR("failed to alloc cmd_work.\n"); > + return ERR_PTR(-ENOMEM); > + } > + > + INIT_WORK((struct work_struct *)cmd_work, ipp_sched_cmd); > + > + return cmd_work; > +} > + > +static struct drm_exynos_ipp_event_work *ipp_create_event_work(void) > +{ > + struct drm_exynos_ipp_event_work *event_work; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + event_work = kzalloc(sizeof(*event_work), GFP_KERNEL); > + if (!event_work) { > + DRM_ERROR("failed to alloc event_work.\n"); > + return ERR_PTR(-ENOMEM); > + } > + > + INIT_WORK((struct work_struct *)event_work, ipp_sched_event); > + > + return event_work; > +} > + > +int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_property *property = data; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret, i; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!property) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + /* > + * This is log print for user application property. > + * user application set various property. > + */ > + for_each_ipp_ops(i) > + ipp_print_property(property, i); > + > + /* > + * set property ioctl generated new prop_id. > + * but in this case already asigned prop_id using old set property. > + * e.g PAUSE state. this case supports find current prop_id and use > it > + * instead of allocation. > + */ > + if (property->prop_id) { > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property- > >prop_id); > + return ipp_find_and_set_property(property); > + } > + > + /* find ipp driver using ipp id */ > + ippdrv = ipp_find_driver(ctx, property); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; > + } > + > + /* allocate command node */ > + c_node = kzalloc(sizeof(*c_node), GFP_KERNEL); > + if (!c_node) { > + DRM_ERROR("failed to allocate map node.\n"); > + return -ENOMEM; > + } > + > + /* create property id */ > + ret = ipp_create_id(&ctx->prop_idr, &ctx->prop_lock, c_node, > + &property->prop_id); > + if (ret) { > + DRM_ERROR("failed to create id.\n"); > + goto err_clear; > + } > + > + DRM_DEBUG_KMS("%s:created prop_id[%d]cmd[%d]ippdrv[0x%x]\n", > + __func__, property->prop_id, property->cmd, (int)ippdrv); > + > + /* stored property information and ippdrv in private data */ > + c_node->priv = priv; > + c_node->property = *property; > + c_node->state = IPP_STATE_IDLE; > + > + c_node->start_work = ipp_create_cmd_work(); > + if (IS_ERR_OR_NULL(c_node->start_work)) { > + DRM_ERROR("failed to create start work.\n"); > + goto err_clear; > + } > + > + c_node->stop_work = ipp_create_cmd_work(); > + if (IS_ERR_OR_NULL(c_node->stop_work)) { > + DRM_ERROR("failed to create stop work.\n"); > + goto err_free_start; > + } > + > + c_node->event_work = ipp_create_event_work(); > + if (IS_ERR_OR_NULL(c_node->event_work)) { > + DRM_ERROR("failed to create event work.\n"); > + goto err_free_stop; > + } > + > + mutex_init(&c_node->cmd_lock); > + mutex_init(&c_node->mem_lock); > + mutex_init(&c_node->event_lock); > + > + init_completion(&c_node->start_complete); > + init_completion(&c_node->stop_complete); > + > + for_each_ipp_ops(i) > + INIT_LIST_HEAD(&c_node->mem_list[i]); > + > + INIT_LIST_HEAD(&c_node->event_list); > + list_splice_init(&priv->event_list, &c_node->event_list); > + list_add_tail(&c_node->list, &ippdrv->cmd_list); > + > + /* make dedicated state without m2m */ > + if (!ipp_is_m2m_cmd(property->cmd)) > + ippdrv->dedicated = true; > + > + return 0; > + > +err_free_stop: > + kfree(c_node->stop_work); > +err_free_start: > + kfree(c_node->start_work); > +err_clear: > + kfree(c_node); > + return ret; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_set_property); > + > +static void ipp_clean_cmd_node(struct drm_exynos_ipp_cmd_node *c_node) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* delete list */ > + list_del(&c_node->list); > + > + /* destroy mutex */ > + mutex_destroy(&c_node->cmd_lock); > + mutex_destroy(&c_node->mem_lock); > + mutex_destroy(&c_node->event_lock); > + > + /* free command node */ > + kfree(c_node->start_work); > + kfree(c_node->stop_work); > + kfree(c_node->event_work); > + kfree(c_node); > +} > + > +static int ipp_check_mem_list(struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct drm_exynos_ipp_mem_node *m_node; > + struct list_head *head; > + int ret, i, count[EXYNOS_DRM_OPS_MAX] = { 0, }; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_lock(&c_node->mem_lock); > + > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:%s memory empty.\n", __func__, > + i ? "dst" : "src"); > + continue; > + } > + > + /* find memory node entry */ > + list_for_each_entry(m_node, head, list) { > + DRM_DEBUG_KMS("%s:%s,count[%d]m_node[0x%x]\n", > __func__, > + i ? "dst" : "src", count[i], (int)m_node); > + count[i]++; > + } > + } > + > + DRM_DEBUG_KMS("%s:min[%d]max[%d]\n", __func__, > + min(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST]), > + max(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST])); > + > + /* > + * M2M operations should be need paired memory address. > + * so, need to check minimum count about src, dst. > + * other case not use paired memory, so use maximum count > + */ > + if (ipp_is_m2m_cmd(property->cmd)) > + ret = min(count[EXYNOS_DRM_OPS_SRC], > + count[EXYNOS_DRM_OPS_DST]); > + else > + ret = max(count[EXYNOS_DRM_OPS_SRC], > + count[EXYNOS_DRM_OPS_DST]); > + > + mutex_unlock(&c_node->mem_lock); > + > + return ret; > +} > + > +static struct drm_exynos_ipp_mem_node > + *ipp_find_mem_node(struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct list_head *head; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, qbuf->buf_id); > + > + /* source/destination memory list */ > + head = &c_node->mem_list[qbuf->ops_id]; > + > + /* find memory node from memory list */ > + list_for_each_entry(m_node, head, list) { > + DRM_DEBUG_KMS("%s:count[%d]m_node[0x%x]\n", > + __func__, count++, (int)m_node); > + > + /* compare buffer id */ > + if (m_node->buf_id == qbuf->buf_id) > + return m_node; > + } > + > + return NULL; > +} > + > +static int ipp_set_mem_node(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node) > +{ > + struct exynos_drm_ipp_ops *ops = NULL; > + int ret = 0; > + > + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); > + > + if (!m_node) { > + DRM_ERROR("invalid queue node.\n"); > + return -EFAULT; > + } > + > + mutex_lock(&c_node->mem_lock); > + > + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); > + > + /* get operations callback */ > + ops = ippdrv->ops[m_node->ops_id]; > + if (!ops) { > + DRM_ERROR("not support ops.\n"); > + ret = -EFAULT; > + goto err_unlock; > + } > + > + /* set address and enable irq */ > + if (ops->set_addr) { > + ret = ops->set_addr(ippdrv->dev, &m_node->buf_info, > + m_node->buf_id, IPP_BUF_ENQUEUE); > + if (ret) { > + DRM_ERROR("failed to set addr.\n"); > + goto err_unlock; > + } > + } > + > +err_unlock: > + mutex_unlock(&c_node->mem_lock); > + return ret; > +} > + > +static struct drm_exynos_ipp_mem_node > + *ipp_get_mem_node(struct drm_device *drm_dev, > + struct drm_file *file, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_buf_info buf_info; > + void *addr; > + int i; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_lock(&c_node->mem_lock); > + > + m_node = kzalloc(sizeof(*m_node), GFP_KERNEL); > + if (!m_node) { > + DRM_ERROR("failed to allocate queue node.\n"); > + goto err_unlock; > + } > + > + /* clear base address for error handling */ > + memset(&buf_info, 0x0, sizeof(buf_info)); > + > + /* operations, buffer id */ > + m_node->ops_id = qbuf->ops_id; > + m_node->prop_id = qbuf->prop_id; > + m_node->buf_id = qbuf->buf_id; > + > + DRM_DEBUG_KMS("%s:m_node[0x%x]ops_id[%d]\n", __func__, > + (int)m_node, qbuf->ops_id); > + DRM_DEBUG_KMS("%s:prop_id[%d]buf_id[%d]\n", __func__, > + qbuf->prop_id, m_node->buf_id); > + > + for_each_ipp_planar(i) { > + DRM_DEBUG_KMS("%s:i[%d]handle[0x%x]\n", __func__, > + i, qbuf->handle[i]); > + > + /* get dma address by handle */ > + if (qbuf->handle[i] != 0) { if (qbuf->handle[i]) { > + addr = exynos_drm_gem_get_dma_addr(drm_dev, > + qbuf->handle[i], file); > + if (!addr) { if (IS_ERR(addr)) > + DRM_ERROR("failed to get addr.\n"); > + goto err_clear; > + } > + > + buf_info.handles[i] = qbuf->handle[i]; > + buf_info.base[i] = *(dma_addr_t *) addr; > + DRM_DEBUG_KMS("%s:i[%d]base[0x%x]hd[0x%x]\n", > + __func__, i, buf_info.base[i], > + (int)buf_info.handles[i]); > + } > + } > + > + m_node->filp = file; > + m_node->buf_info = buf_info; > + list_add_tail(&m_node->list, &c_node->mem_list[qbuf->ops_id]); > + > + mutex_unlock(&c_node->mem_lock); > + return m_node; > + > +err_clear: > + kfree(m_node); > +err_unlock: > + mutex_unlock(&c_node->mem_lock); > + return ERR_PTR(-EFAULT); > +} > + > +static int ipp_put_mem_node(struct drm_device *drm_dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node) > +{ > + int i; > + > + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); > + > + if (!m_node) { > + DRM_ERROR("invalid dequeue node.\n"); > + return -EFAULT; > + } > + > + if (list_empty(&m_node->list)) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + mutex_lock(&c_node->mem_lock); > + > + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); > + > + /* put gem buffer */ > + for_each_ipp_planar(i) { > + unsigned long handle = m_node->buf_info.handles[i]; > + if (handle) > + exynos_drm_gem_put_dma_addr(drm_dev, handle, > + m_node->filp); > + } > + > + /* delete list in queue */ > + list_del(&m_node->list); > + kfree(m_node); > + > + mutex_unlock(&c_node->mem_lock); > + > + return 0; > +} > + > +static void ipp_free_event(struct drm_pending_event *event) Does this function need? This function just calls kfree. Just call kfree instead of wrapper. > +{ > + kfree(event); > +} > + > +static int ipp_get_event(struct drm_device *drm_dev, > + struct drm_file *file, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_send_event *e; > + unsigned long flags; > + > + DRM_DEBUG_KMS("%s:ops_id[%d]buf_id[%d]\n", __func__, > + qbuf->ops_id, qbuf->buf_id); > + > + e = kzalloc(sizeof(*e), GFP_KERNEL); > + > + if (!e) { > + DRM_ERROR("failed to allocate event.\n"); > + spin_lock_irqsave(&drm_dev->event_lock, flags); > + file->event_space += sizeof(e->event); > + spin_unlock_irqrestore(&drm_dev->event_lock, flags); > + return -ENOMEM; > + } > + > + /* make event */ > + e->event.base.type = DRM_EXYNOS_IPP_EVENT; > + e->event.base.length = sizeof(e->event); > + e->event.user_data = qbuf->user_data; > + e->event.prop_id = qbuf->prop_id; > + e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; > + e->base.event = &e->event.base; > + e->base.file_priv = file; > + e->base.destroy = ipp_free_event; > + list_add_tail(&e->base.link, &c_node->event_list); > + > + return 0; > +} > + > +static void ipp_put_event(struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_send_event *e, *te; > + int count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (list_empty(&c_node->event_list)) { > + DRM_DEBUG_KMS("%s:event_list is empty.\n", __func__); > + return; > + } > + > + list_for_each_entry_safe(e, te, &c_node->event_list, base.link) { > + DRM_DEBUG_KMS("%s:count[%d]e[0x%x]\n", > + __func__, count++, (int)e); > + > + /* > + * quf == NULL condition means all event deletion. > + * stop operations want to delete all event list. > + * another case delete only same buf id. > + */ > + if (!qbuf) { > + /* delete list */ > + list_del(&e->base.link); > + kfree(e); > + } > + > + /* compare buffer id */ > + if (qbuf && (qbuf->buf_id == > + e->event.buf_id[EXYNOS_DRM_OPS_DST])) { > + /* delete list */ > + list_del(&e->base.link); > + kfree(e); > + return; > + } > + } > +} > + > +void ipp_handle_cmd_work(struct device *dev, > + struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_work *cmd_work, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + cmd_work->ippdrv = ippdrv; > + cmd_work->c_node = c_node; > + queue_work(ctx->cmd_workq, (struct work_struct *)cmd_work); > +} > + > +static int ipp_queue_buf_with_run(struct device *dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_mem_node *m_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_property *property; > + struct exynos_drm_ipp_ops *ops; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + ippdrv = ipp_find_drv_by_handle(qbuf->prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EFAULT; > + } > + > + ops = ippdrv->ops[qbuf->ops_id]; > + if (!ops) { > + DRM_ERROR("failed to get ops.\n"); > + return -EFAULT; > + } > + > + property = &c_node->property; > + if (!property) { Property couldn't be NULL. remove it. > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + if (c_node->state != IPP_STATE_START) { > + DRM_DEBUG_KMS("%s:bypass for invalid state.\n" , __func__); > + return 0; No error? > + } > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return 0; No error? > + } > + > + /* > + * If set destination buffer and enabled clock, > + * then m2m operations need start operations at queue_buf > + */ > + if (ipp_is_m2m_cmd(property->cmd)) { > + struct drm_exynos_ipp_cmd_work *cmd_work = c_node- > >start_work; > + > + cmd_work->ctrl = IPP_CTRL_PLAY; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + } else { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +static void ipp_clean_queue_buf(struct drm_device *drm_dev, > + struct drm_exynos_ipp_cmd_node *c_node, > + struct drm_exynos_ipp_queue_buf *qbuf) > +{ > + struct drm_exynos_ipp_mem_node *m_node, *tm_node; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!list_empty(&c_node->mem_list[qbuf->ops_id])) { > + /* delete list */ > + list_for_each_entry_safe(m_node, tm_node, > + &c_node->mem_list[qbuf->ops_id], list) { > + if (m_node->buf_id == qbuf->buf_id && > + m_node->ops_id == qbuf->ops_id) > + ipp_put_mem_node(drm_dev, c_node, m_node); > + } > + } > +} > + > +int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_queue_buf *qbuf = data; > + struct drm_exynos_ipp_cmd_node *c_node; > + struct drm_exynos_ipp_mem_node *m_node; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!qbuf) { > + DRM_ERROR("invalid buf parameter.\n"); > + return -EINVAL; > + } > + > + if (qbuf->ops_id >= EXYNOS_DRM_OPS_MAX) { > + DRM_ERROR("invalid ops parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", > + __func__, qbuf->prop_id, qbuf->ops_id ? "dst" : "src", > + qbuf->buf_id, qbuf->buf_type); > + > + /* find command node */ > + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, > + qbuf->prop_id); > + if (!c_node) { > + DRM_ERROR("failed to get command node.\n"); > + return -EFAULT; > + } > + > + /* buffer control */ > + switch (qbuf->buf_type) { > + case IPP_BUF_ENQUEUE: > + /* get memory node */ > + m_node = ipp_get_mem_node(drm_dev, file, c_node, qbuf); > + if (IS_ERR_OR_NULL(m_node)) { This function never returns error. Use IS_ERR(m_node) instead. > + DRM_ERROR("failed to get m_node.\n"); > + return -EINVAL; And return PTR_ERR(m_node); > + } > + > + /* > + * first step get event for destination buffer. > + * and second step when M2M case run with destination buffer > + * if needed. > + */ > + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) { > + /* get event for destination buffer */ > + ret = ipp_get_event(drm_dev, file, c_node, qbuf); > + if (ret) { > + DRM_ERROR("failed to get event.\n"); > + goto err_clean_node; > + } > + > + /* > + * M2M case run play control for streaming feature. > + * other case set address and waiting. > + */ > + ret = ipp_queue_buf_with_run(dev, c_node, m_node, > qbuf); > + if (ret) { > + DRM_ERROR("failed to run command.\n"); > + goto err_clean_node; > + } > + } > + break; > + case IPP_BUF_DEQUEUE: > + mutex_lock(&c_node->cmd_lock); > + > + /* put event for destination buffer */ > + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) > + ipp_put_event(c_node, qbuf); > + > + ipp_clean_queue_buf(drm_dev, c_node, qbuf); > + > + mutex_unlock(&c_node->cmd_lock); > + break; > + default: > + DRM_ERROR("invalid buffer control.\n"); > + return -EINVAL; > + } > + > + return 0; > + > +err_clean_node: > + DRM_ERROR("clean memory nodes.\n"); > + > + ipp_clean_queue_buf(drm_dev, c_node, qbuf); > + return ret; > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_queue_buf); Remove EXPORT_SYMBOL_GPL. > + > +static bool exynos_drm_ipp_check_valid(struct device *dev, > + enum drm_exynos_ipp_ctrl ctrl, enum drm_exynos_ipp_state > state) > +{ > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (ctrl != IPP_CTRL_PLAY) { > + if (pm_runtime_suspended(dev)) { > + DRM_ERROR("pm:runtime_suspended.\n"); > + goto err_status; > + } > + } > + > + switch (ctrl) { > + case IPP_CTRL_PLAY: > + if (state != IPP_STATE_IDLE) > + goto err_status; > + break; > + case IPP_CTRL_STOP: > + if (state == IPP_STATE_STOP) > + goto err_status; > + break; > + case IPP_CTRL_PAUSE: > + if (state != IPP_STATE_START) > + goto err_status; > + break; > + case IPP_CTRL_RESUME: > + if (state != IPP_STATE_STOP) > + goto err_status; > + break; > + default: > + DRM_ERROR("invalid state.\n"); > + goto err_status; > + break; > + } > + > + return true; > + > +err_status: > + DRM_ERROR("invalid status:ctrl[%d]state[%d]\n", ctrl, state); > + return false; > +} > + > +int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct exynos_drm_ippdrv *ippdrv = NULL; > + struct device *dev = priv->dev; > + struct ipp_context *ctx = get_ipp_context(dev); > + struct drm_exynos_ipp_cmd_ctrl *cmd_ctrl = data; > + struct drm_exynos_ipp_cmd_work *cmd_work; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!ctx) { > + DRM_ERROR("invalid context.\n"); > + return -EINVAL; > + } > + > + if (!cmd_ctrl) { > + DRM_ERROR("invalid control parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:ctrl[%d]prop_id[%d]\n", __func__, > + cmd_ctrl->ctrl, cmd_ctrl->prop_id); > + > + ippdrv = ipp_find_drv_by_handle(cmd_ctrl->prop_id); > + if (IS_ERR_OR_NULL(ippdrv)) { Never return NULL. just use IS_ERR(ippdrv) > + DRM_ERROR("failed to get ipp driver.\n"); > + return -EINVAL; And return PTR_ERR(ippdrv); > + } > + > + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, > + cmd_ctrl->prop_id); > + if (!c_node) { > + DRM_ERROR("invalid command node list.\n"); > + return -EINVAL; > + } > + > + if (!exynos_drm_ipp_check_valid(ippdrv->dev, cmd_ctrl->ctrl, > + c_node->state)) { > + DRM_ERROR("invalid state.\n"); > + return -EINVAL; > + } > + > + switch (cmd_ctrl->ctrl) { > + case IPP_CTRL_PLAY: > + if (pm_runtime_suspended(ippdrv->dev)) > + pm_runtime_get_sync(ippdrv->dev); > + c_node->state = IPP_STATE_START; > + > + cmd_work = c_node->start_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + c_node->state = IPP_STATE_START; > + break; > + case IPP_CTRL_STOP: > + cmd_work = c_node->stop_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + > + if (!wait_for_completion_timeout(&c_node->stop_complete, > + msecs_to_jiffies(300))) { > + DRM_ERROR("timeout stop:prop_id[%d]\n", > + c_node->property.prop_id); > + } > + > + c_node->state = IPP_STATE_STOP; > + ippdrv->dedicated = false; > + ipp_clean_cmd_node(c_node); > + > + if (list_empty(&ippdrv->cmd_list)) > + pm_runtime_put_sync(ippdrv->dev); > + break; > + case IPP_CTRL_PAUSE: > + cmd_work = c_node->stop_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + > + if (!wait_for_completion_timeout(&c_node->stop_complete, > + msecs_to_jiffies(200))) { > + DRM_ERROR("timeout stop:prop_id[%d]\n", > + c_node->property.prop_id); > + } > + > + c_node->state = IPP_STATE_STOP; > + break; > + case IPP_CTRL_RESUME: > + c_node->state = IPP_STATE_START; > + cmd_work = c_node->start_work; > + cmd_work->ctrl = cmd_ctrl->ctrl; > + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); > + break; > + default: > + /* ToDo: expand ctrl operation */ > + DRM_ERROR("could not support this state currently.\n"); > + goto err_clear; Just return error; > + } > + > + DRM_DEBUG_KMS("%s:done ctrl[%d]prop_id[%d]\n", __func__, > + cmd_ctrl->ctrl, cmd_ctrl->prop_id); > + > + return 0; > + > +err_clear: > + return ret; And remove the above label. > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ipp_cmd_ctrl); Remove it. > + > +int exynos_drm_ippnb_register(struct notifier_block *nb) > +{ > + return blocking_notifier_chain_register( > + &exynos_drm_ippnb_list, nb); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_register); Remove it. > + > +int exynos_drm_ippnb_unregister(struct notifier_block *nb) > +{ > + return blocking_notifier_chain_unregister( > + &exynos_drm_ippnb_list, nb); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_unregister); Remove it. > + > +int exynos_drm_ippnb_send_event(unsigned long val, void *v) > +{ > + return blocking_notifier_call_chain( > + &exynos_drm_ippnb_list, val, v); > +} > +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_send_event); Remove it. > + > +static int ipp_set_property(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_property *property) > +{ > + struct exynos_drm_ipp_ops *ops = NULL; > + int ret, i, swap = 0; > + > + if (!property) { > + DRM_ERROR("invalid property parameter.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* reset h/w block */ > + if (ippdrv->reset && > + ippdrv->reset(ippdrv->dev)) { > + DRM_ERROR("failed to reset.\n"); > + return -EINVAL; > + } > + > + /* set source,destination operations */ > + for_each_ipp_ops(i) { > + /* ToDo: integrate property and config */ > + struct drm_exynos_ipp_config *config = > + &property->config[i]; > + > + ops = ippdrv->ops[i]; > + if (!ops || !config) { > + DRM_ERROR("not support ops and config.\n"); > + return -EINVAL; > + } > + > + /* set format */ > + if (ops->set_fmt) { > + ret = ops->set_fmt(ippdrv->dev, config->fmt); > + if (ret) { > + DRM_ERROR("not support format.\n"); > + return ret; > + } > + } > + > + /* set transform for rotation, flip */ > + if (ops->set_transf) { > + swap = ops->set_transf(ippdrv->dev, config->degree, > + config->flip); > + if (swap < 0) { > + DRM_ERROR("not support tranf.\n"); > + return -EINVAL; > + } > + } > + > + /* set size */ > + if (ops->set_size) { > + ret = ops->set_size(ippdrv->dev, swap, &config->pos, > + &config->sz); > + if (ret) { > + DRM_ERROR("not support size.\n"); > + return ret; > + } > + } > + } > + > + return 0; > +} > + > +static int ipp_start_property(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct list_head *head; > + int ret, i; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* store command info in ippdrv */ > + ippdrv->cmd = c_node; > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return -ENOMEM; > + } > + > + /* set current property in ippdrv */ > + ret = ipp_set_property(ippdrv, property); > + if (ret) { > + DRM_ERROR("failed to set property.\n"); > + ippdrv->cmd = NULL; > + return ret; > + } > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("failed to get node.\n"); > + ret = -EFAULT; > + return ret; > + } > + > + DRM_DEBUG_KMS("%s:m_node[0x%x]\n", > + __func__, (int)m_node); > + > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + case IPP_CMD_WB: > + /* destination memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; > + > + list_for_each_entry(m_node, head, list) { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + list_for_each_entry(m_node, head, list) { > + ret = ipp_set_mem_node(ippdrv, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to set m node.\n"); > + return ret; > + } > + } > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + return -EINVAL; > + } > + > + DRM_DEBUG_KMS("%s:cmd[%d]\n", __func__, property->cmd); > + > + /* start operations */ > + if (ippdrv->start) { > + ret = ippdrv->start(ippdrv->dev, property->cmd); > + if (ret) { > + DRM_ERROR("failed to start ops.\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +static int ipp_stop_property(struct drm_device *drm_dev, > + struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node) > +{ > + struct drm_exynos_ipp_mem_node *m_node, *tm_node; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct list_head *head; > + int ret = 0, i; > + > + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); > + > + /* put event */ > + ipp_put_event(c_node, NULL); > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", > + __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, > + head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, > + m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + } > + break; > + case IPP_CMD_WB: > + /* destination memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + if (list_empty(head)) { > + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); > + break; > + } > + > + list_for_each_entry_safe(m_node, tm_node, head, list) { > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) { > + DRM_ERROR("failed to put m_node.\n"); > + goto err_clear; > + } > + } > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + ret = -EINVAL; > + goto err_clear; > + } > + > +err_clear: > + /* stop operations */ > + if (ippdrv->stop) > + ippdrv->stop(ippdrv->dev, property->cmd); > + > + return ret; > +} > + > +void ipp_sched_cmd(struct work_struct *work) > +{ > + struct drm_exynos_ipp_cmd_work *cmd_work = > + (struct drm_exynos_ipp_cmd_work *)work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + struct drm_exynos_ipp_property *property; > + int ret; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + ippdrv = cmd_work->ippdrv; > + if (!ippdrv) { > + DRM_ERROR("invalid ippdrv list.\n"); > + return; > + } > + > + c_node = cmd_work->c_node; > + if (!c_node) { > + DRM_ERROR("invalid command node list.\n"); > + return; > + } > + > + mutex_lock(&c_node->cmd_lock); > + > + property = &c_node->property; > + if (!property) { > + DRM_ERROR("failed to get property:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + > + switch (cmd_work->ctrl) { > + case IPP_CTRL_PLAY: > + case IPP_CTRL_RESUME: > + ret = ipp_start_property(ippdrv, c_node); > + if (ret) { > + DRM_ERROR("failed to start property:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + > + /* > + * M2M case supports wait_completion of transfer. > + * because M2M case supports single unit operation > + * with multiple queue. > + * M2M need to wait completion of data transfer. > + */ > + if (ipp_is_m2m_cmd(property->cmd)) { > + if (!wait_for_completion_timeout > + (&c_node->start_complete, msecs_to_jiffies(200))) { > + DRM_ERROR("timeout event:prop_id[%d]\n", > + c_node->property.prop_id); > + goto err_unlock; > + } > + } > + break; > + case IPP_CTRL_STOP: > + case IPP_CTRL_PAUSE: > + ret = ipp_stop_property(ippdrv->drm_dev, ippdrv, > + c_node); > + if (ret) { > + DRM_ERROR("failed to stop property.\n"); > + goto err_unlock; > + } > + > + complete(&c_node->stop_complete); > + break; > + default: > + DRM_ERROR("unknown control type\n"); > + break; > + } > + > + DRM_DEBUG_KMS("%s:ctrl[%d] done.\n", __func__, cmd_work->ctrl); > + > +err_unlock: > + mutex_unlock(&c_node->cmd_lock); > +} > + > +static int ipp_send_event(struct exynos_drm_ippdrv *ippdrv, > + struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) > +{ > + struct drm_device *drm_dev = ippdrv->drm_dev; > + struct drm_exynos_ipp_property *property = &c_node->property; > + struct drm_exynos_ipp_mem_node *m_node; > + struct drm_exynos_ipp_queue_buf qbuf; > + struct drm_exynos_ipp_send_event *e; > + struct list_head *head; > + struct timeval now; > + unsigned long flags; > + u32 tbuf_id[EXYNOS_DRM_OPS_MAX] = {0, }; > + int ret, i; > + > + for_each_ipp_ops(i) > + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, > + i ? "dst" : "src", buf_id[i]); > + > + if (!drm_dev) { > + DRM_ERROR("failed to get drm_dev.\n"); > + return -EINVAL; > + } > + > + if (!property) { > + DRM_ERROR("failed to get property.\n"); > + return -EINVAL; > + } > + > + if (list_empty(&c_node->event_list)) { > + DRM_DEBUG_KMS("%s:event list is empty.\n", __func__); > + return 0; > + } > + > + if (!ipp_check_mem_list(c_node)) { > + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); > + return 0; > + } > + > + /* check command */ > + switch (property->cmd) { > + case IPP_CMD_M2M: > + for_each_ipp_ops(i) { > + /* source/destination memory list */ > + head = &c_node->mem_list[i]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[i] = m_node->buf_id; > + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, > + i ? "dst" : "src", tbuf_id[i]); > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + } > + break; > + case IPP_CMD_WB: > + /* clear buf for finding */ > + memset(&qbuf, 0x0, sizeof(qbuf)); > + qbuf.ops_id = EXYNOS_DRM_OPS_DST; > + qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST]; > + > + /* get memory node entry */ > + m_node = ipp_find_mem_node(c_node, &qbuf); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id; > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + break; > + case IPP_CMD_OUTPUT: > + /* source memory list */ > + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; > + > + m_node = list_first_entry(head, > + struct drm_exynos_ipp_mem_node, list); > + if (!m_node) { > + DRM_ERROR("empty memory node.\n"); > + return -ENOMEM; > + } > + > + tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id; > + > + ret = ipp_put_mem_node(drm_dev, c_node, m_node); > + if (ret) > + DRM_ERROR("failed to put m_node.\n"); > + break; > + default: > + DRM_ERROR("invalid operations.\n"); > + return -EINVAL; > + } > + > + /* ToDo: Fix buffer id */ > + if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST]) > + DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n", > + tbuf_id[1], buf_id[1], property->prop_id); > + > + /* > + * command node have event list of destination buffer > + * If destination buffer enqueue to mem list, > + * then we make event and link to event list tail. > + * so, we get first event for first enqueued buffer. > + */ > + e = list_first_entry(&c_node->event_list, > + struct drm_exynos_ipp_send_event, base.link); > + > + if (!e) { > + DRM_ERROR("empty event.\n"); > + return -EINVAL; > + } > + > + do_gettimeofday(&now); > + DRM_DEBUG_KMS("%s:tv_sec[%ld]tv_usec[%ld]\n" > + , __func__, now.tv_sec, now.tv_usec); > + e->event.tv_sec = now.tv_sec; > + e->event.tv_usec = now.tv_usec; > + e->event.prop_id = property->prop_id; > + > + /* set buffer id about source destination */ > + for_each_ipp_ops(i) > + e->event.buf_id[i] = tbuf_id[i]; > + > + /* ToDo: compare index. If needed */ > + > + spin_lock_irqsave(&drm_dev->event_lock, flags); > + list_move_tail(&e->base.link, &e->base.file_priv->event_list); > + wake_up_interruptible(&e->base.file_priv->event_wait); > + spin_unlock_irqrestore(&drm_dev->event_lock, flags); > + > + DRM_DEBUG_KMS("%s:done cmd[%d]prop_id[%d]buf_id[%d]\n", __func__, > + property->cmd, property->prop_id, > tbuf_id[EXYNOS_DRM_OPS_DST]); > + > + return 0; > +} > + > +void ipp_sched_event(struct work_struct *work) > +{ > + struct drm_exynos_ipp_event_work *event_work = > + (struct drm_exynos_ipp_event_work *)work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + int ret; > + > + if (!event_work) { > + DRM_ERROR("failed to get event_work.\n"); > + return; > + } > + > + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, > + event_work->buf_id[EXYNOS_DRM_OPS_DST]); > + > + ippdrv = event_work->ippdrv; > + if (!ippdrv) { > + DRM_ERROR("failed to get ipp driver.\n"); > + return; > + } > + > + c_node = ippdrv->cmd; > + if (!c_node) { > + DRM_ERROR("failed to get command node.\n"); > + return; > + } > + > + /* > + * IPP supports command thread, event thread synchronization. > + * If IPP close immediately from user land, then IPP make > + * synchronization with command thread, so make complete event. > + * or going out operations. > + */ > + if (c_node->state != IPP_STATE_START) { > + DRM_DEBUG_KMS("%s:bypass state[%d]prop_id[%d]\n", > + __func__, c_node->state, c_node->property.prop_id); > + goto err_completion; > + } > + > + mutex_lock(&c_node->event_lock); > + > + ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); > + if (ret) { > + DRM_ERROR("failed to send event.\n"); > + goto err_completion; > + } > + > +err_completion: > + if (ipp_is_m2m_cmd(c_node->property.cmd)) > + complete(&c_node->start_complete); > + > + mutex_unlock(&c_node->event_lock); > +} > + > +static int ipp_subdrv_probe(struct drm_device *drm_dev, struct device > *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + struct exynos_drm_ippdrv *ippdrv; > + int ret, count = 0; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* get ipp driver entry */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + ippdrv->drm_dev = drm_dev; > + > + ret = ipp_create_id(&ctx->ipp_idr, &ctx->ipp_lock, ippdrv, > + &ippdrv->ipp_id); > + if (ret) { > + DRM_ERROR("failed to create id.\n"); > + goto err_idr; > + } > + > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]ipp_id[%d]\n", > __func__, > + count++, (int)ippdrv, ippdrv->ipp_id); > + > + if (ippdrv->ipp_id == 0) { > + DRM_ERROR("failed to get ipp_id[%d]\n", > + ippdrv->ipp_id); > + goto err_idr; > + } > + > + /* store parent device for node */ > + ippdrv->parent_dev = dev; > + > + /* store event work queue and handler */ > + ippdrv->event_workq = ctx->event_workq; > + ippdrv->sched_event = ipp_sched_event; > + INIT_LIST_HEAD(&ippdrv->cmd_list); > + } > + > + return 0; > + > +err_idr: > + idr_remove_all(&ctx->ipp_idr); > + idr_remove_all(&ctx->prop_idr); > + idr_destroy(&ctx->ipp_idr); > + idr_destroy(&ctx->prop_idr); > + return ret; > +} > + > +static void ipp_subdrv_remove(struct drm_device *drm_dev, struct device > *dev) > +{ > + struct exynos_drm_ippdrv *ippdrv; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* get ipp driver entry */ > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + ippdrv->drm_dev = NULL; > + exynos_drm_ippdrv_unregister(ippdrv); > + } > + > + /* ToDo: free notifier callback list if needed */ > +} > + > +static int ipp_subdrv_open(struct drm_device *drm_dev, struct device *dev, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* ToDo: multi device open */ > + > + priv = kzalloc(sizeof(*priv), GFP_KERNEL); > + if (!priv) { > + DRM_ERROR("failed to allocate priv.\n"); > + return -ENOMEM; > + } > + priv->dev = dev; > + file_priv->ipp_priv = priv; > + > + INIT_LIST_HEAD(&priv->event_list); > + > + DRM_DEBUG_KMS("%s:done priv[0x%x]\n", __func__, (int)priv); > + > + return 0; > +} > + > +static void ipp_subdrv_close(struct drm_device *drm_dev, struct device > *dev, > + struct drm_file *file) > +{ > + struct drm_exynos_file_private *file_priv = file->driver_priv; > + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; > + struct exynos_drm_ippdrv *ippdrv = NULL; > + struct drm_exynos_ipp_cmd_node *c_node, *tc_node; > + int count = 0; > + > + DRM_DEBUG_KMS("%s:for priv[0x%x]\n", __func__, (int)priv); > + > + if (list_empty(&exynos_drm_ippdrv_list)) { > + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); > + goto err_clear; > + } > + > + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { > + if (list_empty(&ippdrv->cmd_list)) > + continue; > + > + list_for_each_entry_safe(c_node, tc_node, > + &ippdrv->cmd_list, list) { > + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", > + __func__, count++, (int)ippdrv); > + > + if (c_node->priv == priv) { > + /* > + * userland goto unnormal state. process killed. > + * and close the file. > + * so, IPP didn't called stop cmd ctrl. > + * so, we are make stop operation in this state. > + */ > + if (c_node->state == IPP_STATE_START) { > + ipp_stop_property(drm_dev, ippdrv, > + c_node); > + c_node->state = IPP_STATE_STOP; > + } > + > + ippdrv->dedicated = false; > + ipp_clean_cmd_node(c_node); > + if (list_empty(&ippdrv->cmd_list)) > + pm_runtime_put_sync(ippdrv->dev); > + } > + } > + } > + > +err_clear: > + kfree(priv); > + return; > +} > + > +static int __devinit ipp_probe(struct platform_device *pdev) > +{ > + struct device *dev = &pdev->dev; > + struct ipp_context *ctx; > + struct exynos_drm_subdrv *subdrv; > + int ret; > + > + ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); > + if (!ctx) > + return -ENOMEM; > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + mutex_init(&ctx->ipp_lock); > + mutex_init(&ctx->prop_lock); > + > + idr_init(&ctx->ipp_idr); > + idr_init(&ctx->prop_idr); > + > + /* > + * create single thread for ipp event > + * IPP supports event thread for IPP drivers. > + * IPP driver send event_work to this thread. > + * and IPP event thread send event to user process. > + */ > + ctx->event_workq = create_singlethread_workqueue("ipp_event"); > + if (!ctx->event_workq) { > + dev_err(dev, "failed to create event workqueue\n"); > + ret = -EINVAL; > + goto err_clear; > + } > + > + /* > + * create single thread for ipp command > + * IPP supports command thread for user process. > + * user process make command node using set property ioctl. > + * and make start_work and send this work to command thread. > + * and then this command thread start property. > + */ > + ctx->cmd_workq = create_singlethread_workqueue("ipp_cmd"); > + if (!ctx->cmd_workq) { > + dev_err(dev, "failed to create cmd workqueue\n"); > + ret = -EINVAL; > + goto err_event_workq; > + } > + > + /* set sub driver informations */ > + subdrv = &ctx->subdrv; > + subdrv->dev = dev; > + subdrv->probe = ipp_subdrv_probe; > + subdrv->remove = ipp_subdrv_remove; > + subdrv->open = ipp_subdrv_open; > + subdrv->close = ipp_subdrv_close; > + > + platform_set_drvdata(pdev, ctx); > + > + ret = exynos_drm_subdrv_register(subdrv); > + if (ret < 0) { > + DRM_ERROR("failed to register drm ipp device.\n"); > + goto err_cmd_workq; > + } > + > + dev_info(&pdev->dev, "drm ipp registered successfully.\n"); > + > + return 0; > + > +err_cmd_workq: > + destroy_workqueue(ctx->cmd_workq); > +err_event_workq: > + destroy_workqueue(ctx->event_workq); > +err_clear: > + kfree(ctx); > + return ret; > +} > + > +static int __devexit ipp_remove(struct platform_device *pdev) > +{ > + struct ipp_context *ctx = platform_get_drvdata(pdev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + /* unregister sub driver */ > + exynos_drm_subdrv_unregister(&ctx->subdrv); > + > + /* remove,destroy ipp idr */ > + idr_remove_all(&ctx->ipp_idr); > + idr_remove_all(&ctx->prop_idr); > + idr_destroy(&ctx->ipp_idr); > + idr_destroy(&ctx->prop_idr); > + > + mutex_destroy(&ctx->ipp_lock); > + mutex_destroy(&ctx->prop_lock); > + > + /* destroy command, event work queue */ > + destroy_workqueue(ctx->cmd_workq); > + destroy_workqueue(ctx->event_workq); > + > + kfree(ctx); > + > + return 0; > +} > + > +static int ipp_power_ctrl(struct ipp_context *ctx, bool enable) > +{ > + /* ToDo: Need to implement power and sysmmu ctrl. */ > + DRM_DEBUG_KMS("%s:enable[%d]\n", __func__, enable); > + > + return 0; > +} > + > +#ifdef CONFIG_PM_SLEEP > +static int ipp_suspend(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (pm_runtime_suspended(dev)) > + return 0; > + > + return ipp_power_ctrl(ctx, false); > +} > + > +static int ipp_resume(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + if (!pm_runtime_suspended(dev)) > + return ipp_power_ctrl(ctx, true); > + > + return 0; > +} > +#endif > + > +#ifdef CONFIG_PM_RUNTIME > +static int ipp_runtime_suspend(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + return ipp_power_ctrl(ctx, false); > +} > + > +static int ipp_runtime_resume(struct device *dev) > +{ > + struct ipp_context *ctx = get_ipp_context(dev); > + > + DRM_DEBUG_KMS("%s\n", __func__); > + > + return ipp_power_ctrl(ctx, true); > +} > +#endif > + > +static const struct dev_pm_ops ipp_pm_ops = { > + SET_SYSTEM_SLEEP_PM_OPS(ipp_suspend, ipp_resume) > + SET_RUNTIME_PM_OPS(ipp_runtime_suspend, ipp_runtime_resume, NULL) > +}; > + > +struct platform_driver ipp_driver = { > + .probe = ipp_probe, > + .remove = __devexit_p(ipp_remove), > + .driver = { > + .name = "exynos-drm-ipp", > + .owner = THIS_MODULE, > + .pm = &ipp_pm_ops, > + }, > +}; > + > diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.h > b/drivers/gpu/drm/exynos/exynos_drm_ipp.h > new file mode 100644 > index 0000000..baab1f0 > --- /dev/null > +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.h > @@ -0,0 +1,266 @@ > +/* > + * Copyright (c) 2012 Samsung Electronics Co., Ltd. > + * > + * Authors: > + * Eunchul Kim <chulspro.kim@samsung.com> > + * Jinyoung Jeon <jy0.jeon@samsung.com> > + * Sangmin Lee <lsmin.lee@samsung.com> > + * > + * Permission is hereby granted, free of charge, to any person obtaining > a > + * copy of this software and associated documentation files (the > "Software"), > + * to deal in the Software without restriction, including without > limitation > + * the rights to use, copy, modify, merge, publish, distribute, > sublicense, > + * and/or sell copies of the Software, and to permit persons to whom the > + * Software is furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice (including the > next > + * paragraph) shall be included in all copies or substantial portions of > the > + * Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, > EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF > MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT > SHALL > + * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES > OR > + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, > + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR > + * OTHER DEALINGS IN THE SOFTWARE. > + */ > + > +#ifndef _EXYNOS_DRM_IPP_H_ > +#define _EXYNOS_DRM_IPP_H_ > + > +#define for_each_ipp_ops(pos) \ > + for (pos = 0; pos < EXYNOS_DRM_OPS_MAX; pos++) > +#define for_each_ipp_planar(pos) \ > + for (pos = 0; pos < EXYNOS_DRM_PLANAR_MAX; pos++) > + > +#define IPP_GET_LCD_WIDTH _IOR('F', 302, int) > +#define IPP_GET_LCD_HEIGHT _IOR('F', 303, int) > +#define IPP_SET_WRITEBACK _IOW('F', 304, u32) > + > +/* definition of state */ > +enum drm_exynos_ipp_state { > + IPP_STATE_IDLE, > + IPP_STATE_START, > + IPP_STATE_STOP, > +}; > + > +/* > + * A structure of command work information. > + * @work: work structure. > + * @ippdrv: current work ippdrv. > + * @c_node: command node information. > + * @ctrl: command control. > + */ > +struct drm_exynos_ipp_cmd_work { > + struct work_struct work; > + struct exynos_drm_ippdrv *ippdrv; > + struct drm_exynos_ipp_cmd_node *c_node; > + enum drm_exynos_ipp_ctrl ctrl; > +}; > + > +/* > + * A structure of command node. > + * > + * @priv: IPP private infomation. > + * @list: list head to command queue information. > + * @event_list: list head of event. > + * @mem_list: list head to source,destination memory queue information. > + * @cmd_lock: lock for synchronization of access to ioctl. > + * @mem_lock: lock for synchronization of access to memory nodes. > + * @event_lock: lock for synchronization of access to scheduled event. > + * @start_complete: completion of start of command. > + * @stop_complete: completion of stop of command. > + * @property: property information. > + * @start_work: start command work structure. > + * @stop_work: stop command work structure. > + * @event_work: event work structure. > + * @state: state of command node. > + */ > +struct drm_exynos_ipp_cmd_node { > + struct exynos_drm_ipp_private *priv; > + struct list_head list; > + struct list_head event_list; > + struct list_head mem_list[EXYNOS_DRM_OPS_MAX]; > + struct mutex cmd_lock; > + struct mutex mem_lock; > + struct mutex event_lock; > + struct completion start_complete; > + struct completion stop_complete; > + struct drm_exynos_ipp_property property; > + struct drm_exynos_ipp_cmd_work *start_work; > + struct drm_exynos_ipp_cmd_work *stop_work; > + struct drm_exynos_ipp_event_work *event_work; > + enum drm_exynos_ipp_state state; > +}; > + > +/* > + * A structure of buffer information. > + * > + * @gem_objs: Y, Cb, Cr each gem object. > + * @base: Y, Cb, Cr each planar address. > + */ > +struct drm_exynos_ipp_buf_info { > + unsigned long handles[EXYNOS_DRM_PLANAR_MAX]; > + dma_addr_t base[EXYNOS_DRM_PLANAR_MAX]; > +}; > + > +/* > + * A structure of wb setting infomation. > + * > + * @enable: enable flag for wb. > + * @refresh: HZ of the refresh rate. > + */ > +struct drm_exynos_ipp_set_wb { > + __u32 enable; > + __u32 refresh; > +}; > + > +/* > + * A structure of event work information. > + * > + * @work: work structure. > + * @ippdrv: current work ippdrv. > + * @buf_id: id of src, dst buffer. > + */ > +struct drm_exynos_ipp_event_work { > + struct work_struct work; > + struct exynos_drm_ippdrv *ippdrv; > + u32 buf_id[EXYNOS_DRM_OPS_MAX]; > +}; > + > +/* > + * A structure of source,destination operations. > + * > + * @set_fmt: set format of image. > + * @set_transf: set transform(rotations, flip). > + * @set_size: set size of region. > + * @set_addr: set address for dma. > + */ > +struct exynos_drm_ipp_ops { > + int (*set_fmt)(struct device *dev, u32 fmt); > + int (*set_transf)(struct device *dev, > + enum drm_exynos_degree degree, > + enum drm_exynos_flip flip); > + int (*set_size)(struct device *dev, int swap, > + struct drm_exynos_pos *pos, struct drm_exynos_sz *sz); > + int (*set_addr)(struct device *dev, > + struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, > + enum drm_exynos_ipp_buf_type buf_type); > +}; > + > +/* > + * A structure of ipp driver. > + * > + * @drv_list: list head for registed sub driver information. > + * @parent_dev: parent device information. > + * @dev: platform device. > + * @drm_dev: drm device. > + * @ipp_id: id of ipp driver. > + * @dedicated: dedicated ipp device. > + * @ops: source, destination operations. > + * @event_workq: event work queue. > + * @cmd: current command information. > + * @cmd_list: list head for command information. > + * @prop_list: property informations of current ipp driver. > + * @check_property: check property about format, size, buffer. > + * @reset: reset ipp block. > + * @start: ipp each device start. > + * @stop: ipp each device stop. > + * @sched_event: work schedule handler. > + */ > +struct exynos_drm_ippdrv { > + struct list_head drv_list; > + struct device *parent_dev; > + struct device *dev; > + struct drm_device *drm_dev; > + u32 ipp_id; > + bool dedicated; > + struct exynos_drm_ipp_ops *ops[EXYNOS_DRM_OPS_MAX]; > + struct workqueue_struct *event_workq; > + struct drm_exynos_ipp_cmd_node *cmd; > + struct list_head cmd_list; > + struct drm_exynos_ipp_prop_list *prop_list; > + > + int (*check_property)(struct device *dev, > + struct drm_exynos_ipp_property *property); > + int (*reset)(struct device *dev); > + int (*start)(struct device *dev, enum drm_exynos_ipp_cmd cmd); > + void (*stop)(struct device *dev, enum drm_exynos_ipp_cmd cmd); > + void (*sched_event)(struct work_struct *work); > +}; > + > +#ifdef CONFIG_DRM_EXYNOS_IPP > +extern int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv); > +extern int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv); > +extern int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void > *data, > + struct drm_file *file); > +extern int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, > + struct drm_file *file); > +extern int exynos_drm_ippnb_register(struct notifier_block *nb); > +extern int exynos_drm_ippnb_unregister(struct notifier_block *nb); > +extern int exynos_drm_ippnb_send_event(unsigned long val, void *v); > +extern void ipp_sched_cmd(struct work_struct *work); > +extern void ipp_sched_event(struct work_struct *work); > + > +#else > +static inline int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv > *ippdrv) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv > *ippdrv) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ipp_get_property(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file_priv) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_set_property(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file_priv) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, > + void *data, > + struct drm_file *file) > +{ > + return -ENOTTY; > +} > + > +static inline int exynos_drm_ippnb_register(struct notifier_block *nb) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippnb_unregister(struct notifier_block *nb) > +{ > + return -ENODEV; > +} > + > +static inline int exynos_drm_ippnb_send_event(unsigned long val, void *v) > +{ > + return -ENOTTY; > +} > +#endif > + > +#endif /* _EXYNOS_DRM_IPP_H_ */ > + > diff --git a/include/uapi/drm/exynos_drm.h b/include/uapi/drm/exynos_drm.h > index c0494d5..82772d7 100644 > --- a/include/uapi/drm/exynos_drm.h > +++ b/include/uapi/drm/exynos_drm.h > @@ -154,6 +154,170 @@ struct drm_exynos_g2d_exec { > __u64 async; > }; > > +enum drm_exynos_ops_id { > + EXYNOS_DRM_OPS_SRC, > + EXYNOS_DRM_OPS_DST, > + EXYNOS_DRM_OPS_MAX, > +}; > + > +struct drm_exynos_sz { > + __u32 hsize; > + __u32 vsize; > +}; > + > +struct drm_exynos_pos { > + __u32 x; > + __u32 y; > + __u32 w; > + __u32 h; > +}; > + > +enum drm_exynos_flip { > + EXYNOS_DRM_FLIP_NONE = (0 << 0), > + EXYNOS_DRM_FLIP_VERTICAL = (1 << 0), > + EXYNOS_DRM_FLIP_HORIZONTAL = (1 << 1), > +}; > + > +enum drm_exynos_degree { > + EXYNOS_DRM_DEGREE_0, > + EXYNOS_DRM_DEGREE_90, > + EXYNOS_DRM_DEGREE_180, > + EXYNOS_DRM_DEGREE_270, > +}; > + > +enum drm_exynos_planer { > + EXYNOS_DRM_PLANAR_Y, > + EXYNOS_DRM_PLANAR_CB, > + EXYNOS_DRM_PLANAR_CR, > + EXYNOS_DRM_PLANAR_MAX, > +}; > + > +/** > + * A structure for ipp supported property list. > + * > + * @version: version of this structure. > + * @ipp_id: id of ipp driver. > + * @count: count of ipp driver. > + * @writeback: flag of writeback supporting. > + * @flip: flag of flip supporting. > + * @degree: flag of degree information. > + * @csc: flag of csc supporting. > + * @crop: flag of crop supporting. > + * @scale: flag of scale supporting. > + * @refresh_min: min hz of refresh. > + * @refresh_max: max hz of refresh. > + * @crop_min: crop min resolution. > + * @crop_max: crop max resolution. > + * @scale_min: scale min resolution. > + * @scale_max: scale max resolution. > + */ > +struct drm_exynos_ipp_prop_list { > + __u32 version; > + __u32 ipp_id; > + __u32 count; > + __u32 writeback; > + __u32 flip; > + __u32 degree; > + __u32 csc; > + __u32 crop; > + __u32 scale; > + __u32 refresh_min; > + __u32 refresh_max; > + __u32 reserved; > + struct drm_exynos_sz crop_min; > + struct drm_exynos_sz crop_max; > + struct drm_exynos_sz scale_min; > + struct drm_exynos_sz scale_max; > +}; > + > +/** > + * A structure for ipp config. > + * > + * @ops_id: property of operation directions. > + * @flip: property of mirror, flip. > + * @degree: property of rotation degree. > + * @fmt: property of image format. > + * @sz: property of image size. > + * @pos: property of image position(src-cropped,dst-scaler). > + */ > +struct drm_exynos_ipp_config { > + enum drm_exynos_ops_id ops_id; > + enum drm_exynos_flip flip; > + enum drm_exynos_degree degree; > + __u32 fmt; > + struct drm_exynos_sz sz; > + struct drm_exynos_pos pos; > +}; > + > +enum drm_exynos_ipp_cmd { > + IPP_CMD_NONE, > + IPP_CMD_M2M, > + IPP_CMD_WB, > + IPP_CMD_OUTPUT, > + IPP_CMD_MAX, > +}; > + > +/** > + * A structure for ipp property. > + * > + * @config: source, destination config. > + * @cmd: definition of command. > + * @ipp_id: id of ipp driver. > + * @prop_id: id of property. > + * @refresh_rate: refresh rate. > + */ > +struct drm_exynos_ipp_property { > + struct drm_exynos_ipp_config config[EXYNOS_DRM_OPS_MAX]; > + enum drm_exynos_ipp_cmd cmd; > + __u32 ipp_id; > + __u32 prop_id; > + __u32 refresh_rate; > +}; > + > +enum drm_exynos_ipp_buf_type { > + IPP_BUF_ENQUEUE, > + IPP_BUF_DEQUEUE, > +}; > + > +/** > + * A structure for ipp buffer operations. > + * > + * @ops_id: operation directions. > + * @buf_type: definition of buffer. > + * @prop_id: id of property. > + * @buf_id: id of buffer. > + * @handle: Y, Cb, Cr each planar handle. > + * @user_data: user data. > + */ > +struct drm_exynos_ipp_queue_buf { > + enum drm_exynos_ops_id ops_id; > + enum drm_exynos_ipp_buf_type buf_type; > + __u32 prop_id; > + __u32 buf_id; > + __u32 handle[EXYNOS_DRM_PLANAR_MAX]; > + __u32 reserved; > + __u64 user_data; > +}; > + > +enum drm_exynos_ipp_ctrl { > + IPP_CTRL_PLAY, > + IPP_CTRL_STOP, > + IPP_CTRL_PAUSE, > + IPP_CTRL_RESUME, > + IPP_CTRL_MAX, > +}; > + > +/** > + * A structure for ipp start/stop operations. > + * > + * @prop_id: id of property. > + * @ctrl: definition of control. > + */ > +struct drm_exynos_ipp_cmd_ctrl { > + __u32 prop_id; > + enum drm_exynos_ipp_ctrl ctrl; > +}; > + > #define DRM_EXYNOS_GEM_CREATE 0x00 > #define DRM_EXYNOS_GEM_MAP_OFFSET 0x01 > #define DRM_EXYNOS_GEM_MMAP 0x02 > @@ -166,6 +330,12 @@ struct drm_exynos_g2d_exec { > #define DRM_EXYNOS_G2D_SET_CMDLIST 0x21 > #define DRM_EXYNOS_G2D_EXEC 0x22 > > +/* IPP - Image Post Processing */ > +#define DRM_EXYNOS_IPP_GET_PROPERTY 0x30 > +#define DRM_EXYNOS_IPP_SET_PROPERTY 0x31 > +#define DRM_EXYNOS_IPP_QUEUE_BUF 0x32 > +#define DRM_EXYNOS_IPP_CMD_CTRL 0x33 > + > #define DRM_IOCTL_EXYNOS_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + > \ > DRM_EXYNOS_GEM_CREATE, struct drm_exynos_gem_create) > > @@ -188,8 +358,18 @@ struct drm_exynos_g2d_exec { > #define DRM_IOCTL_EXYNOS_G2D_EXEC DRM_IOWR(DRM_COMMAND_BASE + > \ > DRM_EXYNOS_G2D_EXEC, struct drm_exynos_g2d_exec) > > +#define DRM_IOCTL_EXYNOS_IPP_GET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_GET_PROPERTY, struct drm_exynos_ipp_prop_list) > +#define DRM_IOCTL_EXYNOS_IPP_SET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_SET_PROPERTY, struct drm_exynos_ipp_property) > +#define DRM_IOCTL_EXYNOS_IPP_QUEUE_BUF DRM_IOWR(DRM_COMMAND_BASE + \ > + DRM_EXYNOS_IPP_QUEUE_BUF, struct drm_exynos_ipp_queue_buf) > +#define DRM_IOCTL_EXYNOS_IPP_CMD_CTRL DRM_IOWR(DRM_COMMAND_BASE + > \ > + DRM_EXYNOS_IPP_CMD_CTRL, struct drm_exynos_ipp_cmd_ctrl) > + > /* EXYNOS specific events */ > #define DRM_EXYNOS_G2D_EVENT 0x80000000 > +#define DRM_EXYNOS_IPP_EVENT 0x80000001 > > struct drm_exynos_g2d_event { > struct drm_event base; > @@ -200,4 +380,14 @@ struct drm_exynos_g2d_event { > __u32 reserved; > }; > > +struct drm_exynos_ipp_event { > + struct drm_event base; > + __u64 user_data; > + __u32 tv_sec; > + __u32 tv_usec; > + __u32 prop_id; > + __u32 reserved; > + __u32 buf_id[EXYNOS_DRM_OPS_MAX]; > +}; > + > #endif /* _UAPI_EXYNOS_DRM_H_ */ > -- > 1.7.0.4
Thank's your comment. I anwer your comment, please check that. Thank's BR Eunchul Kim. On 12/12/2012 05:18 PM, Inki Dae wrote: > > >> -----Original Message----- >> From: Eunchul Kim [mailto:chulspro.kim@samsung.com] >> Sent: Wednesday, December 12, 2012 4:35 PM >> To: dri-devel@lists.freedesktop.org; inki.dae@samsung.com >> Cc: jy0.jeon@samsung.com; yj44.cho@samsung.com; jmock.shin@samsung.com; >> jaejoon.seo@samsung.com; kyungmin.park@samsung.com; >> chulspro.kim@samsung.com >> Subject: [RFC v3 1/5] drm/exynos: add ipp subsystem >> >> IPP stand for Image Post Processing and supports image scaler/rotator >> /crop/flip/csc(color space conversion) and input/output DMA operations >> using ipp drivers. >> also supports writeback and display output operations. >> ipp driver include FIMC, Rotator, GSC, SC, so on. >> and ipp is integration device driver for each hardware. >> >> Signed-off-by: Eunchul Kim <chulspro.kim@samsung.com> >> Signed-off-by: Jinyoung Jeon <jy0.jeon@samsung.com> >> --- >> drivers/gpu/drm/exynos/Kconfig | 6 + >> drivers/gpu/drm/exynos/Makefile | 1 + >> drivers/gpu/drm/exynos/exynos_drm_drv.c | 24 + >> drivers/gpu/drm/exynos/exynos_drm_drv.h | 7 + >> drivers/gpu/drm/exynos/exynos_drm_ipp.c | 2059 >> +++++++++++++++++++++++++++++++ >> drivers/gpu/drm/exynos/exynos_drm_ipp.h | 266 ++++ >> include/uapi/drm/exynos_drm.h | 190 +++ >> 7 files changed, 2553 insertions(+), 0 deletions(-) >> create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.c >> create mode 100644 drivers/gpu/drm/exynos/exynos_drm_ipp.h >> >> diff --git a/drivers/gpu/drm/exynos/Kconfig >> b/drivers/gpu/drm/exynos/Kconfig >> index 4ea8cdc..bcf1c9d 100644 >> --- a/drivers/gpu/drm/exynos/Kconfig >> +++ b/drivers/gpu/drm/exynos/Kconfig >> @@ -45,3 +45,9 @@ config DRM_EXYNOS_G2D >> depends on DRM_EXYNOS && !VIDEO_SAMSUNG_S5P_G2D >> help >> Choose this option if you want to use Exynos G2D for DRM. >> + >> +config DRM_EXYNOS_IPP >> + bool "Exynos DRM IPP" >> + depends on DRM_EXYNOS >> + help >> + Choose this option if you want to use IPP feature for DRM. >> diff --git a/drivers/gpu/drm/exynos/Makefile >> b/drivers/gpu/drm/exynos/Makefile >> index 26813b8..6c536ce 100644 >> --- a/drivers/gpu/drm/exynos/Makefile >> +++ b/drivers/gpu/drm/exynos/Makefile >> @@ -16,5 +16,6 @@ exynosdrm-$(CONFIG_DRM_EXYNOS_HDMI) += >> exynos_hdmi.o exynos_mixer.o \ >> exynos_drm_hdmi.o >> exynosdrm-$(CONFIG_DRM_EXYNOS_VIDI) += exynos_drm_vidi.o >> exynosdrm-$(CONFIG_DRM_EXYNOS_G2D) += exynos_drm_g2d.o >> +exynosdrm-$(CONFIG_DRM_EXYNOS_IPP) += exynos_drm_ipp.o >> >> obj-$(CONFIG_DRM_EXYNOS) += exynosdrm.o >> diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.c >> b/drivers/gpu/drm/exynos/exynos_drm_drv.c >> index 4a1168d..0eb8a97 100644 >> --- a/drivers/gpu/drm/exynos/exynos_drm_drv.c >> +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.c >> @@ -40,6 +40,7 @@ >> #include "exynos_drm_vidi.h" >> #include "exynos_drm_dmabuf.h" >> #include "exynos_drm_g2d.h" >> +#include "exynos_drm_ipp.h" >> #include "exynos_drm_iommu.h" >> >> #define DRIVER_NAME "exynos" >> @@ -249,6 +250,14 @@ static struct drm_ioctl_desc exynos_ioctls[] = { >> exynos_g2d_set_cmdlist_ioctl, DRM_UNLOCKED | > DRM_AUTH), >> DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, >> exynos_g2d_exec_ioctl, DRM_UNLOCKED | DRM_AUTH), >> + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, >> + exynos_drm_ipp_get_property, DRM_UNLOCKED | > DRM_AUTH), >> + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, >> + exynos_drm_ipp_set_property, DRM_UNLOCKED | > DRM_AUTH), >> + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, >> + exynos_drm_ipp_queue_buf, DRM_UNLOCKED | DRM_AUTH), >> + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, >> + exynos_drm_ipp_cmd_ctrl, DRM_UNLOCKED | DRM_AUTH), >> }; >> >> static const struct file_operations exynos_drm_driver_fops = { >> @@ -363,6 +372,12 @@ static int __init exynos_drm_init(void) >> goto out_g2d; >> #endif >> >> +#ifdef CONFIG_DRM_EXYNOS_IPP >> + ret = platform_driver_register(&ipp_driver); >> + if (ret < 0) >> + goto out_ipp; >> +#endif >> + >> ret = platform_driver_register(&exynos_drm_platform_driver); >> if (ret < 0) >> goto out_drm; >> @@ -380,6 +395,11 @@ out: >> platform_driver_unregister(&exynos_drm_platform_driver); >> >> out_drm: >> +#ifdef CONFIG_DRM_EXYNOS_IPP >> + platform_driver_unregister(&ipp_driver); >> +out_ipp: >> +#endif >> + >> #ifdef CONFIG_DRM_EXYNOS_G2D >> platform_driver_unregister(&g2d_driver); >> out_g2d: >> @@ -416,6 +436,10 @@ static void __exit exynos_drm_exit(void) >> >> platform_driver_unregister(&exynos_drm_platform_driver); >> >> +#ifdef CONFIG_DRM_EXYNOS_IPP >> + platform_driver_unregister(&ipp_driver); >> +#endif >> + >> #ifdef CONFIG_DRM_EXYNOS_G2D >> platform_driver_unregister(&g2d_driver); >> #endif >> diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.h >> b/drivers/gpu/drm/exynos/exynos_drm_drv.h >> index a9db025..a365788 100644 >> --- a/drivers/gpu/drm/exynos/exynos_drm_drv.h >> +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.h >> @@ -235,8 +235,14 @@ struct exynos_drm_g2d_private { >> unsigned int gem_nr; >> }; >> >> +struct exynos_drm_ipp_private { >> + struct device *dev; >> + struct list_head event_list; >> +}; >> + >> struct drm_exynos_file_private { >> struct exynos_drm_g2d_private *g2d_priv; >> + struct exynos_drm_ipp_private *ipp_priv; >> }; >> >> /* >> @@ -346,4 +352,5 @@ extern struct platform_driver mixer_driver; >> extern struct platform_driver exynos_drm_common_hdmi_driver; >> extern struct platform_driver vidi_driver; >> extern struct platform_driver g2d_driver; >> +extern struct platform_driver ipp_driver; >> #endif >> diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.c >> b/drivers/gpu/drm/exynos/exynos_drm_ipp.c >> new file mode 100644 >> index 0000000..eb12cb5 >> --- /dev/null >> +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.c >> @@ -0,0 +1,2059 @@ >> +/* >> + * Copyright (C) 2012 Samsung Electronics Co.Ltd >> + * Authors: >> + * Eunchul Kim <chulspro.kim@samsung.com> >> + * Jinyoung Jeon <jy0.jeon@samsung.com> >> + * Sangmin Lee <lsmin.lee@samsung.com> >> + * >> + * This program is free software; you can redistribute it and/or modify >> it >> + * under the terms of the GNU General Public License as published by >> the >> + * Free Software Foundation; either version 2 of the License, or (at >> your >> + * option) any later version. >> + * >> + */ >> +#include <linux/kernel.h> >> +#include <linux/module.h> >> +#include <linux/platform_device.h> >> +#include <linux/clk.h> >> +#include <linux/pm_runtime.h> >> +#include <plat/map-base.h> >> + >> +#include <drm/drmP.h> >> +#include <drm/exynos_drm.h> >> +#include "exynos_drm_drv.h" >> +#include "exynos_drm_gem.h" >> +#include "exynos_drm_ipp.h" >> + >> +/* >> + * IPP is stand for Image Post Processing and >> + * supports image scaler/rotator and input/output DMA operations. >> + * using FIMC, GSC, Rotator, so on. >> + * IPP is integration device driver of same attribute h/w >> + */ >> + >> +#define get_ipp_context(dev) >> platform_get_drvdata(to_platform_device(dev)) >> +#define ipp_is_m2m_cmd(c) (c == IPP_CMD_M2M) >> + >> +/* >> + * A structure of event. >> + * >> + * @base: base of event. >> + * @event: ipp event. >> + */ >> +struct drm_exynos_ipp_send_event { >> + struct drm_pending_event base; >> + struct drm_exynos_ipp_event event; >> +}; >> + >> +/* >> + * A structure of memory node. >> + * >> + * @list: list head to memory queue information. >> + * @ops_id: id of operations. >> + * @prop_id: id of property. >> + * @buf_id: id of buffer. >> + * @buf_info: gem objects and dma address, size. >> + * @filp: a pointer to drm_file. >> + */ >> +struct drm_exynos_ipp_mem_node { >> + struct list_head list; >> + enum drm_exynos_ops_id ops_id; >> + u32 prop_id; >> + u32 buf_id; >> + struct drm_exynos_ipp_buf_info buf_info; >> + struct drm_file *filp; >> +}; >> + >> +/* >> + * A structure of ipp context. >> + * >> + * @subdrv: prepare initialization using subdrv. >> + * @ipp_lock: lock for synchronization of access to ipp_idr. >> + * @prop_lock: lock for synchronization of access to prop_idr. >> + * @ipp_idr: ipp driver idr. >> + * @prop_idr: property idr. >> + * @event_workq: event work queue. >> + * @cmd_workq: command work queue. >> + */ >> +struct ipp_context { >> + struct exynos_drm_subdrv subdrv; >> + struct mutex ipp_lock; >> + struct mutex prop_lock; >> + struct idr ipp_idr; >> + struct idr prop_idr; >> + struct workqueue_struct *event_workq; >> + struct workqueue_struct *cmd_workq; >> +}; >> + >> +static LIST_HEAD(exynos_drm_ippdrv_list); >> +static DEFINE_MUTEX(exynos_drm_ippdrv_lock); >> +static BLOCKING_NOTIFIER_HEAD(exynos_drm_ippnb_list); >> + >> +int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv) >> +{ >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!ippdrv) >> + return -EINVAL; >> + >> + mutex_lock(&exynos_drm_ippdrv_lock); >> + list_add_tail(&ippdrv->drv_list, &exynos_drm_ippdrv_list); >> + mutex_unlock(&exynos_drm_ippdrv_lock); >> + >> + return 0; >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_register); >> + >> +int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv) >> +{ >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!ippdrv) >> + return -EINVAL; >> + >> + mutex_lock(&exynos_drm_ippdrv_lock); >> + list_del(&ippdrv->drv_list); >> + mutex_unlock(&exynos_drm_ippdrv_lock); >> + >> + return 0; >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_unregister); >> + >> +static int ipp_create_id(struct idr *id_idr, struct mutex *lock, void >> *obj, >> + u32 *idp) >> +{ >> + int ret; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> +again: >> + /* ensure there is space available to allocate a handle */ >> + if (idr_pre_get(id_idr, GFP_KERNEL) == 0) { >> + DRM_ERROR("failed to get idr.\n"); >> + return -ENOMEM; >> + } >> + >> + /* do the allocation under our mutexlock */ >> + mutex_lock(lock); >> + ret = idr_get_new_above(id_idr, obj, 1, (int *)idp); >> + mutex_unlock(lock); >> + if (ret == -EAGAIN) >> + goto again; >> + >> + return ret; >> +} >> + >> +static void *ipp_find_obj(struct idr *id_idr, struct mutex *lock, u32 id) >> +{ >> + void *obj; >> + >> + DRM_DEBUG_KMS("%s:id[%d]\n", __func__, id); >> + >> + mutex_lock(lock); >> + >> + /* find object using handle */ >> + obj = idr_find(id_idr, id); >> + if (!obj) { >> + DRM_ERROR("failed to find object.\n"); >> + mutex_unlock(lock); >> + return ERR_PTR(-ENODEV); >> + } >> + >> + mutex_unlock(lock); >> + >> + return obj; >> +} >> + >> +static inline bool ipp_check_dedicated(struct exynos_drm_ippdrv *ippdrv, >> + enum drm_exynos_ipp_cmd cmd) >> +{ >> + /* >> + * check dedicated flag and WB, OUTPUT operation with >> + * power on state. >> + */ >> + if (ippdrv->dedicated || (!ipp_is_m2m_cmd(cmd) && >> + !pm_runtime_suspended(ippdrv->dev))) >> + return true; >> + >> + return false; >> +} >> + >> +static struct exynos_drm_ippdrv *ipp_find_driver(struct ipp_context *ctx, >> + struct drm_exynos_ipp_property *property) >> +{ >> + struct exynos_drm_ippdrv *ippdrv; >> + u32 ipp_id = property->ipp_id; >> + >> + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, ipp_id); >> + >> + if (ipp_id) { >> + /* find ipp driver using idr */ >> + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, >> + ipp_id); >> + if (IS_ERR_OR_NULL(ippdrv)) { >> + DRM_ERROR("not found ipp%d driver.\n", ipp_id); >> + goto err_null; > > Just return ippdrv; - done. > >> + } >> + >> + /* >> + * WB, OUTPUT opertion not supported multi-operation. >> + * so, make dedicated state at set property ioctl. >> + * when ipp driver finished operations, clear dedicated > flags. >> + */ >> + if (ipp_check_dedicated(ippdrv, property->cmd)) { >> + DRM_ERROR("already used choose device.\n"); >> + goto err_null; > > Just return proper error; - done ERR_PTR(-EBUSY); > >> + } >> + >> + /* >> + * This is necessary to find correct device in ipp drivers. >> + * ipp drivers have different abilities, >> + * so need to check property. >> + */ >> + if (ippdrv->check_property && >> + ippdrv->check_property(ippdrv->dev, property)) { >> + DRM_ERROR("not support property.\n"); >> + goto err_null; > > Just return proper error; - done ERR_PTR(-EINVAL); > >> + } >> + >> + return ippdrv; >> + } else { >> + /* >> + * This case is search all ipp driver for finding. >> + * user application don't set ipp_id in this case, >> + * so ipp subsystem search correct driver in driver list. >> + */ >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, >> drv_list) { >> + if (ipp_check_dedicated(ippdrv, property->cmd)) { >> + DRM_DEBUG_KMS("%s:used device.\n", > __func__); >> + continue; >> + } >> + >> + if (ippdrv->check_property && >> + ippdrv->check_property(ippdrv->dev, property)) { >> + DRM_DEBUG_KMS("%s:not support property.\n", >> + __func__); >> + continue; >> + } >> + >> + return ippdrv; >> + } >> + >> + DRM_ERROR("not support ipp driver operations.\n"); >> + } >> + >> + return ERR_PTR(-ENODEV); >> + >> +err_null: >> + return NULL; > > And remove the above codes. - removed. > >> +} >> + >> +static struct exynos_drm_ippdrv *ipp_find_drv_by_handle(u32 prop_id) >> +{ >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + int count = 0; >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); >> + >> + if (list_empty(&exynos_drm_ippdrv_list)) { >> + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); >> + return ERR_PTR(-ENODEV); >> + } >> + >> + /* >> + * This case is search ipp driver by prop_id handle. >> + * sometimes, ipp subsystem find driver by prop_id. >> + * e.g PAUSE state, queue buf, command contro. >> + */ >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { >> + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", __func__, >> + count++, (int)ippdrv); >> + >> + if (!list_empty(&ippdrv->cmd_list)) { >> + list_for_each_entry(c_node, &ippdrv->cmd_list, list) >> + if (c_node->property.prop_id == prop_id) >> + return ippdrv; >> + } >> + } >> + >> + return ERR_PTR(-ENODEV); >> +} >> + >> +int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void *data, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; >> + struct device *dev = priv->dev; >> + struct ipp_context *ctx = get_ipp_context(dev); >> + struct drm_exynos_ipp_prop_list *prop_list = data; >> + struct exynos_drm_ippdrv *ippdrv; >> + int count = 0; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!ctx) { >> + DRM_ERROR("invalid context.\n"); >> + return -EINVAL; >> + } >> + >> + if (!prop_list) { >> + DRM_ERROR("invalid property parameter.\n"); >> + return -EINVAL; >> + } >> + >> + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, prop_list->ipp_id); >> + >> + if (!prop_list->ipp_id) { >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, >> drv_list) >> + count++; >> + /* >> + * Supports ippdrv list count for user application. >> + * First step user application getting ippdrv count. >> + * and second step getting ippdrv capability using ipp_id. >> + */ >> + prop_list->count = count; >> + } else { >> + /* >> + * Getting ippdrv capability by ipp_id. >> + * some deivce not supported wb, output interface. >> + * so, user application detect correct ipp driver >> + * using this ioctl. >> + */ >> + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, >> + prop_list->ipp_id); >> + if (!ippdrv) { >> + DRM_ERROR("not found ipp%d driver.\n", >> + prop_list->ipp_id); >> + return -EINVAL; >> + } >> + >> + prop_list = ippdrv->prop_list; >> + } >> + >> + return 0; >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ipp_get_property); > > Should this function really be exported? Which external module calls this > function? If not so, remove it. - removed all EXPORT_SYMBOL_GPL > >> + >> +static void ipp_print_property(struct drm_exynos_ipp_property *property, >> + int idx) >> +{ >> + struct drm_exynos_ipp_config *config = &property->config[idx]; >> + struct drm_exynos_pos *pos = &config->pos; >> + struct drm_exynos_sz *sz = &config->sz; >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]ops[%s]fmt[0x%x]\n", >> + __func__, property->prop_id, idx ? "dst" : "src", config- >>> fmt); >> + >> + DRM_DEBUG_KMS("%s:pos[%d %d %d %d]sz[%d %d]f[%d]r[%d]\n", >> + __func__, pos->x, pos->y, pos->w, pos->h, >> + sz->hsize, sz->vsize, config->flip, config->degree); >> +} >> + >> +static int ipp_find_and_set_property(struct drm_exynos_ipp_property >> *property) >> +{ >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + u32 prop_id = property->prop_id; >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); >> + >> + ippdrv = ipp_find_drv_by_handle(prop_id); >> + if (IS_ERR_OR_NULL(ippdrv)) { >> + DRM_ERROR("failed to get ipp driver.\n"); >> + return -EINVAL; >> + } >> + >> + /* >> + * Find command node using command list in ippdrv. >> + * when we find this command no using prop_id. >> + * return property information set in this command node. >> + */ >> + list_for_each_entry(c_node, &ippdrv->cmd_list, list) { >> + if ((c_node->property.prop_id == prop_id) && >> + (c_node->state == IPP_STATE_STOP)) { >> + DRM_DEBUG_KMS("%s:found cmd[%d]ippdrv[0x%x]\n", >> + __func__, property->cmd, (int)ippdrv); >> + >> + c_node->property = *property; >> + return 0; >> + } >> + } >> + >> + DRM_ERROR("failed to search property.\n"); >> + >> + return -EINVAL; >> +} >> + >> +static struct drm_exynos_ipp_cmd_work *ipp_create_cmd_work(void) >> +{ >> + struct drm_exynos_ipp_cmd_work *cmd_work; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + cmd_work = kzalloc(sizeof(*cmd_work), GFP_KERNEL); >> + if (!cmd_work) { >> + DRM_ERROR("failed to alloc cmd_work.\n"); >> + return ERR_PTR(-ENOMEM); >> + } >> + >> + INIT_WORK((struct work_struct *)cmd_work, ipp_sched_cmd); >> + >> + return cmd_work; >> +} >> + >> +static struct drm_exynos_ipp_event_work *ipp_create_event_work(void) >> +{ >> + struct drm_exynos_ipp_event_work *event_work; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + event_work = kzalloc(sizeof(*event_work), GFP_KERNEL); >> + if (!event_work) { >> + DRM_ERROR("failed to alloc event_work.\n"); >> + return ERR_PTR(-ENOMEM); >> + } >> + >> + INIT_WORK((struct work_struct *)event_work, ipp_sched_event); >> + >> + return event_work; >> +} >> + >> +int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void *data, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; >> + struct device *dev = priv->dev; >> + struct ipp_context *ctx = get_ipp_context(dev); >> + struct drm_exynos_ipp_property *property = data; >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + int ret, i; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!ctx) { >> + DRM_ERROR("invalid context.\n"); >> + return -EINVAL; >> + } >> + >> + if (!property) { >> + DRM_ERROR("invalid property parameter.\n"); >> + return -EINVAL; >> + } >> + >> + /* >> + * This is log print for user application property. >> + * user application set various property. >> + */ >> + for_each_ipp_ops(i) >> + ipp_print_property(property, i); >> + >> + /* >> + * set property ioctl generated new prop_id. >> + * but in this case already asigned prop_id using old set property. >> + * e.g PAUSE state. this case supports find current prop_id and use >> it >> + * instead of allocation. >> + */ >> + if (property->prop_id) { >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property- >>> prop_id); >> + return ipp_find_and_set_property(property); >> + } >> + >> + /* find ipp driver using ipp id */ >> + ippdrv = ipp_find_driver(ctx, property); >> + if (IS_ERR_OR_NULL(ippdrv)) { >> + DRM_ERROR("failed to get ipp driver.\n"); >> + return -EINVAL; >> + } >> + >> + /* allocate command node */ >> + c_node = kzalloc(sizeof(*c_node), GFP_KERNEL); >> + if (!c_node) { >> + DRM_ERROR("failed to allocate map node.\n"); >> + return -ENOMEM; >> + } >> + >> + /* create property id */ >> + ret = ipp_create_id(&ctx->prop_idr, &ctx->prop_lock, c_node, >> + &property->prop_id); >> + if (ret) { >> + DRM_ERROR("failed to create id.\n"); >> + goto err_clear; >> + } >> + >> + DRM_DEBUG_KMS("%s:created prop_id[%d]cmd[%d]ippdrv[0x%x]\n", >> + __func__, property->prop_id, property->cmd, (int)ippdrv); >> + >> + /* stored property information and ippdrv in private data */ >> + c_node->priv = priv; >> + c_node->property = *property; >> + c_node->state = IPP_STATE_IDLE; >> + >> + c_node->start_work = ipp_create_cmd_work(); >> + if (IS_ERR_OR_NULL(c_node->start_work)) { >> + DRM_ERROR("failed to create start work.\n"); >> + goto err_clear; >> + } >> + >> + c_node->stop_work = ipp_create_cmd_work(); >> + if (IS_ERR_OR_NULL(c_node->stop_work)) { >> + DRM_ERROR("failed to create stop work.\n"); >> + goto err_free_start; >> + } >> + >> + c_node->event_work = ipp_create_event_work(); >> + if (IS_ERR_OR_NULL(c_node->event_work)) { >> + DRM_ERROR("failed to create event work.\n"); >> + goto err_free_stop; >> + } >> + >> + mutex_init(&c_node->cmd_lock); >> + mutex_init(&c_node->mem_lock); >> + mutex_init(&c_node->event_lock); >> + >> + init_completion(&c_node->start_complete); >> + init_completion(&c_node->stop_complete); >> + >> + for_each_ipp_ops(i) >> + INIT_LIST_HEAD(&c_node->mem_list[i]); >> + >> + INIT_LIST_HEAD(&c_node->event_list); >> + list_splice_init(&priv->event_list, &c_node->event_list); >> + list_add_tail(&c_node->list, &ippdrv->cmd_list); >> + >> + /* make dedicated state without m2m */ >> + if (!ipp_is_m2m_cmd(property->cmd)) >> + ippdrv->dedicated = true; >> + >> + return 0; >> + >> +err_free_stop: >> + kfree(c_node->stop_work); >> +err_free_start: >> + kfree(c_node->start_work); >> +err_clear: >> + kfree(c_node); >> + return ret; >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ipp_set_property); >> + >> +static void ipp_clean_cmd_node(struct drm_exynos_ipp_cmd_node *c_node) >> +{ >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + /* delete list */ >> + list_del(&c_node->list); >> + >> + /* destroy mutex */ >> + mutex_destroy(&c_node->cmd_lock); >> + mutex_destroy(&c_node->mem_lock); >> + mutex_destroy(&c_node->event_lock); >> + >> + /* free command node */ >> + kfree(c_node->start_work); >> + kfree(c_node->stop_work); >> + kfree(c_node->event_work); >> + kfree(c_node); >> +} >> + >> +static int ipp_check_mem_list(struct drm_exynos_ipp_cmd_node *c_node) >> +{ >> + struct drm_exynos_ipp_property *property = &c_node->property; >> + struct drm_exynos_ipp_mem_node *m_node; >> + struct list_head *head; >> + int ret, i, count[EXYNOS_DRM_OPS_MAX] = { 0, }; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + mutex_lock(&c_node->mem_lock); >> + >> + for_each_ipp_ops(i) { >> + /* source/destination memory list */ >> + head = &c_node->mem_list[i]; >> + >> + if (list_empty(head)) { >> + DRM_DEBUG_KMS("%s:%s memory empty.\n", __func__, >> + i ? "dst" : "src"); >> + continue; >> + } >> + >> + /* find memory node entry */ >> + list_for_each_entry(m_node, head, list) { >> + DRM_DEBUG_KMS("%s:%s,count[%d]m_node[0x%x]\n", >> __func__, >> + i ? "dst" : "src", count[i], (int)m_node); >> + count[i]++; >> + } >> + } >> + >> + DRM_DEBUG_KMS("%s:min[%d]max[%d]\n", __func__, >> + min(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST]), >> + max(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST])); >> + >> + /* >> + * M2M operations should be need paired memory address. >> + * so, need to check minimum count about src, dst. >> + * other case not use paired memory, so use maximum count >> + */ >> + if (ipp_is_m2m_cmd(property->cmd)) >> + ret = min(count[EXYNOS_DRM_OPS_SRC], >> + count[EXYNOS_DRM_OPS_DST]); >> + else >> + ret = max(count[EXYNOS_DRM_OPS_SRC], >> + count[EXYNOS_DRM_OPS_DST]); >> + >> + mutex_unlock(&c_node->mem_lock); >> + >> + return ret; >> +} >> + >> +static struct drm_exynos_ipp_mem_node >> + *ipp_find_mem_node(struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct drm_exynos_ipp_mem_node *m_node; >> + struct list_head *head; >> + int count = 0; >> + >> + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, qbuf->buf_id); >> + >> + /* source/destination memory list */ >> + head = &c_node->mem_list[qbuf->ops_id]; >> + >> + /* find memory node from memory list */ >> + list_for_each_entry(m_node, head, list) { >> + DRM_DEBUG_KMS("%s:count[%d]m_node[0x%x]\n", >> + __func__, count++, (int)m_node); >> + >> + /* compare buffer id */ >> + if (m_node->buf_id == qbuf->buf_id) >> + return m_node; >> + } >> + >> + return NULL; >> +} >> + >> +static int ipp_set_mem_node(struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_mem_node *m_node) >> +{ >> + struct exynos_drm_ipp_ops *ops = NULL; >> + int ret = 0; >> + >> + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); >> + >> + if (!m_node) { >> + DRM_ERROR("invalid queue node.\n"); >> + return -EFAULT; >> + } >> + >> + mutex_lock(&c_node->mem_lock); >> + >> + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); >> + >> + /* get operations callback */ >> + ops = ippdrv->ops[m_node->ops_id]; >> + if (!ops) { >> + DRM_ERROR("not support ops.\n"); >> + ret = -EFAULT; >> + goto err_unlock; >> + } >> + >> + /* set address and enable irq */ >> + if (ops->set_addr) { >> + ret = ops->set_addr(ippdrv->dev, &m_node->buf_info, >> + m_node->buf_id, IPP_BUF_ENQUEUE); >> + if (ret) { >> + DRM_ERROR("failed to set addr.\n"); >> + goto err_unlock; >> + } >> + } >> + >> +err_unlock: >> + mutex_unlock(&c_node->mem_lock); >> + return ret; >> +} >> + >> +static struct drm_exynos_ipp_mem_node >> + *ipp_get_mem_node(struct drm_device *drm_dev, >> + struct drm_file *file, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct drm_exynos_ipp_mem_node *m_node; >> + struct drm_exynos_ipp_buf_info buf_info; >> + void *addr; >> + int i; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + mutex_lock(&c_node->mem_lock); >> + >> + m_node = kzalloc(sizeof(*m_node), GFP_KERNEL); >> + if (!m_node) { >> + DRM_ERROR("failed to allocate queue node.\n"); >> + goto err_unlock; >> + } >> + >> + /* clear base address for error handling */ >> + memset(&buf_info, 0x0, sizeof(buf_info)); >> + >> + /* operations, buffer id */ >> + m_node->ops_id = qbuf->ops_id; >> + m_node->prop_id = qbuf->prop_id; >> + m_node->buf_id = qbuf->buf_id; >> + >> + DRM_DEBUG_KMS("%s:m_node[0x%x]ops_id[%d]\n", __func__, >> + (int)m_node, qbuf->ops_id); >> + DRM_DEBUG_KMS("%s:prop_id[%d]buf_id[%d]\n", __func__, >> + qbuf->prop_id, m_node->buf_id); >> + >> + for_each_ipp_planar(i) { >> + DRM_DEBUG_KMS("%s:i[%d]handle[0x%x]\n", __func__, >> + i, qbuf->handle[i]); >> + >> + /* get dma address by handle */ >> + if (qbuf->handle[i] != 0) { > > if (qbuf->handle[i]) { - done. > >> + addr = exynos_drm_gem_get_dma_addr(drm_dev, >> + qbuf->handle[i], file); >> + if (!addr) { > > if (IS_ERR(addr)) - done. > >> + DRM_ERROR("failed to get addr.\n"); >> + goto err_clear; >> + } >> + >> + buf_info.handles[i] = qbuf->handle[i]; >> + buf_info.base[i] = *(dma_addr_t *) addr; >> + DRM_DEBUG_KMS("%s:i[%d]base[0x%x]hd[0x%x]\n", >> + __func__, i, buf_info.base[i], >> + (int)buf_info.handles[i]); >> + } >> + } >> + >> + m_node->filp = file; >> + m_node->buf_info = buf_info; >> + list_add_tail(&m_node->list, &c_node->mem_list[qbuf->ops_id]); >> + >> + mutex_unlock(&c_node->mem_lock); >> + return m_node; >> + >> +err_clear: >> + kfree(m_node); >> +err_unlock: >> + mutex_unlock(&c_node->mem_lock); >> + return ERR_PTR(-EFAULT); >> +} >> + >> +static int ipp_put_mem_node(struct drm_device *drm_dev, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_mem_node *m_node) >> +{ >> + int i; >> + >> + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); >> + >> + if (!m_node) { >> + DRM_ERROR("invalid dequeue node.\n"); >> + return -EFAULT; >> + } >> + >> + if (list_empty(&m_node->list)) { >> + DRM_ERROR("empty memory node.\n"); >> + return -ENOMEM; >> + } >> + >> + mutex_lock(&c_node->mem_lock); >> + >> + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); >> + >> + /* put gem buffer */ >> + for_each_ipp_planar(i) { >> + unsigned long handle = m_node->buf_info.handles[i]; >> + if (handle) >> + exynos_drm_gem_put_dma_addr(drm_dev, handle, >> + m_node->filp); >> + } >> + >> + /* delete list in queue */ >> + list_del(&m_node->list); >> + kfree(m_node); >> + >> + mutex_unlock(&c_node->mem_lock); >> + >> + return 0; >> +} >> + >> +static void ipp_free_event(struct drm_pending_event *event) > > Does this function need? This function just calls kfree. Just call kfree > instead of wrapper. - It is Kyungmin.Park review. > >> +{ >> + kfree(event); >> +} >> + >> +static int ipp_get_event(struct drm_device *drm_dev, >> + struct drm_file *file, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct drm_exynos_ipp_send_event *e; >> + unsigned long flags; >> + >> + DRM_DEBUG_KMS("%s:ops_id[%d]buf_id[%d]\n", __func__, >> + qbuf->ops_id, qbuf->buf_id); >> + >> + e = kzalloc(sizeof(*e), GFP_KERNEL); >> + >> + if (!e) { >> + DRM_ERROR("failed to allocate event.\n"); >> + spin_lock_irqsave(&drm_dev->event_lock, flags); >> + file->event_space += sizeof(e->event); >> + spin_unlock_irqrestore(&drm_dev->event_lock, flags); >> + return -ENOMEM; >> + } >> + >> + /* make event */ >> + e->event.base.type = DRM_EXYNOS_IPP_EVENT; >> + e->event.base.length = sizeof(e->event); >> + e->event.user_data = qbuf->user_data; >> + e->event.prop_id = qbuf->prop_id; >> + e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; >> + e->base.event = &e->event.base; >> + e->base.file_priv = file; >> + e->base.destroy = ipp_free_event; >> + list_add_tail(&e->base.link, &c_node->event_list); >> + >> + return 0; >> +} >> + >> +static void ipp_put_event(struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct drm_exynos_ipp_send_event *e, *te; >> + int count = 0; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (list_empty(&c_node->event_list)) { >> + DRM_DEBUG_KMS("%s:event_list is empty.\n", __func__); >> + return; >> + } >> + >> + list_for_each_entry_safe(e, te, &c_node->event_list, base.link) { >> + DRM_DEBUG_KMS("%s:count[%d]e[0x%x]\n", >> + __func__, count++, (int)e); >> + >> + /* >> + * quf == NULL condition means all event deletion. >> + * stop operations want to delete all event list. >> + * another case delete only same buf id. >> + */ >> + if (!qbuf) { >> + /* delete list */ >> + list_del(&e->base.link); >> + kfree(e); >> + } >> + >> + /* compare buffer id */ >> + if (qbuf && (qbuf->buf_id == >> + e->event.buf_id[EXYNOS_DRM_OPS_DST])) { >> + /* delete list */ >> + list_del(&e->base.link); >> + kfree(e); >> + return; >> + } >> + } >> +} >> + >> +void ipp_handle_cmd_work(struct device *dev, >> + struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_cmd_work *cmd_work, >> + struct drm_exynos_ipp_cmd_node *c_node) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + >> + cmd_work->ippdrv = ippdrv; >> + cmd_work->c_node = c_node; >> + queue_work(ctx->cmd_workq, (struct work_struct *)cmd_work); >> +} >> + >> +static int ipp_queue_buf_with_run(struct device *dev, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_mem_node *m_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_property *property; >> + struct exynos_drm_ipp_ops *ops; >> + int ret; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + ippdrv = ipp_find_drv_by_handle(qbuf->prop_id); >> + if (IS_ERR_OR_NULL(ippdrv)) { >> + DRM_ERROR("failed to get ipp driver.\n"); >> + return -EFAULT; >> + } >> + >> + ops = ippdrv->ops[qbuf->ops_id]; >> + if (!ops) { >> + DRM_ERROR("failed to get ops.\n"); >> + return -EFAULT; >> + } >> + >> + property = &c_node->property; >> + if (!property) { > > Property couldn't be NULL. remove it. - done. > >> + DRM_ERROR("invalid property parameter.\n"); >> + return -EINVAL; >> + } >> + >> + if (c_node->state != IPP_STATE_START) { >> + DRM_DEBUG_KMS("%s:bypass for invalid state.\n" , __func__); >> + return 0; > > No error? - This is no error. command started yet. user application make play command control. > >> + } >> + >> + if (!ipp_check_mem_list(c_node)) { >> + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); >> + return 0; > > No error? - also, this case no error. not finished memory list. > >> + } >> + >> + /* >> + * If set destination buffer and enabled clock, >> + * then m2m operations need start operations at queue_buf >> + */ >> + if (ipp_is_m2m_cmd(property->cmd)) { >> + struct drm_exynos_ipp_cmd_work *cmd_work = c_node- >>> start_work; >> + >> + cmd_work->ctrl = IPP_CTRL_PLAY; >> + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); >> + } else { >> + ret = ipp_set_mem_node(ippdrv, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to set m node.\n"); >> + return ret; >> + } >> + } >> + >> + return 0; >> +} >> + >> +static void ipp_clean_queue_buf(struct drm_device *drm_dev, >> + struct drm_exynos_ipp_cmd_node *c_node, >> + struct drm_exynos_ipp_queue_buf *qbuf) >> +{ >> + struct drm_exynos_ipp_mem_node *m_node, *tm_node; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!list_empty(&c_node->mem_list[qbuf->ops_id])) { >> + /* delete list */ >> + list_for_each_entry_safe(m_node, tm_node, >> + &c_node->mem_list[qbuf->ops_id], list) { >> + if (m_node->buf_id == qbuf->buf_id && >> + m_node->ops_id == qbuf->ops_id) >> + ipp_put_mem_node(drm_dev, c_node, m_node); >> + } >> + } >> +} >> + >> +int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void *data, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; >> + struct device *dev = priv->dev; >> + struct ipp_context *ctx = get_ipp_context(dev); >> + struct drm_exynos_ipp_queue_buf *qbuf = data; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + struct drm_exynos_ipp_mem_node *m_node; >> + int ret; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!qbuf) { >> + DRM_ERROR("invalid buf parameter.\n"); >> + return -EINVAL; >> + } >> + >> + if (qbuf->ops_id >= EXYNOS_DRM_OPS_MAX) { >> + DRM_ERROR("invalid ops parameter.\n"); >> + return -EINVAL; >> + } >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", >> + __func__, qbuf->prop_id, qbuf->ops_id ? "dst" : "src", >> + qbuf->buf_id, qbuf->buf_type); >> + >> + /* find command node */ >> + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, >> + qbuf->prop_id); >> + if (!c_node) { >> + DRM_ERROR("failed to get command node.\n"); >> + return -EFAULT; >> + } >> + >> + /* buffer control */ >> + switch (qbuf->buf_type) { >> + case IPP_BUF_ENQUEUE: >> + /* get memory node */ >> + m_node = ipp_get_mem_node(drm_dev, file, c_node, qbuf); >> + if (IS_ERR_OR_NULL(m_node)) { > > This function never returns error. Use IS_ERR(m_node) instead. - done. > >> + DRM_ERROR("failed to get m_node.\n"); >> + return -EINVAL; > > And return PTR_ERR(m_node); - done. > >> + } >> + >> + /* >> + * first step get event for destination buffer. >> + * and second step when M2M case run with destination buffer >> + * if needed. >> + */ >> + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) { >> + /* get event for destination buffer */ >> + ret = ipp_get_event(drm_dev, file, c_node, qbuf); >> + if (ret) { >> + DRM_ERROR("failed to get event.\n"); >> + goto err_clean_node; >> + } >> + >> + /* >> + * M2M case run play control for streaming feature. >> + * other case set address and waiting. >> + */ >> + ret = ipp_queue_buf_with_run(dev, c_node, m_node, >> qbuf); >> + if (ret) { >> + DRM_ERROR("failed to run command.\n"); >> + goto err_clean_node; >> + } >> + } >> + break; >> + case IPP_BUF_DEQUEUE: >> + mutex_lock(&c_node->cmd_lock); >> + >> + /* put event for destination buffer */ >> + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) >> + ipp_put_event(c_node, qbuf); >> + >> + ipp_clean_queue_buf(drm_dev, c_node, qbuf); >> + >> + mutex_unlock(&c_node->cmd_lock); >> + break; >> + default: >> + DRM_ERROR("invalid buffer control.\n"); >> + return -EINVAL; >> + } >> + >> + return 0; >> + >> +err_clean_node: >> + DRM_ERROR("clean memory nodes.\n"); >> + >> + ipp_clean_queue_buf(drm_dev, c_node, qbuf); >> + return ret; >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ipp_queue_buf); > > Remove EXPORT_SYMBOL_GPL. - done. > >> + >> +static bool exynos_drm_ipp_check_valid(struct device *dev, >> + enum drm_exynos_ipp_ctrl ctrl, enum drm_exynos_ipp_state >> state) >> +{ >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (ctrl != IPP_CTRL_PLAY) { >> + if (pm_runtime_suspended(dev)) { >> + DRM_ERROR("pm:runtime_suspended.\n"); >> + goto err_status; >> + } >> + } >> + >> + switch (ctrl) { >> + case IPP_CTRL_PLAY: >> + if (state != IPP_STATE_IDLE) >> + goto err_status; >> + break; >> + case IPP_CTRL_STOP: >> + if (state == IPP_STATE_STOP) >> + goto err_status; >> + break; >> + case IPP_CTRL_PAUSE: >> + if (state != IPP_STATE_START) >> + goto err_status; >> + break; >> + case IPP_CTRL_RESUME: >> + if (state != IPP_STATE_STOP) >> + goto err_status; >> + break; >> + default: >> + DRM_ERROR("invalid state.\n"); >> + goto err_status; >> + break; >> + } >> + >> + return true; >> + >> +err_status: >> + DRM_ERROR("invalid status:ctrl[%d]state[%d]\n", ctrl, state); >> + return false; >> +} >> + >> +int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; >> + struct exynos_drm_ippdrv *ippdrv = NULL; >> + struct device *dev = priv->dev; >> + struct ipp_context *ctx = get_ipp_context(dev); >> + struct drm_exynos_ipp_cmd_ctrl *cmd_ctrl = data; >> + struct drm_exynos_ipp_cmd_work *cmd_work; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + int ret = 0; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!ctx) { >> + DRM_ERROR("invalid context.\n"); >> + return -EINVAL; >> + } >> + >> + if (!cmd_ctrl) { >> + DRM_ERROR("invalid control parameter.\n"); >> + return -EINVAL; >> + } >> + >> + DRM_DEBUG_KMS("%s:ctrl[%d]prop_id[%d]\n", __func__, >> + cmd_ctrl->ctrl, cmd_ctrl->prop_id); >> + >> + ippdrv = ipp_find_drv_by_handle(cmd_ctrl->prop_id); >> + if (IS_ERR_OR_NULL(ippdrv)) { > > Never return NULL. just use IS_ERR(ippdrv) - done. > >> + DRM_ERROR("failed to get ipp driver.\n"); >> + return -EINVAL; > > And return PTR_ERR(ippdrv); - done. > >> + } >> + >> + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, >> + cmd_ctrl->prop_id); >> + if (!c_node) { >> + DRM_ERROR("invalid command node list.\n"); >> + return -EINVAL; >> + } >> + >> + if (!exynos_drm_ipp_check_valid(ippdrv->dev, cmd_ctrl->ctrl, >> + c_node->state)) { >> + DRM_ERROR("invalid state.\n"); >> + return -EINVAL; >> + } >> + >> + switch (cmd_ctrl->ctrl) { >> + case IPP_CTRL_PLAY: >> + if (pm_runtime_suspended(ippdrv->dev)) >> + pm_runtime_get_sync(ippdrv->dev); >> + c_node->state = IPP_STATE_START; >> + >> + cmd_work = c_node->start_work; >> + cmd_work->ctrl = cmd_ctrl->ctrl; >> + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); >> + c_node->state = IPP_STATE_START; >> + break; >> + case IPP_CTRL_STOP: >> + cmd_work = c_node->stop_work; >> + cmd_work->ctrl = cmd_ctrl->ctrl; >> + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); >> + >> + if (!wait_for_completion_timeout(&c_node->stop_complete, >> + msecs_to_jiffies(300))) { >> + DRM_ERROR("timeout stop:prop_id[%d]\n", >> + c_node->property.prop_id); >> + } >> + >> + c_node->state = IPP_STATE_STOP; >> + ippdrv->dedicated = false; >> + ipp_clean_cmd_node(c_node); >> + >> + if (list_empty(&ippdrv->cmd_list)) >> + pm_runtime_put_sync(ippdrv->dev); >> + break; >> + case IPP_CTRL_PAUSE: >> + cmd_work = c_node->stop_work; >> + cmd_work->ctrl = cmd_ctrl->ctrl; >> + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); >> + >> + if (!wait_for_completion_timeout(&c_node->stop_complete, >> + msecs_to_jiffies(200))) { >> + DRM_ERROR("timeout stop:prop_id[%d]\n", >> + c_node->property.prop_id); >> + } >> + >> + c_node->state = IPP_STATE_STOP; >> + break; >> + case IPP_CTRL_RESUME: >> + c_node->state = IPP_STATE_START; >> + cmd_work = c_node->start_work; >> + cmd_work->ctrl = cmd_ctrl->ctrl; >> + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); >> + break; >> + default: >> + /* ToDo: expand ctrl operation */ >> + DRM_ERROR("could not support this state currently.\n"); >> + goto err_clear; > > Just return error; - added return -EINVAL; > >> + } >> + >> + DRM_DEBUG_KMS("%s:done ctrl[%d]prop_id[%d]\n", __func__, >> + cmd_ctrl->ctrl, cmd_ctrl->prop_id); >> + >> + return 0; >> + >> +err_clear: >> + return ret; > > And remove the above label. - done. > >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ipp_cmd_ctrl); > > Remove it. - done. > >> + >> +int exynos_drm_ippnb_register(struct notifier_block *nb) >> +{ >> + return blocking_notifier_chain_register( >> + &exynos_drm_ippnb_list, nb); >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_register); > > Remove it. - done. > >> + >> +int exynos_drm_ippnb_unregister(struct notifier_block *nb) >> +{ >> + return blocking_notifier_chain_unregister( >> + &exynos_drm_ippnb_list, nb); >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_unregister); > > Remove it. - done. > >> + >> +int exynos_drm_ippnb_send_event(unsigned long val, void *v) >> +{ >> + return blocking_notifier_call_chain( >> + &exynos_drm_ippnb_list, val, v); >> +} >> +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_send_event); > > Remove it. - done. > >> + >> +static int ipp_set_property(struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_property *property) >> +{ >> + struct exynos_drm_ipp_ops *ops = NULL; >> + int ret, i, swap = 0; >> + >> + if (!property) { >> + DRM_ERROR("invalid property parameter.\n"); >> + return -EINVAL; >> + } >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); >> + >> + /* reset h/w block */ >> + if (ippdrv->reset && >> + ippdrv->reset(ippdrv->dev)) { >> + DRM_ERROR("failed to reset.\n"); >> + return -EINVAL; >> + } >> + >> + /* set source,destination operations */ >> + for_each_ipp_ops(i) { >> + /* ToDo: integrate property and config */ >> + struct drm_exynos_ipp_config *config = >> + &property->config[i]; >> + >> + ops = ippdrv->ops[i]; >> + if (!ops || !config) { >> + DRM_ERROR("not support ops and config.\n"); >> + return -EINVAL; >> + } >> + >> + /* set format */ >> + if (ops->set_fmt) { >> + ret = ops->set_fmt(ippdrv->dev, config->fmt); >> + if (ret) { >> + DRM_ERROR("not support format.\n"); >> + return ret; >> + } >> + } >> + >> + /* set transform for rotation, flip */ >> + if (ops->set_transf) { >> + swap = ops->set_transf(ippdrv->dev, config->degree, >> + config->flip); >> + if (swap < 0) { >> + DRM_ERROR("not support tranf.\n"); >> + return -EINVAL; >> + } >> + } >> + >> + /* set size */ >> + if (ops->set_size) { >> + ret = ops->set_size(ippdrv->dev, swap, &config->pos, >> + &config->sz); >> + if (ret) { >> + DRM_ERROR("not support size.\n"); >> + return ret; >> + } >> + } >> + } >> + >> + return 0; >> +} >> + >> +static int ipp_start_property(struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_cmd_node *c_node) >> +{ >> + struct drm_exynos_ipp_mem_node *m_node; >> + struct drm_exynos_ipp_property *property = &c_node->property; >> + struct list_head *head; >> + int ret, i; >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); >> + >> + /* store command info in ippdrv */ >> + ippdrv->cmd = c_node; >> + >> + if (!ipp_check_mem_list(c_node)) { >> + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); >> + return -ENOMEM; >> + } >> + >> + /* set current property in ippdrv */ >> + ret = ipp_set_property(ippdrv, property); >> + if (ret) { >> + DRM_ERROR("failed to set property.\n"); >> + ippdrv->cmd = NULL; >> + return ret; >> + } >> + >> + /* check command */ >> + switch (property->cmd) { >> + case IPP_CMD_M2M: >> + for_each_ipp_ops(i) { >> + /* source/destination memory list */ >> + head = &c_node->mem_list[i]; >> + >> + m_node = list_first_entry(head, >> + struct drm_exynos_ipp_mem_node, list); >> + if (!m_node) { >> + DRM_ERROR("failed to get node.\n"); >> + ret = -EFAULT; >> + return ret; >> + } >> + >> + DRM_DEBUG_KMS("%s:m_node[0x%x]\n", >> + __func__, (int)m_node); >> + >> + ret = ipp_set_mem_node(ippdrv, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to set m node.\n"); >> + return ret; >> + } >> + } >> + break; >> + case IPP_CMD_WB: >> + /* destination memory list */ >> + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; >> + >> + list_for_each_entry(m_node, head, list) { >> + ret = ipp_set_mem_node(ippdrv, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to set m node.\n"); >> + return ret; >> + } >> + } >> + break; >> + case IPP_CMD_OUTPUT: >> + /* source memory list */ >> + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; >> + >> + list_for_each_entry(m_node, head, list) { >> + ret = ipp_set_mem_node(ippdrv, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to set m node.\n"); >> + return ret; >> + } >> + } >> + break; >> + default: >> + DRM_ERROR("invalid operations.\n"); >> + return -EINVAL; >> + } >> + >> + DRM_DEBUG_KMS("%s:cmd[%d]\n", __func__, property->cmd); >> + >> + /* start operations */ >> + if (ippdrv->start) { >> + ret = ippdrv->start(ippdrv->dev, property->cmd); >> + if (ret) { >> + DRM_ERROR("failed to start ops.\n"); >> + return ret; >> + } >> + } >> + >> + return 0; >> +} >> + >> +static int ipp_stop_property(struct drm_device *drm_dev, >> + struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_cmd_node *c_node) >> +{ >> + struct drm_exynos_ipp_mem_node *m_node, *tm_node; >> + struct drm_exynos_ipp_property *property = &c_node->property; >> + struct list_head *head; >> + int ret = 0, i; >> + >> + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); >> + >> + /* put event */ >> + ipp_put_event(c_node, NULL); >> + >> + /* check command */ >> + switch (property->cmd) { >> + case IPP_CMD_M2M: >> + for_each_ipp_ops(i) { >> + /* source/destination memory list */ >> + head = &c_node->mem_list[i]; >> + >> + if (list_empty(head)) { >> + DRM_DEBUG_KMS("%s:mem_list is empty.\n", >> + __func__); >> + break; >> + } >> + >> + list_for_each_entry_safe(m_node, tm_node, >> + head, list) { >> + ret = ipp_put_mem_node(drm_dev, c_node, >> + m_node); >> + if (ret) { >> + DRM_ERROR("failed to put > m_node.\n"); >> + goto err_clear; >> + } >> + } >> + } >> + break; >> + case IPP_CMD_WB: >> + /* destination memory list */ >> + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; >> + >> + if (list_empty(head)) { >> + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); >> + break; >> + } >> + >> + list_for_each_entry_safe(m_node, tm_node, head, list) { >> + ret = ipp_put_mem_node(drm_dev, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to put m_node.\n"); >> + goto err_clear; >> + } >> + } >> + break; >> + case IPP_CMD_OUTPUT: >> + /* source memory list */ >> + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; >> + >> + if (list_empty(head)) { >> + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); >> + break; >> + } >> + >> + list_for_each_entry_safe(m_node, tm_node, head, list) { >> + ret = ipp_put_mem_node(drm_dev, c_node, m_node); >> + if (ret) { >> + DRM_ERROR("failed to put m_node.\n"); >> + goto err_clear; >> + } >> + } >> + break; >> + default: >> + DRM_ERROR("invalid operations.\n"); >> + ret = -EINVAL; >> + goto err_clear; >> + } >> + >> +err_clear: >> + /* stop operations */ >> + if (ippdrv->stop) >> + ippdrv->stop(ippdrv->dev, property->cmd); >> + >> + return ret; >> +} >> + >> +void ipp_sched_cmd(struct work_struct *work) >> +{ >> + struct drm_exynos_ipp_cmd_work *cmd_work = >> + (struct drm_exynos_ipp_cmd_work *)work; >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + struct drm_exynos_ipp_property *property; >> + int ret; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + ippdrv = cmd_work->ippdrv; >> + if (!ippdrv) { >> + DRM_ERROR("invalid ippdrv list.\n"); >> + return; >> + } >> + >> + c_node = cmd_work->c_node; >> + if (!c_node) { >> + DRM_ERROR("invalid command node list.\n"); >> + return; >> + } >> + >> + mutex_lock(&c_node->cmd_lock); >> + >> + property = &c_node->property; >> + if (!property) { >> + DRM_ERROR("failed to get property:prop_id[%d]\n", >> + c_node->property.prop_id); >> + goto err_unlock; >> + } >> + >> + switch (cmd_work->ctrl) { >> + case IPP_CTRL_PLAY: >> + case IPP_CTRL_RESUME: >> + ret = ipp_start_property(ippdrv, c_node); >> + if (ret) { >> + DRM_ERROR("failed to start property:prop_id[%d]\n", >> + c_node->property.prop_id); >> + goto err_unlock; >> + } >> + >> + /* >> + * M2M case supports wait_completion of transfer. >> + * because M2M case supports single unit operation >> + * with multiple queue. >> + * M2M need to wait completion of data transfer. >> + */ >> + if (ipp_is_m2m_cmd(property->cmd)) { >> + if (!wait_for_completion_timeout >> + (&c_node->start_complete, > msecs_to_jiffies(200))) { >> + DRM_ERROR("timeout event:prop_id[%d]\n", >> + c_node->property.prop_id); >> + goto err_unlock; >> + } >> + } >> + break; >> + case IPP_CTRL_STOP: >> + case IPP_CTRL_PAUSE: >> + ret = ipp_stop_property(ippdrv->drm_dev, ippdrv, >> + c_node); >> + if (ret) { >> + DRM_ERROR("failed to stop property.\n"); >> + goto err_unlock; >> + } >> + >> + complete(&c_node->stop_complete); >> + break; >> + default: >> + DRM_ERROR("unknown control type\n"); >> + break; >> + } >> + >> + DRM_DEBUG_KMS("%s:ctrl[%d] done.\n", __func__, cmd_work->ctrl); >> + >> +err_unlock: >> + mutex_unlock(&c_node->cmd_lock); >> +} >> + >> +static int ipp_send_event(struct exynos_drm_ippdrv *ippdrv, >> + struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) >> +{ >> + struct drm_device *drm_dev = ippdrv->drm_dev; >> + struct drm_exynos_ipp_property *property = &c_node->property; >> + struct drm_exynos_ipp_mem_node *m_node; >> + struct drm_exynos_ipp_queue_buf qbuf; >> + struct drm_exynos_ipp_send_event *e; >> + struct list_head *head; >> + struct timeval now; >> + unsigned long flags; >> + u32 tbuf_id[EXYNOS_DRM_OPS_MAX] = {0, }; >> + int ret, i; >> + >> + for_each_ipp_ops(i) >> + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, >> + i ? "dst" : "src", buf_id[i]); >> + >> + if (!drm_dev) { >> + DRM_ERROR("failed to get drm_dev.\n"); >> + return -EINVAL; >> + } >> + >> + if (!property) { >> + DRM_ERROR("failed to get property.\n"); >> + return -EINVAL; >> + } >> + >> + if (list_empty(&c_node->event_list)) { >> + DRM_DEBUG_KMS("%s:event list is empty.\n", __func__); >> + return 0; >> + } >> + >> + if (!ipp_check_mem_list(c_node)) { >> + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); >> + return 0; >> + } >> + >> + /* check command */ >> + switch (property->cmd) { >> + case IPP_CMD_M2M: >> + for_each_ipp_ops(i) { >> + /* source/destination memory list */ >> + head = &c_node->mem_list[i]; >> + >> + m_node = list_first_entry(head, >> + struct drm_exynos_ipp_mem_node, list); >> + if (!m_node) { >> + DRM_ERROR("empty memory node.\n"); >> + return -ENOMEM; >> + } >> + >> + tbuf_id[i] = m_node->buf_id; >> + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, >> + i ? "dst" : "src", tbuf_id[i]); >> + >> + ret = ipp_put_mem_node(drm_dev, c_node, m_node); >> + if (ret) >> + DRM_ERROR("failed to put m_node.\n"); >> + } >> + break; >> + case IPP_CMD_WB: >> + /* clear buf for finding */ >> + memset(&qbuf, 0x0, sizeof(qbuf)); >> + qbuf.ops_id = EXYNOS_DRM_OPS_DST; >> + qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST]; >> + >> + /* get memory node entry */ >> + m_node = ipp_find_mem_node(c_node, &qbuf); >> + if (!m_node) { >> + DRM_ERROR("empty memory node.\n"); >> + return -ENOMEM; >> + } >> + >> + tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id; >> + >> + ret = ipp_put_mem_node(drm_dev, c_node, m_node); >> + if (ret) >> + DRM_ERROR("failed to put m_node.\n"); >> + break; >> + case IPP_CMD_OUTPUT: >> + /* source memory list */ >> + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; >> + >> + m_node = list_first_entry(head, >> + struct drm_exynos_ipp_mem_node, list); >> + if (!m_node) { >> + DRM_ERROR("empty memory node.\n"); >> + return -ENOMEM; >> + } >> + >> + tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id; >> + >> + ret = ipp_put_mem_node(drm_dev, c_node, m_node); >> + if (ret) >> + DRM_ERROR("failed to put m_node.\n"); >> + break; >> + default: >> + DRM_ERROR("invalid operations.\n"); >> + return -EINVAL; >> + } >> + >> + /* ToDo: Fix buffer id */ >> + if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST]) >> + DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n", >> + tbuf_id[1], buf_id[1], property->prop_id); >> + >> + /* >> + * command node have event list of destination buffer >> + * If destination buffer enqueue to mem list, >> + * then we make event and link to event list tail. >> + * so, we get first event for first enqueued buffer. >> + */ >> + e = list_first_entry(&c_node->event_list, >> + struct drm_exynos_ipp_send_event, base.link); >> + >> + if (!e) { >> + DRM_ERROR("empty event.\n"); >> + return -EINVAL; >> + } >> + >> + do_gettimeofday(&now); >> + DRM_DEBUG_KMS("%s:tv_sec[%ld]tv_usec[%ld]\n" >> + , __func__, now.tv_sec, now.tv_usec); >> + e->event.tv_sec = now.tv_sec; >> + e->event.tv_usec = now.tv_usec; >> + e->event.prop_id = property->prop_id; >> + >> + /* set buffer id about source destination */ >> + for_each_ipp_ops(i) >> + e->event.buf_id[i] = tbuf_id[i]; >> + >> + /* ToDo: compare index. If needed */ >> + >> + spin_lock_irqsave(&drm_dev->event_lock, flags); >> + list_move_tail(&e->base.link, &e->base.file_priv->event_list); >> + wake_up_interruptible(&e->base.file_priv->event_wait); >> + spin_unlock_irqrestore(&drm_dev->event_lock, flags); >> + >> + DRM_DEBUG_KMS("%s:done cmd[%d]prop_id[%d]buf_id[%d]\n", __func__, >> + property->cmd, property->prop_id, >> tbuf_id[EXYNOS_DRM_OPS_DST]); >> + >> + return 0; >> +} >> + >> +void ipp_sched_event(struct work_struct *work) >> +{ >> + struct drm_exynos_ipp_event_work *event_work = >> + (struct drm_exynos_ipp_event_work *)work; >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + int ret; >> + >> + if (!event_work) { >> + DRM_ERROR("failed to get event_work.\n"); >> + return; >> + } >> + >> + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, >> + event_work->buf_id[EXYNOS_DRM_OPS_DST]); >> + >> + ippdrv = event_work->ippdrv; >> + if (!ippdrv) { >> + DRM_ERROR("failed to get ipp driver.\n"); >> + return; >> + } >> + >> + c_node = ippdrv->cmd; >> + if (!c_node) { >> + DRM_ERROR("failed to get command node.\n"); >> + return; >> + } >> + >> + /* >> + * IPP supports command thread, event thread synchronization. >> + * If IPP close immediately from user land, then IPP make >> + * synchronization with command thread, so make complete event. >> + * or going out operations. >> + */ >> + if (c_node->state != IPP_STATE_START) { >> + DRM_DEBUG_KMS("%s:bypass state[%d]prop_id[%d]\n", >> + __func__, c_node->state, c_node->property.prop_id); >> + goto err_completion; >> + } >> + >> + mutex_lock(&c_node->event_lock); >> + >> + ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); >> + if (ret) { >> + DRM_ERROR("failed to send event.\n"); >> + goto err_completion; >> + } >> + >> +err_completion: >> + if (ipp_is_m2m_cmd(c_node->property.cmd)) >> + complete(&c_node->start_complete); >> + >> + mutex_unlock(&c_node->event_lock); >> +} >> + >> +static int ipp_subdrv_probe(struct drm_device *drm_dev, struct device >> *dev) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + struct exynos_drm_ippdrv *ippdrv; >> + int ret, count = 0; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + /* get ipp driver entry */ >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { >> + ippdrv->drm_dev = drm_dev; >> + >> + ret = ipp_create_id(&ctx->ipp_idr, &ctx->ipp_lock, ippdrv, >> + &ippdrv->ipp_id); >> + if (ret) { >> + DRM_ERROR("failed to create id.\n"); >> + goto err_idr; >> + } >> + >> + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]ipp_id[%d]\n", >> __func__, >> + count++, (int)ippdrv, ippdrv->ipp_id); >> + >> + if (ippdrv->ipp_id == 0) { >> + DRM_ERROR("failed to get ipp_id[%d]\n", >> + ippdrv->ipp_id); >> + goto err_idr; >> + } >> + >> + /* store parent device for node */ >> + ippdrv->parent_dev = dev; >> + >> + /* store event work queue and handler */ >> + ippdrv->event_workq = ctx->event_workq; >> + ippdrv->sched_event = ipp_sched_event; >> + INIT_LIST_HEAD(&ippdrv->cmd_list); >> + } >> + >> + return 0; >> + >> +err_idr: >> + idr_remove_all(&ctx->ipp_idr); >> + idr_remove_all(&ctx->prop_idr); >> + idr_destroy(&ctx->ipp_idr); >> + idr_destroy(&ctx->prop_idr); >> + return ret; >> +} >> + >> +static void ipp_subdrv_remove(struct drm_device *drm_dev, struct device >> *dev) >> +{ >> + struct exynos_drm_ippdrv *ippdrv; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + /* get ipp driver entry */ >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { >> + ippdrv->drm_dev = NULL; >> + exynos_drm_ippdrv_unregister(ippdrv); >> + } >> + >> + /* ToDo: free notifier callback list if needed */ >> +} >> + >> +static int ipp_subdrv_open(struct drm_device *drm_dev, struct device > *dev, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + /* ToDo: multi device open */ >> + >> + priv = kzalloc(sizeof(*priv), GFP_KERNEL); >> + if (!priv) { >> + DRM_ERROR("failed to allocate priv.\n"); >> + return -ENOMEM; >> + } >> + priv->dev = dev; >> + file_priv->ipp_priv = priv; >> + >> + INIT_LIST_HEAD(&priv->event_list); >> + >> + DRM_DEBUG_KMS("%s:done priv[0x%x]\n", __func__, (int)priv); >> + >> + return 0; >> +} >> + >> +static void ipp_subdrv_close(struct drm_device *drm_dev, struct device >> *dev, >> + struct drm_file *file) >> +{ >> + struct drm_exynos_file_private *file_priv = file->driver_priv; >> + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; >> + struct exynos_drm_ippdrv *ippdrv = NULL; >> + struct drm_exynos_ipp_cmd_node *c_node, *tc_node; >> + int count = 0; >> + >> + DRM_DEBUG_KMS("%s:for priv[0x%x]\n", __func__, (int)priv); >> + >> + if (list_empty(&exynos_drm_ippdrv_list)) { >> + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); >> + goto err_clear; >> + } >> + >> + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { >> + if (list_empty(&ippdrv->cmd_list)) >> + continue; >> + >> + list_for_each_entry_safe(c_node, tc_node, >> + &ippdrv->cmd_list, list) { >> + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", >> + __func__, count++, (int)ippdrv); >> + >> + if (c_node->priv == priv) { >> + /* >> + * userland goto unnormal state. process > killed. >> + * and close the file. >> + * so, IPP didn't called stop cmd ctrl. >> + * so, we are make stop operation in this > state. >> + */ >> + if (c_node->state == IPP_STATE_START) { >> + ipp_stop_property(drm_dev, ippdrv, >> + c_node); >> + c_node->state = IPP_STATE_STOP; >> + } >> + >> + ippdrv->dedicated = false; >> + ipp_clean_cmd_node(c_node); >> + if (list_empty(&ippdrv->cmd_list)) >> + pm_runtime_put_sync(ippdrv->dev); >> + } >> + } >> + } >> + >> +err_clear: >> + kfree(priv); >> + return; >> +} >> + >> +static int __devinit ipp_probe(struct platform_device *pdev) >> +{ >> + struct device *dev = &pdev->dev; >> + struct ipp_context *ctx; >> + struct exynos_drm_subdrv *subdrv; >> + int ret; >> + >> + ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); >> + if (!ctx) >> + return -ENOMEM; >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + mutex_init(&ctx->ipp_lock); >> + mutex_init(&ctx->prop_lock); >> + >> + idr_init(&ctx->ipp_idr); >> + idr_init(&ctx->prop_idr); >> + >> + /* >> + * create single thread for ipp event >> + * IPP supports event thread for IPP drivers. >> + * IPP driver send event_work to this thread. >> + * and IPP event thread send event to user process. >> + */ >> + ctx->event_workq = create_singlethread_workqueue("ipp_event"); >> + if (!ctx->event_workq) { >> + dev_err(dev, "failed to create event workqueue\n"); >> + ret = -EINVAL; >> + goto err_clear; >> + } >> + >> + /* >> + * create single thread for ipp command >> + * IPP supports command thread for user process. >> + * user process make command node using set property ioctl. >> + * and make start_work and send this work to command thread. >> + * and then this command thread start property. >> + */ >> + ctx->cmd_workq = create_singlethread_workqueue("ipp_cmd"); >> + if (!ctx->cmd_workq) { >> + dev_err(dev, "failed to create cmd workqueue\n"); >> + ret = -EINVAL; >> + goto err_event_workq; >> + } >> + >> + /* set sub driver informations */ >> + subdrv = &ctx->subdrv; >> + subdrv->dev = dev; >> + subdrv->probe = ipp_subdrv_probe; >> + subdrv->remove = ipp_subdrv_remove; >> + subdrv->open = ipp_subdrv_open; >> + subdrv->close = ipp_subdrv_close; >> + >> + platform_set_drvdata(pdev, ctx); >> + >> + ret = exynos_drm_subdrv_register(subdrv); >> + if (ret < 0) { >> + DRM_ERROR("failed to register drm ipp device.\n"); >> + goto err_cmd_workq; >> + } >> + >> + dev_info(&pdev->dev, "drm ipp registered successfully.\n"); >> + >> + return 0; >> + >> +err_cmd_workq: >> + destroy_workqueue(ctx->cmd_workq); >> +err_event_workq: >> + destroy_workqueue(ctx->event_workq); >> +err_clear: >> + kfree(ctx); >> + return ret; >> +} >> + >> +static int __devexit ipp_remove(struct platform_device *pdev) >> +{ >> + struct ipp_context *ctx = platform_get_drvdata(pdev); >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + /* unregister sub driver */ >> + exynos_drm_subdrv_unregister(&ctx->subdrv); >> + >> + /* remove,destroy ipp idr */ >> + idr_remove_all(&ctx->ipp_idr); >> + idr_remove_all(&ctx->prop_idr); >> + idr_destroy(&ctx->ipp_idr); >> + idr_destroy(&ctx->prop_idr); >> + >> + mutex_destroy(&ctx->ipp_lock); >> + mutex_destroy(&ctx->prop_lock); >> + >> + /* destroy command, event work queue */ >> + destroy_workqueue(ctx->cmd_workq); >> + destroy_workqueue(ctx->event_workq); >> + >> + kfree(ctx); >> + >> + return 0; >> +} >> + >> +static int ipp_power_ctrl(struct ipp_context *ctx, bool enable) >> +{ >> + /* ToDo: Need to implement power and sysmmu ctrl. */ >> + DRM_DEBUG_KMS("%s:enable[%d]\n", __func__, enable); >> + >> + return 0; >> +} >> + >> +#ifdef CONFIG_PM_SLEEP >> +static int ipp_suspend(struct device *dev) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (pm_runtime_suspended(dev)) >> + return 0; >> + >> + return ipp_power_ctrl(ctx, false); >> +} >> + >> +static int ipp_resume(struct device *dev) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + if (!pm_runtime_suspended(dev)) >> + return ipp_power_ctrl(ctx, true); >> + >> + return 0; >> +} >> +#endif >> + >> +#ifdef CONFIG_PM_RUNTIME >> +static int ipp_runtime_suspend(struct device *dev) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + return ipp_power_ctrl(ctx, false); >> +} >> + >> +static int ipp_runtime_resume(struct device *dev) >> +{ >> + struct ipp_context *ctx = get_ipp_context(dev); >> + >> + DRM_DEBUG_KMS("%s\n", __func__); >> + >> + return ipp_power_ctrl(ctx, true); >> +} >> +#endif >> + >> +static const struct dev_pm_ops ipp_pm_ops = { >> + SET_SYSTEM_SLEEP_PM_OPS(ipp_suspend, ipp_resume) >> + SET_RUNTIME_PM_OPS(ipp_runtime_suspend, ipp_runtime_resume, NULL) >> +}; >> + >> +struct platform_driver ipp_driver = { >> + .probe = ipp_probe, >> + .remove = __devexit_p(ipp_remove), >> + .driver = { >> + .name = "exynos-drm-ipp", >> + .owner = THIS_MODULE, >> + .pm = &ipp_pm_ops, >> + }, >> +}; >> + >> diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.h >> b/drivers/gpu/drm/exynos/exynos_drm_ipp.h >> new file mode 100644 >> index 0000000..baab1f0 >> --- /dev/null >> +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.h >> @@ -0,0 +1,266 @@ >> +/* >> + * Copyright (c) 2012 Samsung Electronics Co., Ltd. >> + * >> + * Authors: >> + * Eunchul Kim <chulspro.kim@samsung.com> >> + * Jinyoung Jeon <jy0.jeon@samsung.com> >> + * Sangmin Lee <lsmin.lee@samsung.com> >> + * >> + * Permission is hereby granted, free of charge, to any person obtaining >> a >> + * copy of this software and associated documentation files (the >> "Software"), >> + * to deal in the Software without restriction, including without >> limitation >> + * the rights to use, copy, modify, merge, publish, distribute, >> sublicense, >> + * and/or sell copies of the Software, and to permit persons to whom the >> + * Software is furnished to do so, subject to the following conditions: >> + * >> + * The above copyright notice and this permission notice (including the >> next >> + * paragraph) shall be included in all copies or substantial portions of >> the >> + * Software. >> + * >> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, >> EXPRESS OR >> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF >> MERCHANTABILITY, >> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT >> SHALL >> + * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES >> OR >> + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, >> + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR >> + * OTHER DEALINGS IN THE SOFTWARE. >> + */ >> + >> +#ifndef _EXYNOS_DRM_IPP_H_ >> +#define _EXYNOS_DRM_IPP_H_ >> + >> +#define for_each_ipp_ops(pos) \ >> + for (pos = 0; pos < EXYNOS_DRM_OPS_MAX; pos++) >> +#define for_each_ipp_planar(pos) \ >> + for (pos = 0; pos < EXYNOS_DRM_PLANAR_MAX; pos++) >> + >> +#define IPP_GET_LCD_WIDTH _IOR('F', 302, int) >> +#define IPP_GET_LCD_HEIGHT _IOR('F', 303, int) >> +#define IPP_SET_WRITEBACK _IOW('F', 304, u32) >> + >> +/* definition of state */ >> +enum drm_exynos_ipp_state { >> + IPP_STATE_IDLE, >> + IPP_STATE_START, >> + IPP_STATE_STOP, >> +}; >> + >> +/* >> + * A structure of command work information. >> + * @work: work structure. >> + * @ippdrv: current work ippdrv. >> + * @c_node: command node information. >> + * @ctrl: command control. >> + */ >> +struct drm_exynos_ipp_cmd_work { >> + struct work_struct work; >> + struct exynos_drm_ippdrv *ippdrv; >> + struct drm_exynos_ipp_cmd_node *c_node; >> + enum drm_exynos_ipp_ctrl ctrl; >> +}; >> + >> +/* >> + * A structure of command node. >> + * >> + * @priv: IPP private infomation. >> + * @list: list head to command queue information. >> + * @event_list: list head of event. >> + * @mem_list: list head to source,destination memory queue information. >> + * @cmd_lock: lock for synchronization of access to ioctl. >> + * @mem_lock: lock for synchronization of access to memory nodes. >> + * @event_lock: lock for synchronization of access to scheduled event. >> + * @start_complete: completion of start of command. >> + * @stop_complete: completion of stop of command. >> + * @property: property information. >> + * @start_work: start command work structure. >> + * @stop_work: stop command work structure. >> + * @event_work: event work structure. >> + * @state: state of command node. >> + */ >> +struct drm_exynos_ipp_cmd_node { >> + struct exynos_drm_ipp_private *priv; >> + struct list_head list; >> + struct list_head event_list; >> + struct list_head mem_list[EXYNOS_DRM_OPS_MAX]; >> + struct mutex cmd_lock; >> + struct mutex mem_lock; >> + struct mutex event_lock; >> + struct completion start_complete; >> + struct completion stop_complete; >> + struct drm_exynos_ipp_property property; >> + struct drm_exynos_ipp_cmd_work *start_work; >> + struct drm_exynos_ipp_cmd_work *stop_work; >> + struct drm_exynos_ipp_event_work *event_work; >> + enum drm_exynos_ipp_state state; >> +}; >> + >> +/* >> + * A structure of buffer information. >> + * >> + * @gem_objs: Y, Cb, Cr each gem object. >> + * @base: Y, Cb, Cr each planar address. >> + */ >> +struct drm_exynos_ipp_buf_info { >> + unsigned long handles[EXYNOS_DRM_PLANAR_MAX]; >> + dma_addr_t base[EXYNOS_DRM_PLANAR_MAX]; >> +}; >> + >> +/* >> + * A structure of wb setting infomation. >> + * >> + * @enable: enable flag for wb. >> + * @refresh: HZ of the refresh rate. >> + */ >> +struct drm_exynos_ipp_set_wb { >> + __u32 enable; >> + __u32 refresh; >> +}; >> + >> +/* >> + * A structure of event work information. >> + * >> + * @work: work structure. >> + * @ippdrv: current work ippdrv. >> + * @buf_id: id of src, dst buffer. >> + */ >> +struct drm_exynos_ipp_event_work { >> + struct work_struct work; >> + struct exynos_drm_ippdrv *ippdrv; >> + u32 buf_id[EXYNOS_DRM_OPS_MAX]; >> +}; >> + >> +/* >> + * A structure of source,destination operations. >> + * >> + * @set_fmt: set format of image. >> + * @set_transf: set transform(rotations, flip). >> + * @set_size: set size of region. >> + * @set_addr: set address for dma. >> + */ >> +struct exynos_drm_ipp_ops { >> + int (*set_fmt)(struct device *dev, u32 fmt); >> + int (*set_transf)(struct device *dev, >> + enum drm_exynos_degree degree, >> + enum drm_exynos_flip flip); >> + int (*set_size)(struct device *dev, int swap, >> + struct drm_exynos_pos *pos, struct drm_exynos_sz *sz); >> + int (*set_addr)(struct device *dev, >> + struct drm_exynos_ipp_buf_info *buf_info, u32 > buf_id, >> + enum drm_exynos_ipp_buf_type buf_type); >> +}; >> + >> +/* >> + * A structure of ipp driver. >> + * >> + * @drv_list: list head for registed sub driver information. >> + * @parent_dev: parent device information. >> + * @dev: platform device. >> + * @drm_dev: drm device. >> + * @ipp_id: id of ipp driver. >> + * @dedicated: dedicated ipp device. >> + * @ops: source, destination operations. >> + * @event_workq: event work queue. >> + * @cmd: current command information. >> + * @cmd_list: list head for command information. >> + * @prop_list: property informations of current ipp driver. >> + * @check_property: check property about format, size, buffer. >> + * @reset: reset ipp block. >> + * @start: ipp each device start. >> + * @stop: ipp each device stop. >> + * @sched_event: work schedule handler. >> + */ >> +struct exynos_drm_ippdrv { >> + struct list_head drv_list; >> + struct device *parent_dev; >> + struct device *dev; >> + struct drm_device *drm_dev; >> + u32 ipp_id; >> + bool dedicated; >> + struct exynos_drm_ipp_ops *ops[EXYNOS_DRM_OPS_MAX]; >> + struct workqueue_struct *event_workq; >> + struct drm_exynos_ipp_cmd_node *cmd; >> + struct list_head cmd_list; >> + struct drm_exynos_ipp_prop_list *prop_list; >> + >> + int (*check_property)(struct device *dev, >> + struct drm_exynos_ipp_property *property); >> + int (*reset)(struct device *dev); >> + int (*start)(struct device *dev, enum drm_exynos_ipp_cmd cmd); >> + void (*stop)(struct device *dev, enum drm_exynos_ipp_cmd cmd); >> + void (*sched_event)(struct work_struct *work); >> +}; >> + >> +#ifdef CONFIG_DRM_EXYNOS_IPP >> +extern int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv); >> +extern int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv > *ippdrv); >> +extern int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void >> *data, >> + struct drm_file *file); >> +extern int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void >> *data, >> + struct drm_file *file); >> +extern int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void >> *data, >> + struct drm_file *file); >> +extern int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void > *data, >> + struct drm_file *file); >> +extern int exynos_drm_ippnb_register(struct notifier_block *nb); >> +extern int exynos_drm_ippnb_unregister(struct notifier_block *nb); >> +extern int exynos_drm_ippnb_send_event(unsigned long val, void *v); >> +extern void ipp_sched_cmd(struct work_struct *work); >> +extern void ipp_sched_event(struct work_struct *work); >> + >> +#else >> +static inline int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv >> *ippdrv) >> +{ >> + return -ENODEV; >> +} >> + >> +static inline int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv >> *ippdrv) >> +{ >> + return -ENODEV; >> +} >> + >> +static inline int exynos_drm_ipp_get_property(struct drm_device *drm_dev, >> + void *data, >> + struct drm_file *file_priv) >> +{ >> + return -ENOTTY; >> +} >> + >> +static inline int exynos_drm_ipp_set_property(struct drm_device *drm_dev, >> + void *data, >> + struct drm_file *file_priv) >> +{ >> + return -ENOTTY; >> +} >> + >> +static inline int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, >> + void *data, >> + struct drm_file *file) >> +{ >> + return -ENOTTY; >> +} >> + >> +static inline int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, >> + void *data, >> + struct drm_file *file) >> +{ >> + return -ENOTTY; >> +} >> + >> +static inline int exynos_drm_ippnb_register(struct notifier_block *nb) >> +{ >> + return -ENODEV; >> +} >> + >> +static inline int exynos_drm_ippnb_unregister(struct notifier_block *nb) >> +{ >> + return -ENODEV; >> +} >> + >> +static inline int exynos_drm_ippnb_send_event(unsigned long val, void *v) >> +{ >> + return -ENOTTY; >> +} >> +#endif >> + >> +#endif /* _EXYNOS_DRM_IPP_H_ */ >> + >> diff --git a/include/uapi/drm/exynos_drm.h b/include/uapi/drm/exynos_drm.h >> index c0494d5..82772d7 100644 >> --- a/include/uapi/drm/exynos_drm.h >> +++ b/include/uapi/drm/exynos_drm.h >> @@ -154,6 +154,170 @@ struct drm_exynos_g2d_exec { >> __u64 async; >> }; >> >> +enum drm_exynos_ops_id { >> + EXYNOS_DRM_OPS_SRC, >> + EXYNOS_DRM_OPS_DST, >> + EXYNOS_DRM_OPS_MAX, >> +}; >> + >> +struct drm_exynos_sz { >> + __u32 hsize; >> + __u32 vsize; >> +}; >> + >> +struct drm_exynos_pos { >> + __u32 x; >> + __u32 y; >> + __u32 w; >> + __u32 h; >> +}; >> + >> +enum drm_exynos_flip { >> + EXYNOS_DRM_FLIP_NONE = (0 << 0), >> + EXYNOS_DRM_FLIP_VERTICAL = (1 << 0), >> + EXYNOS_DRM_FLIP_HORIZONTAL = (1 << 1), >> +}; >> + >> +enum drm_exynos_degree { >> + EXYNOS_DRM_DEGREE_0, >> + EXYNOS_DRM_DEGREE_90, >> + EXYNOS_DRM_DEGREE_180, >> + EXYNOS_DRM_DEGREE_270, >> +}; >> + >> +enum drm_exynos_planer { >> + EXYNOS_DRM_PLANAR_Y, >> + EXYNOS_DRM_PLANAR_CB, >> + EXYNOS_DRM_PLANAR_CR, >> + EXYNOS_DRM_PLANAR_MAX, >> +}; >> + >> +/** >> + * A structure for ipp supported property list. >> + * >> + * @version: version of this structure. >> + * @ipp_id: id of ipp driver. >> + * @count: count of ipp driver. >> + * @writeback: flag of writeback supporting. >> + * @flip: flag of flip supporting. >> + * @degree: flag of degree information. >> + * @csc: flag of csc supporting. >> + * @crop: flag of crop supporting. >> + * @scale: flag of scale supporting. >> + * @refresh_min: min hz of refresh. >> + * @refresh_max: max hz of refresh. >> + * @crop_min: crop min resolution. >> + * @crop_max: crop max resolution. >> + * @scale_min: scale min resolution. >> + * @scale_max: scale max resolution. >> + */ >> +struct drm_exynos_ipp_prop_list { >> + __u32 version; >> + __u32 ipp_id; >> + __u32 count; >> + __u32 writeback; >> + __u32 flip; >> + __u32 degree; >> + __u32 csc; >> + __u32 crop; >> + __u32 scale; >> + __u32 refresh_min; >> + __u32 refresh_max; >> + __u32 reserved; >> + struct drm_exynos_sz crop_min; >> + struct drm_exynos_sz crop_max; >> + struct drm_exynos_sz scale_min; >> + struct drm_exynos_sz scale_max; >> +}; >> + >> +/** >> + * A structure for ipp config. >> + * >> + * @ops_id: property of operation directions. >> + * @flip: property of mirror, flip. >> + * @degree: property of rotation degree. >> + * @fmt: property of image format. >> + * @sz: property of image size. >> + * @pos: property of image position(src-cropped,dst-scaler). >> + */ >> +struct drm_exynos_ipp_config { >> + enum drm_exynos_ops_id ops_id; >> + enum drm_exynos_flip flip; >> + enum drm_exynos_degree degree; >> + __u32 fmt; >> + struct drm_exynos_sz sz; >> + struct drm_exynos_pos pos; >> +}; >> + >> +enum drm_exynos_ipp_cmd { >> + IPP_CMD_NONE, >> + IPP_CMD_M2M, >> + IPP_CMD_WB, >> + IPP_CMD_OUTPUT, >> + IPP_CMD_MAX, >> +}; >> + >> +/** >> + * A structure for ipp property. >> + * >> + * @config: source, destination config. >> + * @cmd: definition of command. >> + * @ipp_id: id of ipp driver. >> + * @prop_id: id of property. >> + * @refresh_rate: refresh rate. >> + */ >> +struct drm_exynos_ipp_property { >> + struct drm_exynos_ipp_config config[EXYNOS_DRM_OPS_MAX]; >> + enum drm_exynos_ipp_cmd cmd; >> + __u32 ipp_id; >> + __u32 prop_id; >> + __u32 refresh_rate; >> +}; >> + >> +enum drm_exynos_ipp_buf_type { >> + IPP_BUF_ENQUEUE, >> + IPP_BUF_DEQUEUE, >> +}; >> + >> +/** >> + * A structure for ipp buffer operations. >> + * >> + * @ops_id: operation directions. >> + * @buf_type: definition of buffer. >> + * @prop_id: id of property. >> + * @buf_id: id of buffer. >> + * @handle: Y, Cb, Cr each planar handle. >> + * @user_data: user data. >> + */ >> +struct drm_exynos_ipp_queue_buf { >> + enum drm_exynos_ops_id ops_id; >> + enum drm_exynos_ipp_buf_type buf_type; >> + __u32 prop_id; >> + __u32 buf_id; >> + __u32 handle[EXYNOS_DRM_PLANAR_MAX]; >> + __u32 reserved; >> + __u64 user_data; >> +}; >> + >> +enum drm_exynos_ipp_ctrl { >> + IPP_CTRL_PLAY, >> + IPP_CTRL_STOP, >> + IPP_CTRL_PAUSE, >> + IPP_CTRL_RESUME, >> + IPP_CTRL_MAX, >> +}; >> + >> +/** >> + * A structure for ipp start/stop operations. >> + * >> + * @prop_id: id of property. >> + * @ctrl: definition of control. >> + */ >> +struct drm_exynos_ipp_cmd_ctrl { >> + __u32 prop_id; >> + enum drm_exynos_ipp_ctrl ctrl; >> +}; >> + >> #define DRM_EXYNOS_GEM_CREATE 0x00 >> #define DRM_EXYNOS_GEM_MAP_OFFSET 0x01 >> #define DRM_EXYNOS_GEM_MMAP 0x02 >> @@ -166,6 +330,12 @@ struct drm_exynos_g2d_exec { >> #define DRM_EXYNOS_G2D_SET_CMDLIST 0x21 >> #define DRM_EXYNOS_G2D_EXEC 0x22 >> >> +/* IPP - Image Post Processing */ >> +#define DRM_EXYNOS_IPP_GET_PROPERTY 0x30 >> +#define DRM_EXYNOS_IPP_SET_PROPERTY 0x31 >> +#define DRM_EXYNOS_IPP_QUEUE_BUF 0x32 >> +#define DRM_EXYNOS_IPP_CMD_CTRL 0x33 >> + >> #define DRM_IOCTL_EXYNOS_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + >> \ >> DRM_EXYNOS_GEM_CREATE, struct drm_exynos_gem_create) >> >> @@ -188,8 +358,18 @@ struct drm_exynos_g2d_exec { >> #define DRM_IOCTL_EXYNOS_G2D_EXEC DRM_IOWR(DRM_COMMAND_BASE + >> \ >> DRM_EXYNOS_G2D_EXEC, struct drm_exynos_g2d_exec) >> >> +#define DRM_IOCTL_EXYNOS_IPP_GET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + >> \ >> + DRM_EXYNOS_IPP_GET_PROPERTY, struct > drm_exynos_ipp_prop_list) >> +#define DRM_IOCTL_EXYNOS_IPP_SET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + >> \ >> + DRM_EXYNOS_IPP_SET_PROPERTY, struct drm_exynos_ipp_property) >> +#define DRM_IOCTL_EXYNOS_IPP_QUEUE_BUF DRM_IOWR(DRM_COMMAND_BASE + > \ >> + DRM_EXYNOS_IPP_QUEUE_BUF, struct drm_exynos_ipp_queue_buf) >> +#define DRM_IOCTL_EXYNOS_IPP_CMD_CTRL > DRM_IOWR(DRM_COMMAND_BASE + >> \ >> + DRM_EXYNOS_IPP_CMD_CTRL, struct drm_exynos_ipp_cmd_ctrl) >> + >> /* EXYNOS specific events */ >> #define DRM_EXYNOS_G2D_EVENT 0x80000000 >> +#define DRM_EXYNOS_IPP_EVENT 0x80000001 >> >> struct drm_exynos_g2d_event { >> struct drm_event base; >> @@ -200,4 +380,14 @@ struct drm_exynos_g2d_event { >> __u32 reserved; >> }; >> >> +struct drm_exynos_ipp_event { >> + struct drm_event base; >> + __u64 user_data; >> + __u32 tv_sec; >> + __u32 tv_usec; >> + __u32 prop_id; >> + __u32 reserved; >> + __u32 buf_id[EXYNOS_DRM_OPS_MAX]; >> +}; >> + >> #endif /* _UAPI_EXYNOS_DRM_H_ */ >> -- >> 1.7.0.4 > >
diff --git a/drivers/gpu/drm/exynos/Kconfig b/drivers/gpu/drm/exynos/Kconfig index 4ea8cdc..bcf1c9d 100644 --- a/drivers/gpu/drm/exynos/Kconfig +++ b/drivers/gpu/drm/exynos/Kconfig @@ -45,3 +45,9 @@ config DRM_EXYNOS_G2D depends on DRM_EXYNOS && !VIDEO_SAMSUNG_S5P_G2D help Choose this option if you want to use Exynos G2D for DRM. + +config DRM_EXYNOS_IPP + bool "Exynos DRM IPP" + depends on DRM_EXYNOS + help + Choose this option if you want to use IPP feature for DRM. diff --git a/drivers/gpu/drm/exynos/Makefile b/drivers/gpu/drm/exynos/Makefile index 26813b8..6c536ce 100644 --- a/drivers/gpu/drm/exynos/Makefile +++ b/drivers/gpu/drm/exynos/Makefile @@ -16,5 +16,6 @@ exynosdrm-$(CONFIG_DRM_EXYNOS_HDMI) += exynos_hdmi.o exynos_mixer.o \ exynos_drm_hdmi.o exynosdrm-$(CONFIG_DRM_EXYNOS_VIDI) += exynos_drm_vidi.o exynosdrm-$(CONFIG_DRM_EXYNOS_G2D) += exynos_drm_g2d.o +exynosdrm-$(CONFIG_DRM_EXYNOS_IPP) += exynos_drm_ipp.o obj-$(CONFIG_DRM_EXYNOS) += exynosdrm.o diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.c b/drivers/gpu/drm/exynos/exynos_drm_drv.c index 4a1168d..0eb8a97 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_drv.c +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.c @@ -40,6 +40,7 @@ #include "exynos_drm_vidi.h" #include "exynos_drm_dmabuf.h" #include "exynos_drm_g2d.h" +#include "exynos_drm_ipp.h" #include "exynos_drm_iommu.h" #define DRIVER_NAME "exynos" @@ -249,6 +250,14 @@ static struct drm_ioctl_desc exynos_ioctls[] = { exynos_g2d_set_cmdlist_ioctl, DRM_UNLOCKED | DRM_AUTH), DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, exynos_g2d_exec_ioctl, DRM_UNLOCKED | DRM_AUTH), + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, + exynos_drm_ipp_get_property, DRM_UNLOCKED | DRM_AUTH), + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, + exynos_drm_ipp_set_property, DRM_UNLOCKED | DRM_AUTH), + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, + exynos_drm_ipp_queue_buf, DRM_UNLOCKED | DRM_AUTH), + DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, + exynos_drm_ipp_cmd_ctrl, DRM_UNLOCKED | DRM_AUTH), }; static const struct file_operations exynos_drm_driver_fops = { @@ -363,6 +372,12 @@ static int __init exynos_drm_init(void) goto out_g2d; #endif +#ifdef CONFIG_DRM_EXYNOS_IPP + ret = platform_driver_register(&ipp_driver); + if (ret < 0) + goto out_ipp; +#endif + ret = platform_driver_register(&exynos_drm_platform_driver); if (ret < 0) goto out_drm; @@ -380,6 +395,11 @@ out: platform_driver_unregister(&exynos_drm_platform_driver); out_drm: +#ifdef CONFIG_DRM_EXYNOS_IPP + platform_driver_unregister(&ipp_driver); +out_ipp: +#endif + #ifdef CONFIG_DRM_EXYNOS_G2D platform_driver_unregister(&g2d_driver); out_g2d: @@ -416,6 +436,10 @@ static void __exit exynos_drm_exit(void) platform_driver_unregister(&exynos_drm_platform_driver); +#ifdef CONFIG_DRM_EXYNOS_IPP + platform_driver_unregister(&ipp_driver); +#endif + #ifdef CONFIG_DRM_EXYNOS_G2D platform_driver_unregister(&g2d_driver); #endif diff --git a/drivers/gpu/drm/exynos/exynos_drm_drv.h b/drivers/gpu/drm/exynos/exynos_drm_drv.h index a9db025..a365788 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_drv.h +++ b/drivers/gpu/drm/exynos/exynos_drm_drv.h @@ -235,8 +235,14 @@ struct exynos_drm_g2d_private { unsigned int gem_nr; }; +struct exynos_drm_ipp_private { + struct device *dev; + struct list_head event_list; +}; + struct drm_exynos_file_private { struct exynos_drm_g2d_private *g2d_priv; + struct exynos_drm_ipp_private *ipp_priv; }; /* @@ -346,4 +352,5 @@ extern struct platform_driver mixer_driver; extern struct platform_driver exynos_drm_common_hdmi_driver; extern struct platform_driver vidi_driver; extern struct platform_driver g2d_driver; +extern struct platform_driver ipp_driver; #endif diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.c b/drivers/gpu/drm/exynos/exynos_drm_ipp.c new file mode 100644 index 0000000..eb12cb5 --- /dev/null +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.c @@ -0,0 +1,2059 @@ +/* + * Copyright (C) 2012 Samsung Electronics Co.Ltd + * Authors: + * Eunchul Kim <chulspro.kim@samsung.com> + * Jinyoung Jeon <jy0.jeon@samsung.com> + * Sangmin Lee <lsmin.lee@samsung.com> + * + * This program is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License as published by the + * Free Software Foundation; either version 2 of the License, or (at your + * option) any later version. + * + */ +#include <linux/kernel.h> +#include <linux/module.h> +#include <linux/platform_device.h> +#include <linux/clk.h> +#include <linux/pm_runtime.h> +#include <plat/map-base.h> + +#include <drm/drmP.h> +#include <drm/exynos_drm.h> +#include "exynos_drm_drv.h" +#include "exynos_drm_gem.h" +#include "exynos_drm_ipp.h" + +/* + * IPP is stand for Image Post Processing and + * supports image scaler/rotator and input/output DMA operations. + * using FIMC, GSC, Rotator, so on. + * IPP is integration device driver of same attribute h/w + */ + +#define get_ipp_context(dev) platform_get_drvdata(to_platform_device(dev)) +#define ipp_is_m2m_cmd(c) (c == IPP_CMD_M2M) + +/* + * A structure of event. + * + * @base: base of event. + * @event: ipp event. + */ +struct drm_exynos_ipp_send_event { + struct drm_pending_event base; + struct drm_exynos_ipp_event event; +}; + +/* + * A structure of memory node. + * + * @list: list head to memory queue information. + * @ops_id: id of operations. + * @prop_id: id of property. + * @buf_id: id of buffer. + * @buf_info: gem objects and dma address, size. + * @filp: a pointer to drm_file. + */ +struct drm_exynos_ipp_mem_node { + struct list_head list; + enum drm_exynos_ops_id ops_id; + u32 prop_id; + u32 buf_id; + struct drm_exynos_ipp_buf_info buf_info; + struct drm_file *filp; +}; + +/* + * A structure of ipp context. + * + * @subdrv: prepare initialization using subdrv. + * @ipp_lock: lock for synchronization of access to ipp_idr. + * @prop_lock: lock for synchronization of access to prop_idr. + * @ipp_idr: ipp driver idr. + * @prop_idr: property idr. + * @event_workq: event work queue. + * @cmd_workq: command work queue. + */ +struct ipp_context { + struct exynos_drm_subdrv subdrv; + struct mutex ipp_lock; + struct mutex prop_lock; + struct idr ipp_idr; + struct idr prop_idr; + struct workqueue_struct *event_workq; + struct workqueue_struct *cmd_workq; +}; + +static LIST_HEAD(exynos_drm_ippdrv_list); +static DEFINE_MUTEX(exynos_drm_ippdrv_lock); +static BLOCKING_NOTIFIER_HEAD(exynos_drm_ippnb_list); + +int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv) +{ + DRM_DEBUG_KMS("%s\n", __func__); + + if (!ippdrv) + return -EINVAL; + + mutex_lock(&exynos_drm_ippdrv_lock); + list_add_tail(&ippdrv->drv_list, &exynos_drm_ippdrv_list); + mutex_unlock(&exynos_drm_ippdrv_lock); + + return 0; +} +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_register); + +int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv) +{ + DRM_DEBUG_KMS("%s\n", __func__); + + if (!ippdrv) + return -EINVAL; + + mutex_lock(&exynos_drm_ippdrv_lock); + list_del(&ippdrv->drv_list); + mutex_unlock(&exynos_drm_ippdrv_lock); + + return 0; +} +EXPORT_SYMBOL_GPL(exynos_drm_ippdrv_unregister); + +static int ipp_create_id(struct idr *id_idr, struct mutex *lock, void *obj, + u32 *idp) +{ + int ret; + + DRM_DEBUG_KMS("%s\n", __func__); + +again: + /* ensure there is space available to allocate a handle */ + if (idr_pre_get(id_idr, GFP_KERNEL) == 0) { + DRM_ERROR("failed to get idr.\n"); + return -ENOMEM; + } + + /* do the allocation under our mutexlock */ + mutex_lock(lock); + ret = idr_get_new_above(id_idr, obj, 1, (int *)idp); + mutex_unlock(lock); + if (ret == -EAGAIN) + goto again; + + return ret; +} + +static void *ipp_find_obj(struct idr *id_idr, struct mutex *lock, u32 id) +{ + void *obj; + + DRM_DEBUG_KMS("%s:id[%d]\n", __func__, id); + + mutex_lock(lock); + + /* find object using handle */ + obj = idr_find(id_idr, id); + if (!obj) { + DRM_ERROR("failed to find object.\n"); + mutex_unlock(lock); + return ERR_PTR(-ENODEV); + } + + mutex_unlock(lock); + + return obj; +} + +static inline bool ipp_check_dedicated(struct exynos_drm_ippdrv *ippdrv, + enum drm_exynos_ipp_cmd cmd) +{ + /* + * check dedicated flag and WB, OUTPUT operation with + * power on state. + */ + if (ippdrv->dedicated || (!ipp_is_m2m_cmd(cmd) && + !pm_runtime_suspended(ippdrv->dev))) + return true; + + return false; +} + +static struct exynos_drm_ippdrv *ipp_find_driver(struct ipp_context *ctx, + struct drm_exynos_ipp_property *property) +{ + struct exynos_drm_ippdrv *ippdrv; + u32 ipp_id = property->ipp_id; + + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, ipp_id); + + if (ipp_id) { + /* find ipp driver using idr */ + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, + ipp_id); + if (IS_ERR_OR_NULL(ippdrv)) { + DRM_ERROR("not found ipp%d driver.\n", ipp_id); + goto err_null; + } + + /* + * WB, OUTPUT opertion not supported multi-operation. + * so, make dedicated state at set property ioctl. + * when ipp driver finished operations, clear dedicated flags. + */ + if (ipp_check_dedicated(ippdrv, property->cmd)) { + DRM_ERROR("already used choose device.\n"); + goto err_null; + } + + /* + * This is necessary to find correct device in ipp drivers. + * ipp drivers have different abilities, + * so need to check property. + */ + if (ippdrv->check_property && + ippdrv->check_property(ippdrv->dev, property)) { + DRM_ERROR("not support property.\n"); + goto err_null; + } + + return ippdrv; + } else { + /* + * This case is search all ipp driver for finding. + * user application don't set ipp_id in this case, + * so ipp subsystem search correct driver in driver list. + */ + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { + if (ipp_check_dedicated(ippdrv, property->cmd)) { + DRM_DEBUG_KMS("%s:used device.\n", __func__); + continue; + } + + if (ippdrv->check_property && + ippdrv->check_property(ippdrv->dev, property)) { + DRM_DEBUG_KMS("%s:not support property.\n", + __func__); + continue; + } + + return ippdrv; + } + + DRM_ERROR("not support ipp driver operations.\n"); + } + + return ERR_PTR(-ENODEV); + +err_null: + return NULL; +} + +static struct exynos_drm_ippdrv *ipp_find_drv_by_handle(u32 prop_id) +{ + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + int count = 0; + + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); + + if (list_empty(&exynos_drm_ippdrv_list)) { + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); + return ERR_PTR(-ENODEV); + } + + /* + * This case is search ipp driver by prop_id handle. + * sometimes, ipp subsystem find driver by prop_id. + * e.g PAUSE state, queue buf, command contro. + */ + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", __func__, + count++, (int)ippdrv); + + if (!list_empty(&ippdrv->cmd_list)) { + list_for_each_entry(c_node, &ippdrv->cmd_list, list) + if (c_node->property.prop_id == prop_id) + return ippdrv; + } + } + + return ERR_PTR(-ENODEV); +} + +int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void *data, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; + struct device *dev = priv->dev; + struct ipp_context *ctx = get_ipp_context(dev); + struct drm_exynos_ipp_prop_list *prop_list = data; + struct exynos_drm_ippdrv *ippdrv; + int count = 0; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!ctx) { + DRM_ERROR("invalid context.\n"); + return -EINVAL; + } + + if (!prop_list) { + DRM_ERROR("invalid property parameter.\n"); + return -EINVAL; + } + + DRM_DEBUG_KMS("%s:ipp_id[%d]\n", __func__, prop_list->ipp_id); + + if (!prop_list->ipp_id) { + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) + count++; + /* + * Supports ippdrv list count for user application. + * First step user application getting ippdrv count. + * and second step getting ippdrv capability using ipp_id. + */ + prop_list->count = count; + } else { + /* + * Getting ippdrv capability by ipp_id. + * some deivce not supported wb, output interface. + * so, user application detect correct ipp driver + * using this ioctl. + */ + ippdrv = ipp_find_obj(&ctx->ipp_idr, &ctx->ipp_lock, + prop_list->ipp_id); + if (!ippdrv) { + DRM_ERROR("not found ipp%d driver.\n", + prop_list->ipp_id); + return -EINVAL; + } + + prop_list = ippdrv->prop_list; + } + + return 0; +} +EXPORT_SYMBOL_GPL(exynos_drm_ipp_get_property); + +static void ipp_print_property(struct drm_exynos_ipp_property *property, + int idx) +{ + struct drm_exynos_ipp_config *config = &property->config[idx]; + struct drm_exynos_pos *pos = &config->pos; + struct drm_exynos_sz *sz = &config->sz; + + DRM_DEBUG_KMS("%s:prop_id[%d]ops[%s]fmt[0x%x]\n", + __func__, property->prop_id, idx ? "dst" : "src", config->fmt); + + DRM_DEBUG_KMS("%s:pos[%d %d %d %d]sz[%d %d]f[%d]r[%d]\n", + __func__, pos->x, pos->y, pos->w, pos->h, + sz->hsize, sz->vsize, config->flip, config->degree); +} + +static int ipp_find_and_set_property(struct drm_exynos_ipp_property *property) +{ + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + u32 prop_id = property->prop_id; + + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, prop_id); + + ippdrv = ipp_find_drv_by_handle(prop_id); + if (IS_ERR_OR_NULL(ippdrv)) { + DRM_ERROR("failed to get ipp driver.\n"); + return -EINVAL; + } + + /* + * Find command node using command list in ippdrv. + * when we find this command no using prop_id. + * return property information set in this command node. + */ + list_for_each_entry(c_node, &ippdrv->cmd_list, list) { + if ((c_node->property.prop_id == prop_id) && + (c_node->state == IPP_STATE_STOP)) { + DRM_DEBUG_KMS("%s:found cmd[%d]ippdrv[0x%x]\n", + __func__, property->cmd, (int)ippdrv); + + c_node->property = *property; + return 0; + } + } + + DRM_ERROR("failed to search property.\n"); + + return -EINVAL; +} + +static struct drm_exynos_ipp_cmd_work *ipp_create_cmd_work(void) +{ + struct drm_exynos_ipp_cmd_work *cmd_work; + + DRM_DEBUG_KMS("%s\n", __func__); + + cmd_work = kzalloc(sizeof(*cmd_work), GFP_KERNEL); + if (!cmd_work) { + DRM_ERROR("failed to alloc cmd_work.\n"); + return ERR_PTR(-ENOMEM); + } + + INIT_WORK((struct work_struct *)cmd_work, ipp_sched_cmd); + + return cmd_work; +} + +static struct drm_exynos_ipp_event_work *ipp_create_event_work(void) +{ + struct drm_exynos_ipp_event_work *event_work; + + DRM_DEBUG_KMS("%s\n", __func__); + + event_work = kzalloc(sizeof(*event_work), GFP_KERNEL); + if (!event_work) { + DRM_ERROR("failed to alloc event_work.\n"); + return ERR_PTR(-ENOMEM); + } + + INIT_WORK((struct work_struct *)event_work, ipp_sched_event); + + return event_work; +} + +int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void *data, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; + struct device *dev = priv->dev; + struct ipp_context *ctx = get_ipp_context(dev); + struct drm_exynos_ipp_property *property = data; + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + int ret, i; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!ctx) { + DRM_ERROR("invalid context.\n"); + return -EINVAL; + } + + if (!property) { + DRM_ERROR("invalid property parameter.\n"); + return -EINVAL; + } + + /* + * This is log print for user application property. + * user application set various property. + */ + for_each_ipp_ops(i) + ipp_print_property(property, i); + + /* + * set property ioctl generated new prop_id. + * but in this case already asigned prop_id using old set property. + * e.g PAUSE state. this case supports find current prop_id and use it + * instead of allocation. + */ + if (property->prop_id) { + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); + return ipp_find_and_set_property(property); + } + + /* find ipp driver using ipp id */ + ippdrv = ipp_find_driver(ctx, property); + if (IS_ERR_OR_NULL(ippdrv)) { + DRM_ERROR("failed to get ipp driver.\n"); + return -EINVAL; + } + + /* allocate command node */ + c_node = kzalloc(sizeof(*c_node), GFP_KERNEL); + if (!c_node) { + DRM_ERROR("failed to allocate map node.\n"); + return -ENOMEM; + } + + /* create property id */ + ret = ipp_create_id(&ctx->prop_idr, &ctx->prop_lock, c_node, + &property->prop_id); + if (ret) { + DRM_ERROR("failed to create id.\n"); + goto err_clear; + } + + DRM_DEBUG_KMS("%s:created prop_id[%d]cmd[%d]ippdrv[0x%x]\n", + __func__, property->prop_id, property->cmd, (int)ippdrv); + + /* stored property information and ippdrv in private data */ + c_node->priv = priv; + c_node->property = *property; + c_node->state = IPP_STATE_IDLE; + + c_node->start_work = ipp_create_cmd_work(); + if (IS_ERR_OR_NULL(c_node->start_work)) { + DRM_ERROR("failed to create start work.\n"); + goto err_clear; + } + + c_node->stop_work = ipp_create_cmd_work(); + if (IS_ERR_OR_NULL(c_node->stop_work)) { + DRM_ERROR("failed to create stop work.\n"); + goto err_free_start; + } + + c_node->event_work = ipp_create_event_work(); + if (IS_ERR_OR_NULL(c_node->event_work)) { + DRM_ERROR("failed to create event work.\n"); + goto err_free_stop; + } + + mutex_init(&c_node->cmd_lock); + mutex_init(&c_node->mem_lock); + mutex_init(&c_node->event_lock); + + init_completion(&c_node->start_complete); + init_completion(&c_node->stop_complete); + + for_each_ipp_ops(i) + INIT_LIST_HEAD(&c_node->mem_list[i]); + + INIT_LIST_HEAD(&c_node->event_list); + list_splice_init(&priv->event_list, &c_node->event_list); + list_add_tail(&c_node->list, &ippdrv->cmd_list); + + /* make dedicated state without m2m */ + if (!ipp_is_m2m_cmd(property->cmd)) + ippdrv->dedicated = true; + + return 0; + +err_free_stop: + kfree(c_node->stop_work); +err_free_start: + kfree(c_node->start_work); +err_clear: + kfree(c_node); + return ret; +} +EXPORT_SYMBOL_GPL(exynos_drm_ipp_set_property); + +static void ipp_clean_cmd_node(struct drm_exynos_ipp_cmd_node *c_node) +{ + DRM_DEBUG_KMS("%s\n", __func__); + + /* delete list */ + list_del(&c_node->list); + + /* destroy mutex */ + mutex_destroy(&c_node->cmd_lock); + mutex_destroy(&c_node->mem_lock); + mutex_destroy(&c_node->event_lock); + + /* free command node */ + kfree(c_node->start_work); + kfree(c_node->stop_work); + kfree(c_node->event_work); + kfree(c_node); +} + +static int ipp_check_mem_list(struct drm_exynos_ipp_cmd_node *c_node) +{ + struct drm_exynos_ipp_property *property = &c_node->property; + struct drm_exynos_ipp_mem_node *m_node; + struct list_head *head; + int ret, i, count[EXYNOS_DRM_OPS_MAX] = { 0, }; + + DRM_DEBUG_KMS("%s\n", __func__); + + mutex_lock(&c_node->mem_lock); + + for_each_ipp_ops(i) { + /* source/destination memory list */ + head = &c_node->mem_list[i]; + + if (list_empty(head)) { + DRM_DEBUG_KMS("%s:%s memory empty.\n", __func__, + i ? "dst" : "src"); + continue; + } + + /* find memory node entry */ + list_for_each_entry(m_node, head, list) { + DRM_DEBUG_KMS("%s:%s,count[%d]m_node[0x%x]\n", __func__, + i ? "dst" : "src", count[i], (int)m_node); + count[i]++; + } + } + + DRM_DEBUG_KMS("%s:min[%d]max[%d]\n", __func__, + min(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST]), + max(count[EXYNOS_DRM_OPS_SRC], count[EXYNOS_DRM_OPS_DST])); + + /* + * M2M operations should be need paired memory address. + * so, need to check minimum count about src, dst. + * other case not use paired memory, so use maximum count + */ + if (ipp_is_m2m_cmd(property->cmd)) + ret = min(count[EXYNOS_DRM_OPS_SRC], + count[EXYNOS_DRM_OPS_DST]); + else + ret = max(count[EXYNOS_DRM_OPS_SRC], + count[EXYNOS_DRM_OPS_DST]); + + mutex_unlock(&c_node->mem_lock); + + return ret; +} + +static struct drm_exynos_ipp_mem_node + *ipp_find_mem_node(struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct drm_exynos_ipp_mem_node *m_node; + struct list_head *head; + int count = 0; + + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, qbuf->buf_id); + + /* source/destination memory list */ + head = &c_node->mem_list[qbuf->ops_id]; + + /* find memory node from memory list */ + list_for_each_entry(m_node, head, list) { + DRM_DEBUG_KMS("%s:count[%d]m_node[0x%x]\n", + __func__, count++, (int)m_node); + + /* compare buffer id */ + if (m_node->buf_id == qbuf->buf_id) + return m_node; + } + + return NULL; +} + +static int ipp_set_mem_node(struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_mem_node *m_node) +{ + struct exynos_drm_ipp_ops *ops = NULL; + int ret = 0; + + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); + + if (!m_node) { + DRM_ERROR("invalid queue node.\n"); + return -EFAULT; + } + + mutex_lock(&c_node->mem_lock); + + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); + + /* get operations callback */ + ops = ippdrv->ops[m_node->ops_id]; + if (!ops) { + DRM_ERROR("not support ops.\n"); + ret = -EFAULT; + goto err_unlock; + } + + /* set address and enable irq */ + if (ops->set_addr) { + ret = ops->set_addr(ippdrv->dev, &m_node->buf_info, + m_node->buf_id, IPP_BUF_ENQUEUE); + if (ret) { + DRM_ERROR("failed to set addr.\n"); + goto err_unlock; + } + } + +err_unlock: + mutex_unlock(&c_node->mem_lock); + return ret; +} + +static struct drm_exynos_ipp_mem_node + *ipp_get_mem_node(struct drm_device *drm_dev, + struct drm_file *file, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct drm_exynos_ipp_mem_node *m_node; + struct drm_exynos_ipp_buf_info buf_info; + void *addr; + int i; + + DRM_DEBUG_KMS("%s\n", __func__); + + mutex_lock(&c_node->mem_lock); + + m_node = kzalloc(sizeof(*m_node), GFP_KERNEL); + if (!m_node) { + DRM_ERROR("failed to allocate queue node.\n"); + goto err_unlock; + } + + /* clear base address for error handling */ + memset(&buf_info, 0x0, sizeof(buf_info)); + + /* operations, buffer id */ + m_node->ops_id = qbuf->ops_id; + m_node->prop_id = qbuf->prop_id; + m_node->buf_id = qbuf->buf_id; + + DRM_DEBUG_KMS("%s:m_node[0x%x]ops_id[%d]\n", __func__, + (int)m_node, qbuf->ops_id); + DRM_DEBUG_KMS("%s:prop_id[%d]buf_id[%d]\n", __func__, + qbuf->prop_id, m_node->buf_id); + + for_each_ipp_planar(i) { + DRM_DEBUG_KMS("%s:i[%d]handle[0x%x]\n", __func__, + i, qbuf->handle[i]); + + /* get dma address by handle */ + if (qbuf->handle[i] != 0) { + addr = exynos_drm_gem_get_dma_addr(drm_dev, + qbuf->handle[i], file); + if (!addr) { + DRM_ERROR("failed to get addr.\n"); + goto err_clear; + } + + buf_info.handles[i] = qbuf->handle[i]; + buf_info.base[i] = *(dma_addr_t *) addr; + DRM_DEBUG_KMS("%s:i[%d]base[0x%x]hd[0x%x]\n", + __func__, i, buf_info.base[i], + (int)buf_info.handles[i]); + } + } + + m_node->filp = file; + m_node->buf_info = buf_info; + list_add_tail(&m_node->list, &c_node->mem_list[qbuf->ops_id]); + + mutex_unlock(&c_node->mem_lock); + return m_node; + +err_clear: + kfree(m_node); +err_unlock: + mutex_unlock(&c_node->mem_lock); + return ERR_PTR(-EFAULT); +} + +static int ipp_put_mem_node(struct drm_device *drm_dev, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_mem_node *m_node) +{ + int i; + + DRM_DEBUG_KMS("%s:node[0x%x]\n", __func__, (int)m_node); + + if (!m_node) { + DRM_ERROR("invalid dequeue node.\n"); + return -EFAULT; + } + + if (list_empty(&m_node->list)) { + DRM_ERROR("empty memory node.\n"); + return -ENOMEM; + } + + mutex_lock(&c_node->mem_lock); + + DRM_DEBUG_KMS("%s:ops_id[%d]\n", __func__, m_node->ops_id); + + /* put gem buffer */ + for_each_ipp_planar(i) { + unsigned long handle = m_node->buf_info.handles[i]; + if (handle) + exynos_drm_gem_put_dma_addr(drm_dev, handle, + m_node->filp); + } + + /* delete list in queue */ + list_del(&m_node->list); + kfree(m_node); + + mutex_unlock(&c_node->mem_lock); + + return 0; +} + +static void ipp_free_event(struct drm_pending_event *event) +{ + kfree(event); +} + +static int ipp_get_event(struct drm_device *drm_dev, + struct drm_file *file, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct drm_exynos_ipp_send_event *e; + unsigned long flags; + + DRM_DEBUG_KMS("%s:ops_id[%d]buf_id[%d]\n", __func__, + qbuf->ops_id, qbuf->buf_id); + + e = kzalloc(sizeof(*e), GFP_KERNEL); + + if (!e) { + DRM_ERROR("failed to allocate event.\n"); + spin_lock_irqsave(&drm_dev->event_lock, flags); + file->event_space += sizeof(e->event); + spin_unlock_irqrestore(&drm_dev->event_lock, flags); + return -ENOMEM; + } + + /* make event */ + e->event.base.type = DRM_EXYNOS_IPP_EVENT; + e->event.base.length = sizeof(e->event); + e->event.user_data = qbuf->user_data; + e->event.prop_id = qbuf->prop_id; + e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; + e->base.event = &e->event.base; + e->base.file_priv = file; + e->base.destroy = ipp_free_event; + list_add_tail(&e->base.link, &c_node->event_list); + + return 0; +} + +static void ipp_put_event(struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct drm_exynos_ipp_send_event *e, *te; + int count = 0; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (list_empty(&c_node->event_list)) { + DRM_DEBUG_KMS("%s:event_list is empty.\n", __func__); + return; + } + + list_for_each_entry_safe(e, te, &c_node->event_list, base.link) { + DRM_DEBUG_KMS("%s:count[%d]e[0x%x]\n", + __func__, count++, (int)e); + + /* + * quf == NULL condition means all event deletion. + * stop operations want to delete all event list. + * another case delete only same buf id. + */ + if (!qbuf) { + /* delete list */ + list_del(&e->base.link); + kfree(e); + } + + /* compare buffer id */ + if (qbuf && (qbuf->buf_id == + e->event.buf_id[EXYNOS_DRM_OPS_DST])) { + /* delete list */ + list_del(&e->base.link); + kfree(e); + return; + } + } +} + +void ipp_handle_cmd_work(struct device *dev, + struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_cmd_work *cmd_work, + struct drm_exynos_ipp_cmd_node *c_node) +{ + struct ipp_context *ctx = get_ipp_context(dev); + + cmd_work->ippdrv = ippdrv; + cmd_work->c_node = c_node; + queue_work(ctx->cmd_workq, (struct work_struct *)cmd_work); +} + +static int ipp_queue_buf_with_run(struct device *dev, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_mem_node *m_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_property *property; + struct exynos_drm_ipp_ops *ops; + int ret; + + DRM_DEBUG_KMS("%s\n", __func__); + + ippdrv = ipp_find_drv_by_handle(qbuf->prop_id); + if (IS_ERR_OR_NULL(ippdrv)) { + DRM_ERROR("failed to get ipp driver.\n"); + return -EFAULT; + } + + ops = ippdrv->ops[qbuf->ops_id]; + if (!ops) { + DRM_ERROR("failed to get ops.\n"); + return -EFAULT; + } + + property = &c_node->property; + if (!property) { + DRM_ERROR("invalid property parameter.\n"); + return -EINVAL; + } + + if (c_node->state != IPP_STATE_START) { + DRM_DEBUG_KMS("%s:bypass for invalid state.\n" , __func__); + return 0; + } + + if (!ipp_check_mem_list(c_node)) { + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); + return 0; + } + + /* + * If set destination buffer and enabled clock, + * then m2m operations need start operations at queue_buf + */ + if (ipp_is_m2m_cmd(property->cmd)) { + struct drm_exynos_ipp_cmd_work *cmd_work = c_node->start_work; + + cmd_work->ctrl = IPP_CTRL_PLAY; + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); + } else { + ret = ipp_set_mem_node(ippdrv, c_node, m_node); + if (ret) { + DRM_ERROR("failed to set m node.\n"); + return ret; + } + } + + return 0; +} + +static void ipp_clean_queue_buf(struct drm_device *drm_dev, + struct drm_exynos_ipp_cmd_node *c_node, + struct drm_exynos_ipp_queue_buf *qbuf) +{ + struct drm_exynos_ipp_mem_node *m_node, *tm_node; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!list_empty(&c_node->mem_list[qbuf->ops_id])) { + /* delete list */ + list_for_each_entry_safe(m_node, tm_node, + &c_node->mem_list[qbuf->ops_id], list) { + if (m_node->buf_id == qbuf->buf_id && + m_node->ops_id == qbuf->ops_id) + ipp_put_mem_node(drm_dev, c_node, m_node); + } + } +} + +int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void *data, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; + struct device *dev = priv->dev; + struct ipp_context *ctx = get_ipp_context(dev); + struct drm_exynos_ipp_queue_buf *qbuf = data; + struct drm_exynos_ipp_cmd_node *c_node; + struct drm_exynos_ipp_mem_node *m_node; + int ret; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!qbuf) { + DRM_ERROR("invalid buf parameter.\n"); + return -EINVAL; + } + + if (qbuf->ops_id >= EXYNOS_DRM_OPS_MAX) { + DRM_ERROR("invalid ops parameter.\n"); + return -EINVAL; + } + + DRM_DEBUG_KMS("%s:prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", + __func__, qbuf->prop_id, qbuf->ops_id ? "dst" : "src", + qbuf->buf_id, qbuf->buf_type); + + /* find command node */ + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, + qbuf->prop_id); + if (!c_node) { + DRM_ERROR("failed to get command node.\n"); + return -EFAULT; + } + + /* buffer control */ + switch (qbuf->buf_type) { + case IPP_BUF_ENQUEUE: + /* get memory node */ + m_node = ipp_get_mem_node(drm_dev, file, c_node, qbuf); + if (IS_ERR_OR_NULL(m_node)) { + DRM_ERROR("failed to get m_node.\n"); + return -EINVAL; + } + + /* + * first step get event for destination buffer. + * and second step when M2M case run with destination buffer + * if needed. + */ + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) { + /* get event for destination buffer */ + ret = ipp_get_event(drm_dev, file, c_node, qbuf); + if (ret) { + DRM_ERROR("failed to get event.\n"); + goto err_clean_node; + } + + /* + * M2M case run play control for streaming feature. + * other case set address and waiting. + */ + ret = ipp_queue_buf_with_run(dev, c_node, m_node, qbuf); + if (ret) { + DRM_ERROR("failed to run command.\n"); + goto err_clean_node; + } + } + break; + case IPP_BUF_DEQUEUE: + mutex_lock(&c_node->cmd_lock); + + /* put event for destination buffer */ + if (qbuf->ops_id == EXYNOS_DRM_OPS_DST) + ipp_put_event(c_node, qbuf); + + ipp_clean_queue_buf(drm_dev, c_node, qbuf); + + mutex_unlock(&c_node->cmd_lock); + break; + default: + DRM_ERROR("invalid buffer control.\n"); + return -EINVAL; + } + + return 0; + +err_clean_node: + DRM_ERROR("clean memory nodes.\n"); + + ipp_clean_queue_buf(drm_dev, c_node, qbuf); + return ret; +} +EXPORT_SYMBOL_GPL(exynos_drm_ipp_queue_buf); + +static bool exynos_drm_ipp_check_valid(struct device *dev, + enum drm_exynos_ipp_ctrl ctrl, enum drm_exynos_ipp_state state) +{ + DRM_DEBUG_KMS("%s\n", __func__); + + if (ctrl != IPP_CTRL_PLAY) { + if (pm_runtime_suspended(dev)) { + DRM_ERROR("pm:runtime_suspended.\n"); + goto err_status; + } + } + + switch (ctrl) { + case IPP_CTRL_PLAY: + if (state != IPP_STATE_IDLE) + goto err_status; + break; + case IPP_CTRL_STOP: + if (state == IPP_STATE_STOP) + goto err_status; + break; + case IPP_CTRL_PAUSE: + if (state != IPP_STATE_START) + goto err_status; + break; + case IPP_CTRL_RESUME: + if (state != IPP_STATE_STOP) + goto err_status; + break; + default: + DRM_ERROR("invalid state.\n"); + goto err_status; + break; + } + + return true; + +err_status: + DRM_ERROR("invalid status:ctrl[%d]state[%d]\n", ctrl, state); + return false; +} + +int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; + struct exynos_drm_ippdrv *ippdrv = NULL; + struct device *dev = priv->dev; + struct ipp_context *ctx = get_ipp_context(dev); + struct drm_exynos_ipp_cmd_ctrl *cmd_ctrl = data; + struct drm_exynos_ipp_cmd_work *cmd_work; + struct drm_exynos_ipp_cmd_node *c_node; + int ret = 0; + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!ctx) { + DRM_ERROR("invalid context.\n"); + return -EINVAL; + } + + if (!cmd_ctrl) { + DRM_ERROR("invalid control parameter.\n"); + return -EINVAL; + } + + DRM_DEBUG_KMS("%s:ctrl[%d]prop_id[%d]\n", __func__, + cmd_ctrl->ctrl, cmd_ctrl->prop_id); + + ippdrv = ipp_find_drv_by_handle(cmd_ctrl->prop_id); + if (IS_ERR_OR_NULL(ippdrv)) { + DRM_ERROR("failed to get ipp driver.\n"); + return -EINVAL; + } + + c_node = ipp_find_obj(&ctx->prop_idr, &ctx->prop_lock, + cmd_ctrl->prop_id); + if (!c_node) { + DRM_ERROR("invalid command node list.\n"); + return -EINVAL; + } + + if (!exynos_drm_ipp_check_valid(ippdrv->dev, cmd_ctrl->ctrl, + c_node->state)) { + DRM_ERROR("invalid state.\n"); + return -EINVAL; + } + + switch (cmd_ctrl->ctrl) { + case IPP_CTRL_PLAY: + if (pm_runtime_suspended(ippdrv->dev)) + pm_runtime_get_sync(ippdrv->dev); + c_node->state = IPP_STATE_START; + + cmd_work = c_node->start_work; + cmd_work->ctrl = cmd_ctrl->ctrl; + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); + c_node->state = IPP_STATE_START; + break; + case IPP_CTRL_STOP: + cmd_work = c_node->stop_work; + cmd_work->ctrl = cmd_ctrl->ctrl; + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); + + if (!wait_for_completion_timeout(&c_node->stop_complete, + msecs_to_jiffies(300))) { + DRM_ERROR("timeout stop:prop_id[%d]\n", + c_node->property.prop_id); + } + + c_node->state = IPP_STATE_STOP; + ippdrv->dedicated = false; + ipp_clean_cmd_node(c_node); + + if (list_empty(&ippdrv->cmd_list)) + pm_runtime_put_sync(ippdrv->dev); + break; + case IPP_CTRL_PAUSE: + cmd_work = c_node->stop_work; + cmd_work->ctrl = cmd_ctrl->ctrl; + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); + + if (!wait_for_completion_timeout(&c_node->stop_complete, + msecs_to_jiffies(200))) { + DRM_ERROR("timeout stop:prop_id[%d]\n", + c_node->property.prop_id); + } + + c_node->state = IPP_STATE_STOP; + break; + case IPP_CTRL_RESUME: + c_node->state = IPP_STATE_START; + cmd_work = c_node->start_work; + cmd_work->ctrl = cmd_ctrl->ctrl; + ipp_handle_cmd_work(dev, ippdrv, cmd_work, c_node); + break; + default: + /* ToDo: expand ctrl operation */ + DRM_ERROR("could not support this state currently.\n"); + goto err_clear; + } + + DRM_DEBUG_KMS("%s:done ctrl[%d]prop_id[%d]\n", __func__, + cmd_ctrl->ctrl, cmd_ctrl->prop_id); + + return 0; + +err_clear: + return ret; +} +EXPORT_SYMBOL_GPL(exynos_drm_ipp_cmd_ctrl); + +int exynos_drm_ippnb_register(struct notifier_block *nb) +{ + return blocking_notifier_chain_register( + &exynos_drm_ippnb_list, nb); +} +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_register); + +int exynos_drm_ippnb_unregister(struct notifier_block *nb) +{ + return blocking_notifier_chain_unregister( + &exynos_drm_ippnb_list, nb); +} +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_unregister); + +int exynos_drm_ippnb_send_event(unsigned long val, void *v) +{ + return blocking_notifier_call_chain( + &exynos_drm_ippnb_list, val, v); +} +EXPORT_SYMBOL_GPL(exynos_drm_ippnb_send_event); + +static int ipp_set_property(struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_property *property) +{ + struct exynos_drm_ipp_ops *ops = NULL; + int ret, i, swap = 0; + + if (!property) { + DRM_ERROR("invalid property parameter.\n"); + return -EINVAL; + } + + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); + + /* reset h/w block */ + if (ippdrv->reset && + ippdrv->reset(ippdrv->dev)) { + DRM_ERROR("failed to reset.\n"); + return -EINVAL; + } + + /* set source,destination operations */ + for_each_ipp_ops(i) { + /* ToDo: integrate property and config */ + struct drm_exynos_ipp_config *config = + &property->config[i]; + + ops = ippdrv->ops[i]; + if (!ops || !config) { + DRM_ERROR("not support ops and config.\n"); + return -EINVAL; + } + + /* set format */ + if (ops->set_fmt) { + ret = ops->set_fmt(ippdrv->dev, config->fmt); + if (ret) { + DRM_ERROR("not support format.\n"); + return ret; + } + } + + /* set transform for rotation, flip */ + if (ops->set_transf) { + swap = ops->set_transf(ippdrv->dev, config->degree, + config->flip); + if (swap < 0) { + DRM_ERROR("not support tranf.\n"); + return -EINVAL; + } + } + + /* set size */ + if (ops->set_size) { + ret = ops->set_size(ippdrv->dev, swap, &config->pos, + &config->sz); + if (ret) { + DRM_ERROR("not support size.\n"); + return ret; + } + } + } + + return 0; +} + +static int ipp_start_property(struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_cmd_node *c_node) +{ + struct drm_exynos_ipp_mem_node *m_node; + struct drm_exynos_ipp_property *property = &c_node->property; + struct list_head *head; + int ret, i; + + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); + + /* store command info in ippdrv */ + ippdrv->cmd = c_node; + + if (!ipp_check_mem_list(c_node)) { + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); + return -ENOMEM; + } + + /* set current property in ippdrv */ + ret = ipp_set_property(ippdrv, property); + if (ret) { + DRM_ERROR("failed to set property.\n"); + ippdrv->cmd = NULL; + return ret; + } + + /* check command */ + switch (property->cmd) { + case IPP_CMD_M2M: + for_each_ipp_ops(i) { + /* source/destination memory list */ + head = &c_node->mem_list[i]; + + m_node = list_first_entry(head, + struct drm_exynos_ipp_mem_node, list); + if (!m_node) { + DRM_ERROR("failed to get node.\n"); + ret = -EFAULT; + return ret; + } + + DRM_DEBUG_KMS("%s:m_node[0x%x]\n", + __func__, (int)m_node); + + ret = ipp_set_mem_node(ippdrv, c_node, m_node); + if (ret) { + DRM_ERROR("failed to set m node.\n"); + return ret; + } + } + break; + case IPP_CMD_WB: + /* destination memory list */ + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; + + list_for_each_entry(m_node, head, list) { + ret = ipp_set_mem_node(ippdrv, c_node, m_node); + if (ret) { + DRM_ERROR("failed to set m node.\n"); + return ret; + } + } + break; + case IPP_CMD_OUTPUT: + /* source memory list */ + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; + + list_for_each_entry(m_node, head, list) { + ret = ipp_set_mem_node(ippdrv, c_node, m_node); + if (ret) { + DRM_ERROR("failed to set m node.\n"); + return ret; + } + } + break; + default: + DRM_ERROR("invalid operations.\n"); + return -EINVAL; + } + + DRM_DEBUG_KMS("%s:cmd[%d]\n", __func__, property->cmd); + + /* start operations */ + if (ippdrv->start) { + ret = ippdrv->start(ippdrv->dev, property->cmd); + if (ret) { + DRM_ERROR("failed to start ops.\n"); + return ret; + } + } + + return 0; +} + +static int ipp_stop_property(struct drm_device *drm_dev, + struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_cmd_node *c_node) +{ + struct drm_exynos_ipp_mem_node *m_node, *tm_node; + struct drm_exynos_ipp_property *property = &c_node->property; + struct list_head *head; + int ret = 0, i; + + DRM_DEBUG_KMS("%s:prop_id[%d]\n", __func__, property->prop_id); + + /* put event */ + ipp_put_event(c_node, NULL); + + /* check command */ + switch (property->cmd) { + case IPP_CMD_M2M: + for_each_ipp_ops(i) { + /* source/destination memory list */ + head = &c_node->mem_list[i]; + + if (list_empty(head)) { + DRM_DEBUG_KMS("%s:mem_list is empty.\n", + __func__); + break; + } + + list_for_each_entry_safe(m_node, tm_node, + head, list) { + ret = ipp_put_mem_node(drm_dev, c_node, + m_node); + if (ret) { + DRM_ERROR("failed to put m_node.\n"); + goto err_clear; + } + } + } + break; + case IPP_CMD_WB: + /* destination memory list */ + head = &c_node->mem_list[EXYNOS_DRM_OPS_DST]; + + if (list_empty(head)) { + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); + break; + } + + list_for_each_entry_safe(m_node, tm_node, head, list) { + ret = ipp_put_mem_node(drm_dev, c_node, m_node); + if (ret) { + DRM_ERROR("failed to put m_node.\n"); + goto err_clear; + } + } + break; + case IPP_CMD_OUTPUT: + /* source memory list */ + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; + + if (list_empty(head)) { + DRM_DEBUG_KMS("%s:mem_list is empty.\n", __func__); + break; + } + + list_for_each_entry_safe(m_node, tm_node, head, list) { + ret = ipp_put_mem_node(drm_dev, c_node, m_node); + if (ret) { + DRM_ERROR("failed to put m_node.\n"); + goto err_clear; + } + } + break; + default: + DRM_ERROR("invalid operations.\n"); + ret = -EINVAL; + goto err_clear; + } + +err_clear: + /* stop operations */ + if (ippdrv->stop) + ippdrv->stop(ippdrv->dev, property->cmd); + + return ret; +} + +void ipp_sched_cmd(struct work_struct *work) +{ + struct drm_exynos_ipp_cmd_work *cmd_work = + (struct drm_exynos_ipp_cmd_work *)work; + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + struct drm_exynos_ipp_property *property; + int ret; + + DRM_DEBUG_KMS("%s\n", __func__); + + ippdrv = cmd_work->ippdrv; + if (!ippdrv) { + DRM_ERROR("invalid ippdrv list.\n"); + return; + } + + c_node = cmd_work->c_node; + if (!c_node) { + DRM_ERROR("invalid command node list.\n"); + return; + } + + mutex_lock(&c_node->cmd_lock); + + property = &c_node->property; + if (!property) { + DRM_ERROR("failed to get property:prop_id[%d]\n", + c_node->property.prop_id); + goto err_unlock; + } + + switch (cmd_work->ctrl) { + case IPP_CTRL_PLAY: + case IPP_CTRL_RESUME: + ret = ipp_start_property(ippdrv, c_node); + if (ret) { + DRM_ERROR("failed to start property:prop_id[%d]\n", + c_node->property.prop_id); + goto err_unlock; + } + + /* + * M2M case supports wait_completion of transfer. + * because M2M case supports single unit operation + * with multiple queue. + * M2M need to wait completion of data transfer. + */ + if (ipp_is_m2m_cmd(property->cmd)) { + if (!wait_for_completion_timeout + (&c_node->start_complete, msecs_to_jiffies(200))) { + DRM_ERROR("timeout event:prop_id[%d]\n", + c_node->property.prop_id); + goto err_unlock; + } + } + break; + case IPP_CTRL_STOP: + case IPP_CTRL_PAUSE: + ret = ipp_stop_property(ippdrv->drm_dev, ippdrv, + c_node); + if (ret) { + DRM_ERROR("failed to stop property.\n"); + goto err_unlock; + } + + complete(&c_node->stop_complete); + break; + default: + DRM_ERROR("unknown control type\n"); + break; + } + + DRM_DEBUG_KMS("%s:ctrl[%d] done.\n", __func__, cmd_work->ctrl); + +err_unlock: + mutex_unlock(&c_node->cmd_lock); +} + +static int ipp_send_event(struct exynos_drm_ippdrv *ippdrv, + struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) +{ + struct drm_device *drm_dev = ippdrv->drm_dev; + struct drm_exynos_ipp_property *property = &c_node->property; + struct drm_exynos_ipp_mem_node *m_node; + struct drm_exynos_ipp_queue_buf qbuf; + struct drm_exynos_ipp_send_event *e; + struct list_head *head; + struct timeval now; + unsigned long flags; + u32 tbuf_id[EXYNOS_DRM_OPS_MAX] = {0, }; + int ret, i; + + for_each_ipp_ops(i) + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, + i ? "dst" : "src", buf_id[i]); + + if (!drm_dev) { + DRM_ERROR("failed to get drm_dev.\n"); + return -EINVAL; + } + + if (!property) { + DRM_ERROR("failed to get property.\n"); + return -EINVAL; + } + + if (list_empty(&c_node->event_list)) { + DRM_DEBUG_KMS("%s:event list is empty.\n", __func__); + return 0; + } + + if (!ipp_check_mem_list(c_node)) { + DRM_DEBUG_KMS("%s:empty memory.\n", __func__); + return 0; + } + + /* check command */ + switch (property->cmd) { + case IPP_CMD_M2M: + for_each_ipp_ops(i) { + /* source/destination memory list */ + head = &c_node->mem_list[i]; + + m_node = list_first_entry(head, + struct drm_exynos_ipp_mem_node, list); + if (!m_node) { + DRM_ERROR("empty memory node.\n"); + return -ENOMEM; + } + + tbuf_id[i] = m_node->buf_id; + DRM_DEBUG_KMS("%s:%s buf_id[%d]\n", __func__, + i ? "dst" : "src", tbuf_id[i]); + + ret = ipp_put_mem_node(drm_dev, c_node, m_node); + if (ret) + DRM_ERROR("failed to put m_node.\n"); + } + break; + case IPP_CMD_WB: + /* clear buf for finding */ + memset(&qbuf, 0x0, sizeof(qbuf)); + qbuf.ops_id = EXYNOS_DRM_OPS_DST; + qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST]; + + /* get memory node entry */ + m_node = ipp_find_mem_node(c_node, &qbuf); + if (!m_node) { + DRM_ERROR("empty memory node.\n"); + return -ENOMEM; + } + + tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id; + + ret = ipp_put_mem_node(drm_dev, c_node, m_node); + if (ret) + DRM_ERROR("failed to put m_node.\n"); + break; + case IPP_CMD_OUTPUT: + /* source memory list */ + head = &c_node->mem_list[EXYNOS_DRM_OPS_SRC]; + + m_node = list_first_entry(head, + struct drm_exynos_ipp_mem_node, list); + if (!m_node) { + DRM_ERROR("empty memory node.\n"); + return -ENOMEM; + } + + tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id; + + ret = ipp_put_mem_node(drm_dev, c_node, m_node); + if (ret) + DRM_ERROR("failed to put m_node.\n"); + break; + default: + DRM_ERROR("invalid operations.\n"); + return -EINVAL; + } + + /* ToDo: Fix buffer id */ + if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST]) + DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n", + tbuf_id[1], buf_id[1], property->prop_id); + + /* + * command node have event list of destination buffer + * If destination buffer enqueue to mem list, + * then we make event and link to event list tail. + * so, we get first event for first enqueued buffer. + */ + e = list_first_entry(&c_node->event_list, + struct drm_exynos_ipp_send_event, base.link); + + if (!e) { + DRM_ERROR("empty event.\n"); + return -EINVAL; + } + + do_gettimeofday(&now); + DRM_DEBUG_KMS("%s:tv_sec[%ld]tv_usec[%ld]\n" + , __func__, now.tv_sec, now.tv_usec); + e->event.tv_sec = now.tv_sec; + e->event.tv_usec = now.tv_usec; + e->event.prop_id = property->prop_id; + + /* set buffer id about source destination */ + for_each_ipp_ops(i) + e->event.buf_id[i] = tbuf_id[i]; + + /* ToDo: compare index. If needed */ + + spin_lock_irqsave(&drm_dev->event_lock, flags); + list_move_tail(&e->base.link, &e->base.file_priv->event_list); + wake_up_interruptible(&e->base.file_priv->event_wait); + spin_unlock_irqrestore(&drm_dev->event_lock, flags); + + DRM_DEBUG_KMS("%s:done cmd[%d]prop_id[%d]buf_id[%d]\n", __func__, + property->cmd, property->prop_id, tbuf_id[EXYNOS_DRM_OPS_DST]); + + return 0; +} + +void ipp_sched_event(struct work_struct *work) +{ + struct drm_exynos_ipp_event_work *event_work = + (struct drm_exynos_ipp_event_work *)work; + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + int ret; + + if (!event_work) { + DRM_ERROR("failed to get event_work.\n"); + return; + } + + DRM_DEBUG_KMS("%s:buf_id[%d]\n", __func__, + event_work->buf_id[EXYNOS_DRM_OPS_DST]); + + ippdrv = event_work->ippdrv; + if (!ippdrv) { + DRM_ERROR("failed to get ipp driver.\n"); + return; + } + + c_node = ippdrv->cmd; + if (!c_node) { + DRM_ERROR("failed to get command node.\n"); + return; + } + + /* + * IPP supports command thread, event thread synchronization. + * If IPP close immediately from user land, then IPP make + * synchronization with command thread, so make complete event. + * or going out operations. + */ + if (c_node->state != IPP_STATE_START) { + DRM_DEBUG_KMS("%s:bypass state[%d]prop_id[%d]\n", + __func__, c_node->state, c_node->property.prop_id); + goto err_completion; + } + + mutex_lock(&c_node->event_lock); + + ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); + if (ret) { + DRM_ERROR("failed to send event.\n"); + goto err_completion; + } + +err_completion: + if (ipp_is_m2m_cmd(c_node->property.cmd)) + complete(&c_node->start_complete); + + mutex_unlock(&c_node->event_lock); +} + +static int ipp_subdrv_probe(struct drm_device *drm_dev, struct device *dev) +{ + struct ipp_context *ctx = get_ipp_context(dev); + struct exynos_drm_ippdrv *ippdrv; + int ret, count = 0; + + DRM_DEBUG_KMS("%s\n", __func__); + + /* get ipp driver entry */ + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { + ippdrv->drm_dev = drm_dev; + + ret = ipp_create_id(&ctx->ipp_idr, &ctx->ipp_lock, ippdrv, + &ippdrv->ipp_id); + if (ret) { + DRM_ERROR("failed to create id.\n"); + goto err_idr; + } + + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]ipp_id[%d]\n", __func__, + count++, (int)ippdrv, ippdrv->ipp_id); + + if (ippdrv->ipp_id == 0) { + DRM_ERROR("failed to get ipp_id[%d]\n", + ippdrv->ipp_id); + goto err_idr; + } + + /* store parent device for node */ + ippdrv->parent_dev = dev; + + /* store event work queue and handler */ + ippdrv->event_workq = ctx->event_workq; + ippdrv->sched_event = ipp_sched_event; + INIT_LIST_HEAD(&ippdrv->cmd_list); + } + + return 0; + +err_idr: + idr_remove_all(&ctx->ipp_idr); + idr_remove_all(&ctx->prop_idr); + idr_destroy(&ctx->ipp_idr); + idr_destroy(&ctx->prop_idr); + return ret; +} + +static void ipp_subdrv_remove(struct drm_device *drm_dev, struct device *dev) +{ + struct exynos_drm_ippdrv *ippdrv; + + DRM_DEBUG_KMS("%s\n", __func__); + + /* get ipp driver entry */ + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { + ippdrv->drm_dev = NULL; + exynos_drm_ippdrv_unregister(ippdrv); + } + + /* ToDo: free notifier callback list if needed */ +} + +static int ipp_subdrv_open(struct drm_device *drm_dev, struct device *dev, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv; + + DRM_DEBUG_KMS("%s\n", __func__); + + /* ToDo: multi device open */ + + priv = kzalloc(sizeof(*priv), GFP_KERNEL); + if (!priv) { + DRM_ERROR("failed to allocate priv.\n"); + return -ENOMEM; + } + priv->dev = dev; + file_priv->ipp_priv = priv; + + INIT_LIST_HEAD(&priv->event_list); + + DRM_DEBUG_KMS("%s:done priv[0x%x]\n", __func__, (int)priv); + + return 0; +} + +static void ipp_subdrv_close(struct drm_device *drm_dev, struct device *dev, + struct drm_file *file) +{ + struct drm_exynos_file_private *file_priv = file->driver_priv; + struct exynos_drm_ipp_private *priv = file_priv->ipp_priv; + struct exynos_drm_ippdrv *ippdrv = NULL; + struct drm_exynos_ipp_cmd_node *c_node, *tc_node; + int count = 0; + + DRM_DEBUG_KMS("%s:for priv[0x%x]\n", __func__, (int)priv); + + if (list_empty(&exynos_drm_ippdrv_list)) { + DRM_DEBUG_KMS("%s:ippdrv_list is empty.\n", __func__); + goto err_clear; + } + + list_for_each_entry(ippdrv, &exynos_drm_ippdrv_list, drv_list) { + if (list_empty(&ippdrv->cmd_list)) + continue; + + list_for_each_entry_safe(c_node, tc_node, + &ippdrv->cmd_list, list) { + DRM_DEBUG_KMS("%s:count[%d]ippdrv[0x%x]\n", + __func__, count++, (int)ippdrv); + + if (c_node->priv == priv) { + /* + * userland goto unnormal state. process killed. + * and close the file. + * so, IPP didn't called stop cmd ctrl. + * so, we are make stop operation in this state. + */ + if (c_node->state == IPP_STATE_START) { + ipp_stop_property(drm_dev, ippdrv, + c_node); + c_node->state = IPP_STATE_STOP; + } + + ippdrv->dedicated = false; + ipp_clean_cmd_node(c_node); + if (list_empty(&ippdrv->cmd_list)) + pm_runtime_put_sync(ippdrv->dev); + } + } + } + +err_clear: + kfree(priv); + return; +} + +static int __devinit ipp_probe(struct platform_device *pdev) +{ + struct device *dev = &pdev->dev; + struct ipp_context *ctx; + struct exynos_drm_subdrv *subdrv; + int ret; + + ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); + if (!ctx) + return -ENOMEM; + + DRM_DEBUG_KMS("%s\n", __func__); + + mutex_init(&ctx->ipp_lock); + mutex_init(&ctx->prop_lock); + + idr_init(&ctx->ipp_idr); + idr_init(&ctx->prop_idr); + + /* + * create single thread for ipp event + * IPP supports event thread for IPP drivers. + * IPP driver send event_work to this thread. + * and IPP event thread send event to user process. + */ + ctx->event_workq = create_singlethread_workqueue("ipp_event"); + if (!ctx->event_workq) { + dev_err(dev, "failed to create event workqueue\n"); + ret = -EINVAL; + goto err_clear; + } + + /* + * create single thread for ipp command + * IPP supports command thread for user process. + * user process make command node using set property ioctl. + * and make start_work and send this work to command thread. + * and then this command thread start property. + */ + ctx->cmd_workq = create_singlethread_workqueue("ipp_cmd"); + if (!ctx->cmd_workq) { + dev_err(dev, "failed to create cmd workqueue\n"); + ret = -EINVAL; + goto err_event_workq; + } + + /* set sub driver informations */ + subdrv = &ctx->subdrv; + subdrv->dev = dev; + subdrv->probe = ipp_subdrv_probe; + subdrv->remove = ipp_subdrv_remove; + subdrv->open = ipp_subdrv_open; + subdrv->close = ipp_subdrv_close; + + platform_set_drvdata(pdev, ctx); + + ret = exynos_drm_subdrv_register(subdrv); + if (ret < 0) { + DRM_ERROR("failed to register drm ipp device.\n"); + goto err_cmd_workq; + } + + dev_info(&pdev->dev, "drm ipp registered successfully.\n"); + + return 0; + +err_cmd_workq: + destroy_workqueue(ctx->cmd_workq); +err_event_workq: + destroy_workqueue(ctx->event_workq); +err_clear: + kfree(ctx); + return ret; +} + +static int __devexit ipp_remove(struct platform_device *pdev) +{ + struct ipp_context *ctx = platform_get_drvdata(pdev); + + DRM_DEBUG_KMS("%s\n", __func__); + + /* unregister sub driver */ + exynos_drm_subdrv_unregister(&ctx->subdrv); + + /* remove,destroy ipp idr */ + idr_remove_all(&ctx->ipp_idr); + idr_remove_all(&ctx->prop_idr); + idr_destroy(&ctx->ipp_idr); + idr_destroy(&ctx->prop_idr); + + mutex_destroy(&ctx->ipp_lock); + mutex_destroy(&ctx->prop_lock); + + /* destroy command, event work queue */ + destroy_workqueue(ctx->cmd_workq); + destroy_workqueue(ctx->event_workq); + + kfree(ctx); + + return 0; +} + +static int ipp_power_ctrl(struct ipp_context *ctx, bool enable) +{ + /* ToDo: Need to implement power and sysmmu ctrl. */ + DRM_DEBUG_KMS("%s:enable[%d]\n", __func__, enable); + + return 0; +} + +#ifdef CONFIG_PM_SLEEP +static int ipp_suspend(struct device *dev) +{ + struct ipp_context *ctx = get_ipp_context(dev); + + DRM_DEBUG_KMS("%s\n", __func__); + + if (pm_runtime_suspended(dev)) + return 0; + + return ipp_power_ctrl(ctx, false); +} + +static int ipp_resume(struct device *dev) +{ + struct ipp_context *ctx = get_ipp_context(dev); + + DRM_DEBUG_KMS("%s\n", __func__); + + if (!pm_runtime_suspended(dev)) + return ipp_power_ctrl(ctx, true); + + return 0; +} +#endif + +#ifdef CONFIG_PM_RUNTIME +static int ipp_runtime_suspend(struct device *dev) +{ + struct ipp_context *ctx = get_ipp_context(dev); + + DRM_DEBUG_KMS("%s\n", __func__); + + return ipp_power_ctrl(ctx, false); +} + +static int ipp_runtime_resume(struct device *dev) +{ + struct ipp_context *ctx = get_ipp_context(dev); + + DRM_DEBUG_KMS("%s\n", __func__); + + return ipp_power_ctrl(ctx, true); +} +#endif + +static const struct dev_pm_ops ipp_pm_ops = { + SET_SYSTEM_SLEEP_PM_OPS(ipp_suspend, ipp_resume) + SET_RUNTIME_PM_OPS(ipp_runtime_suspend, ipp_runtime_resume, NULL) +}; + +struct platform_driver ipp_driver = { + .probe = ipp_probe, + .remove = __devexit_p(ipp_remove), + .driver = { + .name = "exynos-drm-ipp", + .owner = THIS_MODULE, + .pm = &ipp_pm_ops, + }, +}; + diff --git a/drivers/gpu/drm/exynos/exynos_drm_ipp.h b/drivers/gpu/drm/exynos/exynos_drm_ipp.h new file mode 100644 index 0000000..baab1f0 --- /dev/null +++ b/drivers/gpu/drm/exynos/exynos_drm_ipp.h @@ -0,0 +1,266 @@ +/* + * Copyright (c) 2012 Samsung Electronics Co., Ltd. + * + * Authors: + * Eunchul Kim <chulspro.kim@samsung.com> + * Jinyoung Jeon <jy0.jeon@samsung.com> + * Sangmin Lee <lsmin.lee@samsung.com> + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR + * OTHER DEALINGS IN THE SOFTWARE. + */ + +#ifndef _EXYNOS_DRM_IPP_H_ +#define _EXYNOS_DRM_IPP_H_ + +#define for_each_ipp_ops(pos) \ + for (pos = 0; pos < EXYNOS_DRM_OPS_MAX; pos++) +#define for_each_ipp_planar(pos) \ + for (pos = 0; pos < EXYNOS_DRM_PLANAR_MAX; pos++) + +#define IPP_GET_LCD_WIDTH _IOR('F', 302, int) +#define IPP_GET_LCD_HEIGHT _IOR('F', 303, int) +#define IPP_SET_WRITEBACK _IOW('F', 304, u32) + +/* definition of state */ +enum drm_exynos_ipp_state { + IPP_STATE_IDLE, + IPP_STATE_START, + IPP_STATE_STOP, +}; + +/* + * A structure of command work information. + * @work: work structure. + * @ippdrv: current work ippdrv. + * @c_node: command node information. + * @ctrl: command control. + */ +struct drm_exynos_ipp_cmd_work { + struct work_struct work; + struct exynos_drm_ippdrv *ippdrv; + struct drm_exynos_ipp_cmd_node *c_node; + enum drm_exynos_ipp_ctrl ctrl; +}; + +/* + * A structure of command node. + * + * @priv: IPP private infomation. + * @list: list head to command queue information. + * @event_list: list head of event. + * @mem_list: list head to source,destination memory queue information. + * @cmd_lock: lock for synchronization of access to ioctl. + * @mem_lock: lock for synchronization of access to memory nodes. + * @event_lock: lock for synchronization of access to scheduled event. + * @start_complete: completion of start of command. + * @stop_complete: completion of stop of command. + * @property: property information. + * @start_work: start command work structure. + * @stop_work: stop command work structure. + * @event_work: event work structure. + * @state: state of command node. + */ +struct drm_exynos_ipp_cmd_node { + struct exynos_drm_ipp_private *priv; + struct list_head list; + struct list_head event_list; + struct list_head mem_list[EXYNOS_DRM_OPS_MAX]; + struct mutex cmd_lock; + struct mutex mem_lock; + struct mutex event_lock; + struct completion start_complete; + struct completion stop_complete; + struct drm_exynos_ipp_property property; + struct drm_exynos_ipp_cmd_work *start_work; + struct drm_exynos_ipp_cmd_work *stop_work; + struct drm_exynos_ipp_event_work *event_work; + enum drm_exynos_ipp_state state; +}; + +/* + * A structure of buffer information. + * + * @gem_objs: Y, Cb, Cr each gem object. + * @base: Y, Cb, Cr each planar address. + */ +struct drm_exynos_ipp_buf_info { + unsigned long handles[EXYNOS_DRM_PLANAR_MAX]; + dma_addr_t base[EXYNOS_DRM_PLANAR_MAX]; +}; + +/* + * A structure of wb setting infomation. + * + * @enable: enable flag for wb. + * @refresh: HZ of the refresh rate. + */ +struct drm_exynos_ipp_set_wb { + __u32 enable; + __u32 refresh; +}; + +/* + * A structure of event work information. + * + * @work: work structure. + * @ippdrv: current work ippdrv. + * @buf_id: id of src, dst buffer. + */ +struct drm_exynos_ipp_event_work { + struct work_struct work; + struct exynos_drm_ippdrv *ippdrv; + u32 buf_id[EXYNOS_DRM_OPS_MAX]; +}; + +/* + * A structure of source,destination operations. + * + * @set_fmt: set format of image. + * @set_transf: set transform(rotations, flip). + * @set_size: set size of region. + * @set_addr: set address for dma. + */ +struct exynos_drm_ipp_ops { + int (*set_fmt)(struct device *dev, u32 fmt); + int (*set_transf)(struct device *dev, + enum drm_exynos_degree degree, + enum drm_exynos_flip flip); + int (*set_size)(struct device *dev, int swap, + struct drm_exynos_pos *pos, struct drm_exynos_sz *sz); + int (*set_addr)(struct device *dev, + struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, + enum drm_exynos_ipp_buf_type buf_type); +}; + +/* + * A structure of ipp driver. + * + * @drv_list: list head for registed sub driver information. + * @parent_dev: parent device information. + * @dev: platform device. + * @drm_dev: drm device. + * @ipp_id: id of ipp driver. + * @dedicated: dedicated ipp device. + * @ops: source, destination operations. + * @event_workq: event work queue. + * @cmd: current command information. + * @cmd_list: list head for command information. + * @prop_list: property informations of current ipp driver. + * @check_property: check property about format, size, buffer. + * @reset: reset ipp block. + * @start: ipp each device start. + * @stop: ipp each device stop. + * @sched_event: work schedule handler. + */ +struct exynos_drm_ippdrv { + struct list_head drv_list; + struct device *parent_dev; + struct device *dev; + struct drm_device *drm_dev; + u32 ipp_id; + bool dedicated; + struct exynos_drm_ipp_ops *ops[EXYNOS_DRM_OPS_MAX]; + struct workqueue_struct *event_workq; + struct drm_exynos_ipp_cmd_node *cmd; + struct list_head cmd_list; + struct drm_exynos_ipp_prop_list *prop_list; + + int (*check_property)(struct device *dev, + struct drm_exynos_ipp_property *property); + int (*reset)(struct device *dev); + int (*start)(struct device *dev, enum drm_exynos_ipp_cmd cmd); + void (*stop)(struct device *dev, enum drm_exynos_ipp_cmd cmd); + void (*sched_event)(struct work_struct *work); +}; + +#ifdef CONFIG_DRM_EXYNOS_IPP +extern int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv); +extern int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv); +extern int exynos_drm_ipp_get_property(struct drm_device *drm_dev, void *data, + struct drm_file *file); +extern int exynos_drm_ipp_set_property(struct drm_device *drm_dev, void *data, + struct drm_file *file); +extern int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, void *data, + struct drm_file *file); +extern int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, void *data, + struct drm_file *file); +extern int exynos_drm_ippnb_register(struct notifier_block *nb); +extern int exynos_drm_ippnb_unregister(struct notifier_block *nb); +extern int exynos_drm_ippnb_send_event(unsigned long val, void *v); +extern void ipp_sched_cmd(struct work_struct *work); +extern void ipp_sched_event(struct work_struct *work); + +#else +static inline int exynos_drm_ippdrv_register(struct exynos_drm_ippdrv *ippdrv) +{ + return -ENODEV; +} + +static inline int exynos_drm_ippdrv_unregister(struct exynos_drm_ippdrv *ippdrv) +{ + return -ENODEV; +} + +static inline int exynos_drm_ipp_get_property(struct drm_device *drm_dev, + void *data, + struct drm_file *file_priv) +{ + return -ENOTTY; +} + +static inline int exynos_drm_ipp_set_property(struct drm_device *drm_dev, + void *data, + struct drm_file *file_priv) +{ + return -ENOTTY; +} + +static inline int exynos_drm_ipp_queue_buf(struct drm_device *drm_dev, + void *data, + struct drm_file *file) +{ + return -ENOTTY; +} + +static inline int exynos_drm_ipp_cmd_ctrl(struct drm_device *drm_dev, + void *data, + struct drm_file *file) +{ + return -ENOTTY; +} + +static inline int exynos_drm_ippnb_register(struct notifier_block *nb) +{ + return -ENODEV; +} + +static inline int exynos_drm_ippnb_unregister(struct notifier_block *nb) +{ + return -ENODEV; +} + +static inline int exynos_drm_ippnb_send_event(unsigned long val, void *v) +{ + return -ENOTTY; +} +#endif + +#endif /* _EXYNOS_DRM_IPP_H_ */ + diff --git a/include/uapi/drm/exynos_drm.h b/include/uapi/drm/exynos_drm.h index c0494d5..82772d7 100644 --- a/include/uapi/drm/exynos_drm.h +++ b/include/uapi/drm/exynos_drm.h @@ -154,6 +154,170 @@ struct drm_exynos_g2d_exec { __u64 async; }; +enum drm_exynos_ops_id { + EXYNOS_DRM_OPS_SRC, + EXYNOS_DRM_OPS_DST, + EXYNOS_DRM_OPS_MAX, +}; + +struct drm_exynos_sz { + __u32 hsize; + __u32 vsize; +}; + +struct drm_exynos_pos { + __u32 x; + __u32 y; + __u32 w; + __u32 h; +}; + +enum drm_exynos_flip { + EXYNOS_DRM_FLIP_NONE = (0 << 0), + EXYNOS_DRM_FLIP_VERTICAL = (1 << 0), + EXYNOS_DRM_FLIP_HORIZONTAL = (1 << 1), +}; + +enum drm_exynos_degree { + EXYNOS_DRM_DEGREE_0, + EXYNOS_DRM_DEGREE_90, + EXYNOS_DRM_DEGREE_180, + EXYNOS_DRM_DEGREE_270, +}; + +enum drm_exynos_planer { + EXYNOS_DRM_PLANAR_Y, + EXYNOS_DRM_PLANAR_CB, + EXYNOS_DRM_PLANAR_CR, + EXYNOS_DRM_PLANAR_MAX, +}; + +/** + * A structure for ipp supported property list. + * + * @version: version of this structure. + * @ipp_id: id of ipp driver. + * @count: count of ipp driver. + * @writeback: flag of writeback supporting. + * @flip: flag of flip supporting. + * @degree: flag of degree information. + * @csc: flag of csc supporting. + * @crop: flag of crop supporting. + * @scale: flag of scale supporting. + * @refresh_min: min hz of refresh. + * @refresh_max: max hz of refresh. + * @crop_min: crop min resolution. + * @crop_max: crop max resolution. + * @scale_min: scale min resolution. + * @scale_max: scale max resolution. + */ +struct drm_exynos_ipp_prop_list { + __u32 version; + __u32 ipp_id; + __u32 count; + __u32 writeback; + __u32 flip; + __u32 degree; + __u32 csc; + __u32 crop; + __u32 scale; + __u32 refresh_min; + __u32 refresh_max; + __u32 reserved; + struct drm_exynos_sz crop_min; + struct drm_exynos_sz crop_max; + struct drm_exynos_sz scale_min; + struct drm_exynos_sz scale_max; +}; + +/** + * A structure for ipp config. + * + * @ops_id: property of operation directions. + * @flip: property of mirror, flip. + * @degree: property of rotation degree. + * @fmt: property of image format. + * @sz: property of image size. + * @pos: property of image position(src-cropped,dst-scaler). + */ +struct drm_exynos_ipp_config { + enum drm_exynos_ops_id ops_id; + enum drm_exynos_flip flip; + enum drm_exynos_degree degree; + __u32 fmt; + struct drm_exynos_sz sz; + struct drm_exynos_pos pos; +}; + +enum drm_exynos_ipp_cmd { + IPP_CMD_NONE, + IPP_CMD_M2M, + IPP_CMD_WB, + IPP_CMD_OUTPUT, + IPP_CMD_MAX, +}; + +/** + * A structure for ipp property. + * + * @config: source, destination config. + * @cmd: definition of command. + * @ipp_id: id of ipp driver. + * @prop_id: id of property. + * @refresh_rate: refresh rate. + */ +struct drm_exynos_ipp_property { + struct drm_exynos_ipp_config config[EXYNOS_DRM_OPS_MAX]; + enum drm_exynos_ipp_cmd cmd; + __u32 ipp_id; + __u32 prop_id; + __u32 refresh_rate; +}; + +enum drm_exynos_ipp_buf_type { + IPP_BUF_ENQUEUE, + IPP_BUF_DEQUEUE, +}; + +/** + * A structure for ipp buffer operations. + * + * @ops_id: operation directions. + * @buf_type: definition of buffer. + * @prop_id: id of property. + * @buf_id: id of buffer. + * @handle: Y, Cb, Cr each planar handle. + * @user_data: user data. + */ +struct drm_exynos_ipp_queue_buf { + enum drm_exynos_ops_id ops_id; + enum drm_exynos_ipp_buf_type buf_type; + __u32 prop_id; + __u32 buf_id; + __u32 handle[EXYNOS_DRM_PLANAR_MAX]; + __u32 reserved; + __u64 user_data; +}; + +enum drm_exynos_ipp_ctrl { + IPP_CTRL_PLAY, + IPP_CTRL_STOP, + IPP_CTRL_PAUSE, + IPP_CTRL_RESUME, + IPP_CTRL_MAX, +}; + +/** + * A structure for ipp start/stop operations. + * + * @prop_id: id of property. + * @ctrl: definition of control. + */ +struct drm_exynos_ipp_cmd_ctrl { + __u32 prop_id; + enum drm_exynos_ipp_ctrl ctrl; +}; + #define DRM_EXYNOS_GEM_CREATE 0x00 #define DRM_EXYNOS_GEM_MAP_OFFSET 0x01 #define DRM_EXYNOS_GEM_MMAP 0x02 @@ -166,6 +330,12 @@ struct drm_exynos_g2d_exec { #define DRM_EXYNOS_G2D_SET_CMDLIST 0x21 #define DRM_EXYNOS_G2D_EXEC 0x22 +/* IPP - Image Post Processing */ +#define DRM_EXYNOS_IPP_GET_PROPERTY 0x30 +#define DRM_EXYNOS_IPP_SET_PROPERTY 0x31 +#define DRM_EXYNOS_IPP_QUEUE_BUF 0x32 +#define DRM_EXYNOS_IPP_CMD_CTRL 0x33 + #define DRM_IOCTL_EXYNOS_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + \ DRM_EXYNOS_GEM_CREATE, struct drm_exynos_gem_create) @@ -188,8 +358,18 @@ struct drm_exynos_g2d_exec { #define DRM_IOCTL_EXYNOS_G2D_EXEC DRM_IOWR(DRM_COMMAND_BASE + \ DRM_EXYNOS_G2D_EXEC, struct drm_exynos_g2d_exec) +#define DRM_IOCTL_EXYNOS_IPP_GET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + \ + DRM_EXYNOS_IPP_GET_PROPERTY, struct drm_exynos_ipp_prop_list) +#define DRM_IOCTL_EXYNOS_IPP_SET_PROPERTY DRM_IOWR(DRM_COMMAND_BASE + \ + DRM_EXYNOS_IPP_SET_PROPERTY, struct drm_exynos_ipp_property) +#define DRM_IOCTL_EXYNOS_IPP_QUEUE_BUF DRM_IOWR(DRM_COMMAND_BASE + \ + DRM_EXYNOS_IPP_QUEUE_BUF, struct drm_exynos_ipp_queue_buf) +#define DRM_IOCTL_EXYNOS_IPP_CMD_CTRL DRM_IOWR(DRM_COMMAND_BASE + \ + DRM_EXYNOS_IPP_CMD_CTRL, struct drm_exynos_ipp_cmd_ctrl) + /* EXYNOS specific events */ #define DRM_EXYNOS_G2D_EVENT 0x80000000 +#define DRM_EXYNOS_IPP_EVENT 0x80000001 struct drm_exynos_g2d_event { struct drm_event base; @@ -200,4 +380,14 @@ struct drm_exynos_g2d_event { __u32 reserved; }; +struct drm_exynos_ipp_event { + struct drm_event base; + __u64 user_data; + __u32 tv_sec; + __u32 tv_usec; + __u32 prop_id; + __u32 reserved; + __u32 buf_id[EXYNOS_DRM_OPS_MAX]; +}; + #endif /* _UAPI_EXYNOS_DRM_H_ */