aab174f0df
Pull vfs update from Al Viro: - big one - consolidation of descriptor-related logics; almost all of that is moved to fs/file.c (BTW, I'm seriously tempted to rename the result to fd.c. As it is, we have a situation when file_table.c is about handling of struct file and file.c is about handling of descriptor tables; the reasons are historical - file_table.c used to be about a static array of struct file we used to have way back). A lot of stray ends got cleaned up and converted to saner primitives, disgusting mess in android/binder.c is still disgusting, but at least doesn't poke so much in descriptor table guts anymore. A bunch of relatively minor races got fixed in process, plus an ext4 struct file leak. - related thing - fget_light() partially unuglified; see fdget() in there (and yes, it generates the code as good as we used to have). - also related - bits of Cyrill's procfs stuff that got entangled into that work; _not_ all of it, just the initial move to fs/proc/fd.c and switch of fdinfo to seq_file. - Alex's fs/coredump.c spiltoff - the same story, had been easier to take that commit than mess with conflicts. The rest is a separate pile, this was just a mechanical code movement. - a few misc patches all over the place. Not all for this cycle, there'll be more (and quite a few currently sit in akpm's tree)." Fix up trivial conflicts in the android binder driver, and some fairly simple conflicts due to two different changes to the sock_alloc_file() interface ("take descriptor handling from sock_alloc_file() to callers" vs "net: Providing protocol type via system.sockprotoname xattr of /proc/PID/fd entries" adding a dentry name to the socket) * 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs: (72 commits) MAX_LFS_FILESIZE should be a loff_t compat: fs: Generic compat_sys_sendfile implementation fs: push rcu_barrier() from deactivate_locked_super() to filesystems btrfs: reada_extent doesn't need kref for refcount coredump: move core dump functionality into its own file coredump: prevent double-free on an error path in core dumper usb/gadget: fix misannotations fcntl: fix misannotations ceph: don't abuse d_delete() on failure exits hypfs: ->d_parent is never NULL or negative vfs: delete surplus inode NULL check switch simple cases of fget_light to fdget new helpers: fdget()/fdput() switch o2hb_region_dev_write() to fget_light() proc_map_files_readdir(): don't bother with grabbing files make get_file() return its argument vhost_set_vring(): turn pollstart/pollstop into bool switch prctl_set_mm_exe_file() to fget_light() switch xfs_find_handle() to fget_light() switch xfs_swapext() to fget_light() ...
522 lines
14 KiB
C
522 lines
14 KiB
C
/* drivers/video/msm_fb/mdp.c
|
|
*
|
|
* MSM MDP Interface (used by framebuffer core)
|
|
*
|
|
* Copyright (C) 2007 QUALCOMM Incorporated
|
|
* Copyright (C) 2007 Google Incorporated
|
|
*
|
|
* This software is licensed under the terms of the GNU General Public
|
|
* License version 2, as published by the Free Software Foundation, and
|
|
* may be copied, distributed, and modified under those terms.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/fb.h>
|
|
#include <linux/msm_mdp.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/file.h>
|
|
#include <linux/major.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include <mach/msm_iomap.h>
|
|
#include <linux/platform_data/video-msm_fb.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/export.h>
|
|
|
|
#include "mdp_hw.h"
|
|
|
|
struct class *mdp_class;
|
|
|
|
#define MDP_CMD_DEBUG_ACCESS_BASE (0x10000)
|
|
|
|
static uint16_t mdp_default_ccs[] = {
|
|
0x254, 0x000, 0x331, 0x254, 0xF38, 0xE61, 0x254, 0x409, 0x000,
|
|
0x010, 0x080, 0x080
|
|
};
|
|
|
|
static DECLARE_WAIT_QUEUE_HEAD(mdp_dma2_waitqueue);
|
|
static DECLARE_WAIT_QUEUE_HEAD(mdp_ppp_waitqueue);
|
|
static struct msmfb_callback *dma_callback;
|
|
static struct clk *clk;
|
|
static unsigned int mdp_irq_mask;
|
|
static DEFINE_SPINLOCK(mdp_lock);
|
|
DEFINE_MUTEX(mdp_mutex);
|
|
|
|
static int enable_mdp_irq(struct mdp_info *mdp, uint32_t mask)
|
|
{
|
|
unsigned long irq_flags;
|
|
int ret = 0;
|
|
|
|
BUG_ON(!mask);
|
|
|
|
spin_lock_irqsave(&mdp_lock, irq_flags);
|
|
/* if the mask bits are already set return an error, this interrupt
|
|
* is already enabled */
|
|
if (mdp_irq_mask & mask) {
|
|
printk(KERN_ERR "mdp irq already on already on %x %x\n",
|
|
mdp_irq_mask, mask);
|
|
ret = -1;
|
|
}
|
|
/* if the mdp irq is not already enabled enable it */
|
|
if (!mdp_irq_mask) {
|
|
if (clk)
|
|
clk_enable(clk);
|
|
enable_irq(mdp->irq);
|
|
}
|
|
|
|
/* update the irq mask to reflect the fact that the interrupt is
|
|
* enabled */
|
|
mdp_irq_mask |= mask;
|
|
spin_unlock_irqrestore(&mdp_lock, irq_flags);
|
|
return ret;
|
|
}
|
|
|
|
static int locked_disable_mdp_irq(struct mdp_info *mdp, uint32_t mask)
|
|
{
|
|
/* this interrupt is already disabled! */
|
|
if (!(mdp_irq_mask & mask)) {
|
|
printk(KERN_ERR "mdp irq already off %x %x\n",
|
|
mdp_irq_mask, mask);
|
|
return -1;
|
|
}
|
|
/* update the irq mask to reflect the fact that the interrupt is
|
|
* disabled */
|
|
mdp_irq_mask &= ~(mask);
|
|
/* if no one is waiting on the interrupt, disable it */
|
|
if (!mdp_irq_mask) {
|
|
disable_irq_nosync(mdp->irq);
|
|
if (clk)
|
|
clk_disable(clk);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int disable_mdp_irq(struct mdp_info *mdp, uint32_t mask)
|
|
{
|
|
unsigned long irq_flags;
|
|
int ret;
|
|
|
|
spin_lock_irqsave(&mdp_lock, irq_flags);
|
|
ret = locked_disable_mdp_irq(mdp, mask);
|
|
spin_unlock_irqrestore(&mdp_lock, irq_flags);
|
|
return ret;
|
|
}
|
|
|
|
static irqreturn_t mdp_isr(int irq, void *data)
|
|
{
|
|
uint32_t status;
|
|
unsigned long irq_flags;
|
|
struct mdp_info *mdp = data;
|
|
|
|
spin_lock_irqsave(&mdp_lock, irq_flags);
|
|
|
|
status = mdp_readl(mdp, MDP_INTR_STATUS);
|
|
mdp_writel(mdp, status, MDP_INTR_CLEAR);
|
|
|
|
status &= mdp_irq_mask;
|
|
if (status & DL0_DMA2_TERM_DONE) {
|
|
if (dma_callback) {
|
|
dma_callback->func(dma_callback);
|
|
dma_callback = NULL;
|
|
}
|
|
wake_up(&mdp_dma2_waitqueue);
|
|
}
|
|
|
|
if (status & DL0_ROI_DONE)
|
|
wake_up(&mdp_ppp_waitqueue);
|
|
|
|
if (status)
|
|
locked_disable_mdp_irq(mdp, status);
|
|
|
|
spin_unlock_irqrestore(&mdp_lock, irq_flags);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static uint32_t mdp_check_mask(uint32_t mask)
|
|
{
|
|
uint32_t ret;
|
|
unsigned long irq_flags;
|
|
|
|
spin_lock_irqsave(&mdp_lock, irq_flags);
|
|
ret = mdp_irq_mask & mask;
|
|
spin_unlock_irqrestore(&mdp_lock, irq_flags);
|
|
return ret;
|
|
}
|
|
|
|
static int mdp_wait(struct mdp_info *mdp, uint32_t mask, wait_queue_head_t *wq)
|
|
{
|
|
int ret = 0;
|
|
unsigned long irq_flags;
|
|
|
|
wait_event_timeout(*wq, !mdp_check_mask(mask), HZ);
|
|
|
|
spin_lock_irqsave(&mdp_lock, irq_flags);
|
|
if (mdp_irq_mask & mask) {
|
|
locked_disable_mdp_irq(mdp, mask);
|
|
printk(KERN_WARNING "timeout waiting for mdp to complete %x\n",
|
|
mask);
|
|
ret = -ETIMEDOUT;
|
|
}
|
|
spin_unlock_irqrestore(&mdp_lock, irq_flags);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void mdp_dma_wait(struct mdp_device *mdp_dev)
|
|
{
|
|
#define MDP_MAX_TIMEOUTS 20
|
|
static int timeout_count;
|
|
struct mdp_info *mdp = container_of(mdp_dev, struct mdp_info, mdp_dev);
|
|
|
|
if (mdp_wait(mdp, DL0_DMA2_TERM_DONE, &mdp_dma2_waitqueue) == -ETIMEDOUT)
|
|
timeout_count++;
|
|
else
|
|
timeout_count = 0;
|
|
|
|
if (timeout_count > MDP_MAX_TIMEOUTS) {
|
|
printk(KERN_ERR "mdp: dma failed %d times, somethings wrong!\n",
|
|
MDP_MAX_TIMEOUTS);
|
|
BUG();
|
|
}
|
|
}
|
|
|
|
static int mdp_ppp_wait(struct mdp_info *mdp)
|
|
{
|
|
return mdp_wait(mdp, DL0_ROI_DONE, &mdp_ppp_waitqueue);
|
|
}
|
|
|
|
void mdp_dma_to_mddi(struct mdp_info *mdp, uint32_t addr, uint32_t stride,
|
|
uint32_t width, uint32_t height, uint32_t x, uint32_t y,
|
|
struct msmfb_callback *callback)
|
|
{
|
|
uint32_t dma2_cfg;
|
|
uint16_t ld_param = 0; /* 0=PRIM, 1=SECD, 2=EXT */
|
|
|
|
if (enable_mdp_irq(mdp, DL0_DMA2_TERM_DONE)) {
|
|
printk(KERN_ERR "mdp_dma_to_mddi: busy\n");
|
|
return;
|
|
}
|
|
|
|
dma_callback = callback;
|
|
|
|
dma2_cfg = DMA_PACK_TIGHT |
|
|
DMA_PACK_ALIGN_LSB |
|
|
DMA_PACK_PATTERN_RGB |
|
|
DMA_OUT_SEL_AHB |
|
|
DMA_IBUF_NONCONTIGUOUS;
|
|
|
|
dma2_cfg |= DMA_IBUF_FORMAT_RGB565;
|
|
|
|
dma2_cfg |= DMA_OUT_SEL_MDDI;
|
|
|
|
dma2_cfg |= DMA_MDDI_DMAOUT_LCD_SEL_PRIMARY;
|
|
|
|
dma2_cfg |= DMA_DITHER_EN;
|
|
|
|
/* setup size, address, and stride */
|
|
mdp_writel(mdp, (height << 16) | (width),
|
|
MDP_CMD_DEBUG_ACCESS_BASE + 0x0184);
|
|
mdp_writel(mdp, addr, MDP_CMD_DEBUG_ACCESS_BASE + 0x0188);
|
|
mdp_writel(mdp, stride, MDP_CMD_DEBUG_ACCESS_BASE + 0x018C);
|
|
|
|
/* 666 18BPP */
|
|
dma2_cfg |= DMA_DSTC0G_6BITS | DMA_DSTC1B_6BITS | DMA_DSTC2R_6BITS;
|
|
|
|
/* set y & x offset and MDDI transaction parameters */
|
|
mdp_writel(mdp, (y << 16) | (x), MDP_CMD_DEBUG_ACCESS_BASE + 0x0194);
|
|
mdp_writel(mdp, ld_param, MDP_CMD_DEBUG_ACCESS_BASE + 0x01a0);
|
|
mdp_writel(mdp, (MDDI_VDO_PACKET_DESC << 16) | MDDI_VDO_PACKET_PRIM,
|
|
MDP_CMD_DEBUG_ACCESS_BASE + 0x01a4);
|
|
|
|
mdp_writel(mdp, dma2_cfg, MDP_CMD_DEBUG_ACCESS_BASE + 0x0180);
|
|
|
|
/* start DMA2 */
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0044);
|
|
}
|
|
|
|
void mdp_dma(struct mdp_device *mdp_dev, uint32_t addr, uint32_t stride,
|
|
uint32_t width, uint32_t height, uint32_t x, uint32_t y,
|
|
struct msmfb_callback *callback, int interface)
|
|
{
|
|
struct mdp_info *mdp = container_of(mdp_dev, struct mdp_info, mdp_dev);
|
|
|
|
if (interface == MSM_MDDI_PMDH_INTERFACE) {
|
|
mdp_dma_to_mddi(mdp, addr, stride, width, height, x, y,
|
|
callback);
|
|
}
|
|
}
|
|
|
|
int get_img(struct mdp_img *img, struct fb_info *info,
|
|
unsigned long *start, unsigned long *len,
|
|
struct file **filep)
|
|
{
|
|
int ret = 0;
|
|
struct fd f = fdget(img->memory_id);
|
|
if (f.file == NULL)
|
|
return -1;
|
|
|
|
if (MAJOR(f.file->f_dentry->d_inode->i_rdev) == FB_MAJOR) {
|
|
*start = info->fix.smem_start;
|
|
*len = info->fix.smem_len;
|
|
} else
|
|
ret = -1;
|
|
fdput(f);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void put_img(struct file *src_file, struct file *dst_file)
|
|
{
|
|
}
|
|
|
|
int mdp_blit(struct mdp_device *mdp_dev, struct fb_info *fb,
|
|
struct mdp_blit_req *req)
|
|
{
|
|
int ret;
|
|
unsigned long src_start = 0, src_len = 0, dst_start = 0, dst_len = 0;
|
|
struct mdp_info *mdp = container_of(mdp_dev, struct mdp_info, mdp_dev);
|
|
struct file *src_file = 0, *dst_file = 0;
|
|
|
|
/* WORKAROUND FOR HARDWARE BUG IN BG TILE FETCH */
|
|
if (unlikely(req->src_rect.h == 0 ||
|
|
req->src_rect.w == 0)) {
|
|
printk(KERN_ERR "mpd_ppp: src img of zero size!\n");
|
|
return -EINVAL;
|
|
}
|
|
if (unlikely(req->dst_rect.h == 0 ||
|
|
req->dst_rect.w == 0))
|
|
return -EINVAL;
|
|
|
|
/* do this first so that if this fails, the caller can always
|
|
* safely call put_img */
|
|
if (unlikely(get_img(&req->src, fb, &src_start, &src_len, &src_file))) {
|
|
printk(KERN_ERR "mpd_ppp: could not retrieve src image from "
|
|
"memory\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (unlikely(get_img(&req->dst, fb, &dst_start, &dst_len, &dst_file))) {
|
|
printk(KERN_ERR "mpd_ppp: could not retrieve dst image from "
|
|
"memory\n");
|
|
return -EINVAL;
|
|
}
|
|
mutex_lock(&mdp_mutex);
|
|
|
|
/* transp_masking unimplemented */
|
|
req->transp_mask = MDP_TRANSP_NOP;
|
|
if (unlikely((req->transp_mask != MDP_TRANSP_NOP ||
|
|
req->alpha != MDP_ALPHA_NOP ||
|
|
HAS_ALPHA(req->src.format)) &&
|
|
(req->flags & MDP_ROT_90 &&
|
|
req->dst_rect.w <= 16 && req->dst_rect.h >= 16))) {
|
|
int i;
|
|
unsigned int tiles = req->dst_rect.h / 16;
|
|
unsigned int remainder = req->dst_rect.h % 16;
|
|
req->src_rect.w = 16*req->src_rect.w / req->dst_rect.h;
|
|
req->dst_rect.h = 16;
|
|
for (i = 0; i < tiles; i++) {
|
|
enable_mdp_irq(mdp, DL0_ROI_DONE);
|
|
ret = mdp_ppp_blit(mdp, req, src_file, src_start,
|
|
src_len, dst_file, dst_start,
|
|
dst_len);
|
|
if (ret)
|
|
goto err_bad_blit;
|
|
ret = mdp_ppp_wait(mdp);
|
|
if (ret)
|
|
goto err_wait_failed;
|
|
req->dst_rect.y += 16;
|
|
req->src_rect.x += req->src_rect.w;
|
|
}
|
|
if (!remainder)
|
|
goto end;
|
|
req->src_rect.w = remainder*req->src_rect.w / req->dst_rect.h;
|
|
req->dst_rect.h = remainder;
|
|
}
|
|
enable_mdp_irq(mdp, DL0_ROI_DONE);
|
|
ret = mdp_ppp_blit(mdp, req, src_file, src_start, src_len, dst_file,
|
|
dst_start,
|
|
dst_len);
|
|
if (ret)
|
|
goto err_bad_blit;
|
|
ret = mdp_ppp_wait(mdp);
|
|
if (ret)
|
|
goto err_wait_failed;
|
|
end:
|
|
put_img(src_file, dst_file);
|
|
mutex_unlock(&mdp_mutex);
|
|
return 0;
|
|
err_bad_blit:
|
|
disable_mdp_irq(mdp, DL0_ROI_DONE);
|
|
err_wait_failed:
|
|
put_img(src_file, dst_file);
|
|
mutex_unlock(&mdp_mutex);
|
|
return ret;
|
|
}
|
|
|
|
void mdp_set_grp_disp(struct mdp_device *mdp_dev, unsigned disp_id)
|
|
{
|
|
struct mdp_info *mdp = container_of(mdp_dev, struct mdp_info, mdp_dev);
|
|
|
|
disp_id &= 0xf;
|
|
mdp_writel(mdp, disp_id, MDP_FULL_BYPASS_WORD43);
|
|
}
|
|
|
|
int register_mdp_client(struct class_interface *cint)
|
|
{
|
|
if (!mdp_class) {
|
|
pr_err("mdp: no mdp_class when registering mdp client\n");
|
|
return -ENODEV;
|
|
}
|
|
cint->class = mdp_class;
|
|
return class_interface_register(cint);
|
|
}
|
|
|
|
#include "mdp_csc_table.h"
|
|
#include "mdp_scale_tables.h"
|
|
|
|
int mdp_probe(struct platform_device *pdev)
|
|
{
|
|
struct resource *resource;
|
|
int ret;
|
|
int n;
|
|
struct mdp_info *mdp;
|
|
|
|
resource = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
if (!resource) {
|
|
pr_err("mdp: can not get mdp mem resource!\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
mdp = kzalloc(sizeof(struct mdp_info), GFP_KERNEL);
|
|
if (!mdp)
|
|
return -ENOMEM;
|
|
|
|
mdp->irq = platform_get_irq(pdev, 0);
|
|
if (mdp->irq < 0) {
|
|
pr_err("mdp: can not get mdp irq\n");
|
|
ret = mdp->irq;
|
|
goto error_get_irq;
|
|
}
|
|
|
|
mdp->base = ioremap(resource->start, resource_size(resource));
|
|
if (mdp->base == 0) {
|
|
printk(KERN_ERR "msmfb: cannot allocate mdp regs!\n");
|
|
ret = -ENOMEM;
|
|
goto error_ioremap;
|
|
}
|
|
|
|
mdp->mdp_dev.dma = mdp_dma;
|
|
mdp->mdp_dev.dma_wait = mdp_dma_wait;
|
|
mdp->mdp_dev.blit = mdp_blit;
|
|
mdp->mdp_dev.set_grp_disp = mdp_set_grp_disp;
|
|
|
|
clk = clk_get(&pdev->dev, "mdp_clk");
|
|
if (IS_ERR(clk)) {
|
|
printk(KERN_INFO "mdp: failed to get mdp clk");
|
|
ret = PTR_ERR(clk);
|
|
goto error_get_clk;
|
|
}
|
|
|
|
ret = request_irq(mdp->irq, mdp_isr, 0, "msm_mdp", mdp);
|
|
if (ret)
|
|
goto error_request_irq;
|
|
disable_irq(mdp->irq);
|
|
mdp_irq_mask = 0;
|
|
|
|
/* debug interface write access */
|
|
mdp_writel(mdp, 1, 0x60);
|
|
|
|
mdp_writel(mdp, MDP_ANY_INTR_MASK, MDP_INTR_ENABLE);
|
|
mdp_writel(mdp, 1, MDP_EBI2_PORTMAP_MODE);
|
|
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01f8);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01fc);
|
|
|
|
for (n = 0; n < ARRAY_SIZE(csc_table); n++)
|
|
mdp_writel(mdp, csc_table[n].val, csc_table[n].reg);
|
|
|
|
/* clear up unused fg/main registers */
|
|
/* comp.plane 2&3 ystride */
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0120);
|
|
|
|
/* unpacked pattern */
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x012c);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0130);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0134);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0158);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x015c);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0160);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0170);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0174);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x017c);
|
|
|
|
/* comp.plane 2 & 3 */
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0114);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x0118);
|
|
|
|
/* clear unused bg registers */
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01c8);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01d0);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01dc);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01e0);
|
|
mdp_writel(mdp, 0, MDP_CMD_DEBUG_ACCESS_BASE + 0x01e4);
|
|
|
|
for (n = 0; n < ARRAY_SIZE(mdp_upscale_table); n++)
|
|
mdp_writel(mdp, mdp_upscale_table[n].val,
|
|
mdp_upscale_table[n].reg);
|
|
|
|
for (n = 0; n < 9; n++)
|
|
mdp_writel(mdp, mdp_default_ccs[n], 0x40440 + 4 * n);
|
|
mdp_writel(mdp, mdp_default_ccs[9], 0x40500 + 4 * 0);
|
|
mdp_writel(mdp, mdp_default_ccs[10], 0x40500 + 4 * 0);
|
|
mdp_writel(mdp, mdp_default_ccs[11], 0x40500 + 4 * 0);
|
|
|
|
/* register mdp device */
|
|
mdp->mdp_dev.dev.parent = &pdev->dev;
|
|
mdp->mdp_dev.dev.class = mdp_class;
|
|
dev_set_name(&mdp->mdp_dev.dev, "mdp%d", pdev->id);
|
|
|
|
/* if you can remove the platform device you'd have to implement
|
|
* this:
|
|
mdp_dev.release = mdp_class; */
|
|
|
|
ret = device_register(&mdp->mdp_dev.dev);
|
|
if (ret)
|
|
goto error_device_register;
|
|
return 0;
|
|
|
|
error_device_register:
|
|
free_irq(mdp->irq, mdp);
|
|
error_request_irq:
|
|
error_get_clk:
|
|
iounmap(mdp->base);
|
|
error_get_irq:
|
|
error_ioremap:
|
|
kfree(mdp);
|
|
return ret;
|
|
}
|
|
|
|
static struct platform_driver msm_mdp_driver = {
|
|
.probe = mdp_probe,
|
|
.driver = {.name = "msm_mdp"},
|
|
};
|
|
|
|
static int __init mdp_init(void)
|
|
{
|
|
mdp_class = class_create(THIS_MODULE, "msm_mdp");
|
|
if (IS_ERR(mdp_class)) {
|
|
printk(KERN_ERR "Error creating mdp class\n");
|
|
return PTR_ERR(mdp_class);
|
|
}
|
|
return platform_driver_register(&msm_mdp_driver);
|
|
}
|
|
|
|
subsys_initcall(mdp_init);
|