--- /dev/null
+/* arch/arm/mach-rk29/vpu.c
+ *
+ * Copyright (C) 2010 ROCKCHIP, Inc.
+ *
+ * This software is licensed under the terms of the GNU General Public
+ * License version 2, as published by the Free Software Foundation, and
+ * may be copied, distributed, and modified under those terms.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ */
+
+#ifdef CONFIG_RK29_VPU_DEBUG
+#define DEBUG
+#define pr_fmt(fmt) "VPU: %s: " fmt, __func__
+#else
+#define pr_fmt(fmt) "VPU: " fmt
+#endif
+
+#include <linux/init.h>
+#include <linux/interrupt.h>
+#include <linux/io.h>
+#include <linux/kernel.h>
+#include <linux/module.h>
+#include <linux/fs.h>
+#include <linux/ioport.h>
+#include <linux/miscdevice.h>
+#include <linux/mm.h>
+#include <linux/poll.h>
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+#include <linux/time.h>
+#endif
+
+#include <asm/uaccess.h>
+
+#include <mach/irqs.h>
+#include <mach/vpu.h>
+#include <mach/rk29_iomap.h>
+
+#define DEC_INTERRUPT_REGISTER 1
+#define PP_INTERRUPT_REGISTER 60
+#define ENC_INTERRUPT_REGISTER 1
+
+#define DEC_INTERRUPT_BIT 0x100
+#define PP_INTERRUPT_BIT 0x100
+#define ENC_INTERRUPT_BIT 0x1
+
+#define DEC_IO_SIZE ((100 + 1) * 4) /* bytes */
+#define ENC_IO_SIZE (96 * 4) /* bytes */
+
+static const u16 dec_hw_ids[] = { 0x8190, 0x8170, 0x9170, 0x9190, 0x6731 };
+static const u16 enc_hw_ids[] = { 0x6280, 0x7280, 0x8270 };
+
+struct vpu_device {
+ unsigned long iobaseaddr;
+ unsigned int iosize;
+ volatile u32 *hwregs;
+ unsigned int irq;
+};
+
+static struct vpu_device dec_dev;
+static struct vpu_device pp_dev;
+static struct vpu_device enc_dev;
+
+struct vpu_client {
+ struct vpu_device *dev;
+ s32 event_count; /* for user */
+ atomic_t event; /* inc by irq */
+ struct fasync_struct *async_queue;
+ wait_queue_head_t wait;
+ struct file *filp; /* for /proc/vpu */
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ struct timespec end_time;
+#endif
+};
+static struct vpu_client client;
+
+static void vpu_release_io(void);
+#ifdef CONFIG_RK29_VPU_DEBUG
+static void dump_regs(unsigned long data);
+#endif
+
+static long vpu_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
+{
+ int err = 0;
+ struct vpu_device *dev = client.dev;
+
+ pr_debug("ioctl cmd 0x%08x\n", cmd);
+
+ if (!dev)
+ return -EINVAL;
+
+ /*
+ * extract the type and number bitfields, and don't decode
+ * wrong cmds: return ENOTTY (inappropriate ioctl) before access_ok()
+ */
+ if (_IOC_TYPE(cmd) != VPU_IOC_MAGIC)
+ return -ENOTTY;
+ if (_IOC_NR(cmd) > VPU_IOC_MAXNR)
+ return -ENOTTY;
+
+ /*
+ * the direction is a bitmask, and VERIFY_WRITE catches R/W
+ * transfers. `Type' is user-oriented, while
+ * access_ok is kernel-oriented, so the concept of "read" and
+ * "write" is reversed
+ */
+ if (_IOC_DIR(cmd) & _IOC_READ)
+ err = !access_ok(VERIFY_WRITE, (void *)arg, _IOC_SIZE(cmd));
+ else if (_IOC_DIR(cmd) & _IOC_WRITE)
+ err = !access_ok(VERIFY_READ, (void *)arg, _IOC_SIZE(cmd));
+ if (err)
+ return -EFAULT;
+
+ switch (cmd) {
+ case VPU_IOC_CLI:
+ disable_irq(dev->irq);
+ break;
+
+ case VPU_IOC_STI:
+ enable_irq(dev->irq);
+ break;
+
+ case VPU_IOC_GHWOFFSET:
+ put_user(dev->iobaseaddr, (unsigned long *)arg);
+ break;
+
+ case VPU_IOC_GHWIOSIZE:
+ put_user(dev->iosize, (unsigned int *)arg);
+ break;
+
+ case VPU_IOC_DEC_INSTANCE:
+ client.dev = &dec_dev;
+ break;
+
+ case VPU_IOC_PP_INSTANCE:
+ client.dev = &pp_dev;
+ break;
+
+ case VPU_IOC_ENC_INSTANCE:
+ client.dev = &enc_dev;
+ break;
+
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ case VPU_IOC_HW_PERFORMANCE:
+ put_user(client.end_time.tv_sec, (long *)arg);
+ put_user(client.end_time.tv_nsec, (long *)arg + 1);
+ break;
+#endif
+ }
+
+ return 0;
+}
+
+static int vpu_open(struct inode *inode, struct file *filp)
+{
+ if (client.filp)
+ return -EBUSY;
+ client.dev = &dec_dev;
+ client.event_count = atomic_read(&client.event);
+ client.filp = filp;
+ pr_debug("dev opened\n");
+ return nonseekable_open(inode, filp);
+}
+
+static int vpu_fasync(int fd, struct file *filp, int mode)
+{
+ return fasync_helper(fd, filp, mode, &client.async_queue);
+}
+
+static int vpu_release(struct inode *inode, struct file *filp)
+{
+#ifdef CONFIG_RK29_VPU_DEBUG
+ dump_regs(&enc_dev); /* dump the regs */
+ dump_regs(&dec_dev); /* dump the regs */
+#endif
+
+ /* remove this filp from the asynchronusly notified filp's */
+ vpu_fasync(-1, filp, 0);
+
+ client.async_queue = NULL;
+ client.filp = NULL;
+
+ pr_debug("dev closed\n");
+ return 0;
+}
+
+static int vpu_check_hw_id(struct vpu_device * dev, const u16 *hwids, size_t num)
+{
+ u32 hwid = readl(dev->hwregs);
+ pr_info("HW ID = 0x%08x\n", hwid);
+
+ hwid = (hwid >> 16) & 0xFFFF; /* product version only */
+
+ while (num--) {
+ if (hwid == hwids[num]) {
+ pr_info("Compatible HW found at 0x%08lx\n", dev->iobaseaddr);
+ return 1;
+ }
+ }
+
+ pr_info("No Compatible HW found at 0x%08lx\n", dev->iobaseaddr);
+ return 0;
+}
+
+static int vpu_reserve_io(void)
+{
+ if (!request_mem_region(dec_dev.iobaseaddr, dec_dev.iosize, "hx170dec")) {
+ pr_info("failed to reserve dec HW regs\n");
+ return -EBUSY;
+ }
+
+ dec_dev.hwregs =
+ (volatile u32 *)ioremap_nocache(dec_dev.iobaseaddr, dec_dev.iosize);
+
+ if (dec_dev.hwregs == NULL) {
+ pr_info("failed to ioremap dec HW regs\n");
+ goto err;
+ }
+
+ /* check for correct HW */
+ if (!vpu_check_hw_id(&dec_dev, dec_hw_ids, ARRAY_SIZE(dec_hw_ids))) {
+ goto err;
+ }
+
+ if (!request_mem_region(enc_dev.iobaseaddr, enc_dev.iosize, "hx280enc")) {
+ pr_info("failed to reserve enc HW regs\n");
+ goto err;
+ }
+
+ enc_dev.hwregs =
+ (volatile u32 *)ioremap_nocache(enc_dev.iobaseaddr, enc_dev.iosize);
+
+ if (enc_dev.hwregs == NULL) {
+ pr_info("failed to ioremap enc HW regs\n");
+ goto err;
+ }
+
+ /* check for correct HW */
+ if (!vpu_check_hw_id(&enc_dev, enc_hw_ids, ARRAY_SIZE(enc_hw_ids))) {
+ goto err;
+ }
+ return 0;
+
+err:
+ vpu_release_io();
+ return -EBUSY;
+}
+
+static void vpu_release_io(void)
+{
+ if (dec_dev.hwregs)
+ iounmap((void *)dec_dev.hwregs);
+ release_mem_region(dec_dev.iobaseaddr, dec_dev.iosize);
+
+ if (enc_dev.hwregs)
+ iounmap((void *)enc_dev.hwregs);
+ release_mem_region(enc_dev.iobaseaddr, enc_dev.iosize);
+}
+
+static void vpu_event_notify(void)
+{
+ atomic_inc(&client.event);
+ wake_up_interruptible(&client.wait);
+ if (client.async_queue)
+ kill_fasync(&client.async_queue, SIGIO, POLL_IN);
+}
+
+static irqreturn_t hx170dec_isr(int irq, void *dev_id)
+{
+ unsigned int handled = 0;
+
+ struct vpu_device *dev = (struct vpu_device *) dev_id;
+ u32 irq_status_dec;
+ u32 irq_status_pp;
+
+ handled = 0;
+
+ /* interrupt status register read */
+ irq_status_dec = readl(dev->hwregs + DEC_INTERRUPT_REGISTER);
+ irq_status_pp = readl(dev->hwregs + PP_INTERRUPT_REGISTER);
+
+ if (irq_status_dec & DEC_INTERRUPT_BIT) {
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ ktime_get_ts(&client.end_time);
+#endif
+ /* clear dec IRQ */
+ writel(irq_status_dec & (~DEC_INTERRUPT_BIT),
+ dev->hwregs + DEC_INTERRUPT_REGISTER);
+
+ vpu_event_notify();
+
+ pr_debug("DEC IRQ received!\n");
+ handled = 1;
+ }
+
+ if (irq_status_pp & PP_INTERRUPT_BIT) {
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ ktime_get_ts(&client.end_time);
+#endif
+ /* clear pp IRQ */
+ writel(irq_status_pp & (~DEC_INTERRUPT_BIT),
+ dev->hwregs + PP_INTERRUPT_REGISTER);
+
+ vpu_event_notify();
+
+ pr_debug("PP IRQ received!\n");
+ handled = 1;
+ }
+
+ if (!handled) {
+ pr_debug("IRQ received, but not x170's!\n");
+ }
+
+ return IRQ_RETVAL(handled);
+}
+
+static irqreturn_t hx280enc_isr(int irq, void *dev_id)
+{
+ struct vpu_device *dev = (struct vpu_device *) dev_id;
+ u32 irq_status;
+
+ irq_status = readl(dev->hwregs + ENC_INTERRUPT_REGISTER);
+
+ if (irq_status & ENC_INTERRUPT_BIT) {
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ ktime_get_ts(&client.end_time);
+#endif
+ /* clear enc IRQ */
+ writel(irq_status & (~ENC_INTERRUPT_BIT),
+ dev->hwregs + ENC_INTERRUPT_REGISTER);
+
+ vpu_event_notify();
+
+ pr_debug("ENC IRQ handled!\n");
+ return IRQ_HANDLED;
+ } else {
+ pr_debug("ENC IRQ received, but NOT handled!\n");
+ return IRQ_NONE;
+ }
+}
+
+static void vpu_reset_dec_asic(struct vpu_device * dev)
+{
+ unsigned int i, n = dev->iosize >> 2;
+
+ writel(0, dev->hwregs + DEC_INTERRUPT_REGISTER);
+
+ for (i = 1; i < n; i++) {
+ writel(0, dev->hwregs + i);
+ }
+}
+
+static void vpu_reset_enc_asic(struct vpu_device * dev)
+{
+ unsigned int i, n = dev->iosize >> 2;
+
+ writel(0, dev->hwregs + 14);
+
+ for (i = 4; i < n; i++) {
+ writel(0, dev->hwregs + i);
+ }
+}
+
+#ifdef CONFIG_RK29_VPU_DEBUG
+static void dump_regs(struct vpu_device *dev)
+{
+ unsigned int i, n = dev->iosize >> 2;
+
+ pr_debug("Reg Dump Start\n");
+ for (i = 0; i < n; i++) {
+ pr_debug("\tswreg%d = %08X\n", i, readl(dev->hwregs + i));
+ }
+ pr_debug("Reg Dump End\n");
+}
+#endif
+
+static int vpu_mmap(struct file *fp, struct vm_area_struct *vm)
+{
+ unsigned long pfn;
+
+ /* Only support the simple cases where we map in a register page. */
+ if (((vm->vm_end - vm->vm_start) > RK29_VCODEC_SIZE) || vm->vm_pgoff)
+ return -EINVAL;
+
+ vm->vm_flags |= VM_IO | VM_RESERVED;
+ vm->vm_page_prot = pgprot_noncached(vm->vm_page_prot);
+ pfn = RK29_VCODEC_PHYS >> PAGE_SHIFT;
+ pr_debug("size = 0x%x, page no. = 0x%x\n",
+ (int)(vm->vm_end - vm->vm_start), (int)pfn);
+ return remap_pfn_range(vm, vm->vm_start, pfn, vm->vm_end - vm->vm_start,
+ vm->vm_page_prot) ? -EAGAIN : 0;
+}
+
+static unsigned int vpu_poll(struct file *filep, poll_table *wait)
+{
+ poll_wait(filep, &client.wait, wait);
+ if (client.event_count != atomic_read(&client.event))
+ return POLLIN | POLLRDNORM;
+ return 0;
+}
+
+static ssize_t vpu_read(struct file *filep, char __user *buf,
+ size_t count, loff_t *ppos)
+{
+ DECLARE_WAITQUEUE(wait, current);
+ ssize_t retval;
+ s32 event_count;
+
+ if (count != sizeof(s32))
+ return -EINVAL;
+
+ add_wait_queue(&client.wait, &wait);
+
+ do {
+ set_current_state(TASK_INTERRUPTIBLE);
+
+ event_count = atomic_read(&client.event);
+ if (event_count != client.event_count) {
+ if (copy_to_user(buf, &event_count, count))
+ retval = -EFAULT;
+ else {
+ client.event_count = event_count;
+ retval = count;
+ }
+ break;
+ }
+
+ if (filep->f_flags & O_NONBLOCK) {
+ retval = -EAGAIN;
+ break;
+ }
+
+ if (signal_pending(current)) {
+ retval = -ERESTARTSYS;
+ break;
+ }
+ schedule();
+ } while (1);
+
+ __set_current_state(TASK_RUNNING);
+ remove_wait_queue(&client.wait, &wait);
+
+ return retval;
+}
+
+static const struct file_operations vpu_fops = {
+ .read = vpu_read,
+ .poll = vpu_poll,
+ .unlocked_ioctl = vpu_ioctl,
+ .mmap = vpu_mmap,
+ .open = vpu_open,
+ .release = vpu_release,
+ .fasync = vpu_fasync,
+};
+
+static struct miscdevice vpu_misc_device = {
+ .minor = MISC_DYNAMIC_MINOR,
+ .name = "vpu",
+ .fops = &vpu_fops,
+};
+
+static int __init vpu_init(void)
+{
+ int ret;
+
+ pr_debug("baseaddr = 0x%08x vdpu irq = %d vepu irq = %d\n", RK29_VCODEC_PHYS, IRQ_VDPU, IRQ_VEPU);
+
+ dec_dev.iobaseaddr = RK29_VCODEC_PHYS + 0x200;
+ dec_dev.iosize = DEC_IO_SIZE;
+ dec_dev.irq = IRQ_VDPU;
+
+ enc_dev.iobaseaddr = RK29_VCODEC_PHYS;
+ enc_dev.iosize = ENC_IO_SIZE;
+ enc_dev.irq = IRQ_VEPU;
+
+ ret = vpu_reserve_io();
+ if (ret < 0) {
+ goto err_reserve_io;
+ }
+ pp_dev = dec_dev;
+
+ init_waitqueue_head(&client.wait);
+ atomic_set(&client.event, 0);
+
+ vpu_reset_dec_asic(&dec_dev); /* reset hardware */
+ vpu_reset_enc_asic(&enc_dev); /* reset hardware */
+
+ /* get the IRQ line */
+ ret = request_irq(IRQ_VDPU, hx170dec_isr, 0, "hx170dec", (void *)&dec_dev);
+ if (ret != 0) {
+ pr_err("can't request vdpu irq %d\n", IRQ_VDPU);
+ goto err_req_vdpu_irq;
+ }
+
+ ret = request_irq(IRQ_VEPU, hx280enc_isr, 0, "hx280enc", (void *)&enc_dev);
+ if (ret != 0) {
+ pr_err("can't request vepu irq %d\n", IRQ_VEPU);
+ goto err_req_vepu_irq;
+ }
+
+ ret = misc_register(&vpu_misc_device);
+ if (ret) {
+ pr_err("misc_register failed\n");
+ goto err_register;
+ }
+
+ pr_info("init success\n");
+
+ return 0;
+
+err_register:
+ free_irq(IRQ_VEPU, (void *)&enc_dev);
+err_req_vepu_irq:
+ free_irq(IRQ_VDPU, (void *)&dec_dev);
+err_req_vdpu_irq:
+ vpu_release_io();
+err_reserve_io:
+ pr_info("init failed\n");
+ return ret;
+}
+
+static void __exit vpu_exit(void)
+{
+ /* clear dec IRQ */
+ writel(0, dec_dev.hwregs + DEC_INTERRUPT_REGISTER);
+ /* clear pp IRQ */
+ writel(0, dec_dev.hwregs + PP_INTERRUPT_REGISTER);
+
+ writel(0, enc_dev.hwregs + 14); /* disable HW */
+ /* clear enc IRQ */
+ writel(0, enc_dev.hwregs + ENC_INTERRUPT_REGISTER);
+
+#ifdef CONFIG_RK29_VPU_DEBUG
+ dump_regs(&enc_dev); /* dump the regs */
+ dump_regs(&dec_dev); /* dump the regs */
+#endif
+
+ misc_deregister(&vpu_misc_device);
+ free_irq(IRQ_VEPU, (void *)&enc_dev);
+ free_irq(IRQ_VDPU, (void *)&dec_dev);
+ vpu_release_io();
+}
+
+module_init(vpu_init);
+module_exit(vpu_exit);
+MODULE_LICENSE("GPL");
+
+#ifdef CONFIG_PROC_FS
+#include <linux/proc_fs.h>
+#include <linux/seq_file.h>
+
+static int proc_vpu_show(struct seq_file *s, void *v)
+{
+ unsigned int i, n;
+
+ if (client.filp) {
+ seq_printf(s, "Opened\n");
+ seq_printf(s, "%s instance\n", client.dev == &dec_dev ? "DEC" : client.dev == &pp_dev ? "PP" : "ENC");
+ } else {
+ seq_printf(s, "Closed\n");
+ }
+ seq_printf(s, "event_count %d event %d\n", client.event_count, atomic_read(&client.event));
+#ifdef CONFIG_RK29_VPU_HW_PERFORMANCE
+ seq_printf(s, "end_time: %ld.%09ld\n", client.end_time.tv_sec, client.end_time.tv_nsec);
+#endif
+
+ seq_printf(s, "\nENC Registers:\n");
+ n = enc_dev.iosize >> 2;
+ for (i = 0; i < n; i++) {
+ seq_printf(s, "\tswreg%d = %08X\n", i, readl(enc_dev.hwregs + i));
+ }
+ seq_printf(s, "\nDEC Registers:\n");
+ n = dec_dev.iosize >> 2;
+ for (i = 0; i < n; i++) {
+ seq_printf(s, "\tswreg%d = %08X\n", i, readl(dec_dev.hwregs + i));
+ }
+ return 0;
+}
+
+static int proc_vpu_open(struct inode *inode, struct file *file)
+{
+ return single_open(file, proc_vpu_show, NULL);
+}
+
+static const struct file_operations proc_vpu_fops = {
+ .open = proc_vpu_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = single_release,
+};
+
+static int __init vpu_proc_init(void)
+{
+ proc_create("vpu", 0, NULL, &proc_vpu_fops);
+ return 0;
+
+}
+late_initcall(vpu_proc_init);
+#endif /* CONFIG_PROC_FS */
+