jpu.c 43 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440
  1. // SPDX-License-Identifier: GPL-2.0+ OR BSD-3-Clause
  2. //--=========================================================================--
  3. // This file is linux device driver for JPU.
  4. //-----------------------------------------------------------------------------
  5. //
  6. // This confidential and proprietary software may be used only
  7. // as authorized by a licensing agreement from Chips&Media Inc.
  8. // In the event of publication, the following notice is applicable:
  9. //
  10. // (C) COPYRIGHT 2006 - 2016 CHIPS&MEDIA INC.
  11. // ALL RIGHTS RESERVED
  12. //
  13. // The entire notice above must be reproduced on all authorized
  14. // copies.
  15. // Copyright (C) 2022 StarFive Technology Co., Ltd.
  16. //--=========================================================================-
  17. #include <linux/kernel.h>
  18. #include <linux/device.h>
  19. #include <linux/mm.h>
  20. #include <linux/interrupt.h>
  21. #include <linux/ioport.h>
  22. #include <linux/module.h>
  23. #include <linux/platform_device.h>
  24. #include <linux/dma-mapping.h>
  25. #include <linux/of.h>
  26. #include <linux/of_address.h>
  27. #include <linux/pm_runtime.h>
  28. #include <linux/wait.h>
  29. #include <linux/list.h>
  30. #include <linux/clk.h>
  31. #include <linux/delay.h>
  32. #include <linux/uaccess.h>
  33. #include <linux/cdev.h>
  34. #include <linux/slab.h>
  35. #include <linux/sched.h>
  36. #include <linux/sched/signal.h>
  37. #include <linux/reset.h>
  38. #include <linux/version.h>
  39. #include <soc/sifive/sifive_l2_cache.h>
  40. #include "../../../jpuapi/jpuconfig.h"
  41. #include "jpu.h"
  42. //#define ENABLE_DEBUG_MSG
  43. #ifdef ENABLE_DEBUG_MSG
  44. #define DPRINTK(args...) printk(KERN_INFO args);
  45. #else
  46. #define DPRINTK(args...)
  47. #endif
  48. /* definitions to be changed as customer configuration */
  49. /* if you want to have clock gating scheme frame by frame */
  50. //#define JPU_SUPPORT_CLOCK_CONTROL
  51. #define JPU_SUPPORT_ISR
  52. //#define JPU_IRQ_CONTROL
  53. /* if clktree is work,try this...*/
  54. #define STARFIVE_JPU_SUPPORT_CLOCK_CONTROL
  55. /* if the platform driver knows the name of this driver */
  56. /* JPU_PLATFORM_DEVICE_NAME */
  57. #define JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  58. /* if this driver knows the dedicated video memory address */
  59. //#define JPU_SUPPORT_RESERVED_VIDEO_MEMORY //if this driver knows the dedicated video memory address
  60. static void starfive_flush_dcache(unsigned long start, unsigned long len)
  61. {
  62. sifive_l2_flush64_range(start, len);
  63. }
  64. #define JPU_PLATFORM_DEVICE_NAME "cnm_jpu"
  65. #define JPU_CLK_NAME "jpege"
  66. #define JPU_DEV_NAME "jpu"
  67. #define JPU_REG_BASE_ADDR 0x11900000
  68. #define JPU_REG_SIZE 0x300
  69. #ifdef JPU_SUPPORT_ISR
  70. #define JPU_IRQ_NUM 24
  71. /* if the driver want to disable and enable IRQ whenever interrupt asserted. */
  72. /*#define JPU_IRQ_CONTROL*/
  73. #endif
  74. #ifndef VM_RESERVED /*for kernel up to 3.7.0 version*/
  75. #define VM_RESERVED (VM_DONTEXPAND | VM_DONTDUMP)
  76. #endif
  77. struct device *jpu_dev;
  78. typedef struct jpu_drv_context_t {
  79. struct fasync_struct *async_queue;
  80. u32 open_count; /*!<< device reference count. Not instance count */
  81. u32 interrupt_reason[MAX_NUM_INSTANCE];
  82. } jpu_drv_context_t;
  83. /* To track the allocated memory buffer */
  84. typedef struct jpudrv_buffer_pool_t {
  85. struct list_head list;
  86. struct jpudrv_buffer_t jb;
  87. struct file* filp;
  88. } jpudrv_buffer_pool_t;
  89. /* To track the instance index and buffer in instance pool */
  90. typedef struct jpudrv_instance_list_t {
  91. struct list_head list;
  92. unsigned long inst_idx;
  93. struct file* filp;
  94. } jpudrv_instance_list_t;
  95. typedef struct jpudrv_instance_pool_t {
  96. unsigned char codecInstPool[MAX_NUM_INSTANCE][MAX_INST_HANDLE_SIZE];
  97. } jpudrv_instance_pool_t;
  98. #ifndef STARFIVE_JPU_SUPPORT_CLOCK_CONTROL
  99. typedef struct jpu_clkgen_t {
  100. void __iomem *en_ctrl;
  101. uint32_t rst_mask;
  102. } jpu_clkgen_t;
  103. #endif
  104. struct clk_bulk_data jpu_clks[] = {
  105. { .id = "axi_clk" },
  106. { .id = "core_clk" },
  107. { .id = "apb_clk" },
  108. { .id = "noc_bus" },
  109. };
  110. typedef struct jpu_clk_t {
  111. #ifndef STARFIVE_JPU_SUPPORT_CLOCK_CONTROL
  112. void __iomem *clkgen;
  113. void __iomem *rst_ctrl;
  114. void __iomem *rst_status;
  115. uint32_t en_shift;
  116. uint32_t en_mask;
  117. jpu_clkgen_t apb_clk;
  118. jpu_clkgen_t axi_clk;
  119. jpu_clkgen_t core_clk;
  120. #else
  121. struct clk_bulk_data *clks;
  122. struct reset_control *resets;
  123. int nr_clks;
  124. #endif
  125. struct device *dev;
  126. } jpu_clk_t;
  127. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  128. #include "jmm.h"
  129. static jpu_mm_t s_jmem;
  130. static jpudrv_buffer_t s_video_memory = {0};
  131. #endif /* JPU_SUPPORT_RESERVED_VIDEO_MEMORY */
  132. static int jpu_hw_reset(void);
  133. static void jpu_clk_disable(jpu_clk_t *clk);
  134. static int jpu_clk_enable(jpu_clk_t *clk);
  135. static jpu_clk_t *jpu_clk_get(struct platform_device *pdev);
  136. static void jpu_clk_put(jpu_clk_t *clk);
  137. static int jpu_pmu_enable(struct device *dev);
  138. static void jpu_pmu_disable(struct device *dev);
  139. // end customer definition
  140. static jpudrv_buffer_t s_instance_pool = {0};
  141. static jpu_drv_context_t s_jpu_drv_context;
  142. static dev_t s_jpu_devt;
  143. static int s_jpu_major;
  144. static struct cdev s_jpu_cdev;
  145. static struct class *s_jpu_class;
  146. static jpu_clk_t *s_jpu_clk;
  147. static int s_jpu_open_ref_count;
  148. #ifdef JPU_SUPPORT_ISR
  149. static int s_jpu_irq = JPU_IRQ_NUM;
  150. #endif
  151. static jpudrv_buffer_t s_jpu_register = {0};
  152. static int s_interrupt_flag[MAX_NUM_INSTANCE];
  153. static wait_queue_head_t s_interrupt_wait_q[MAX_NUM_INSTANCE];
  154. static spinlock_t s_jpu_lock = __SPIN_LOCK_UNLOCKED(s_jpu_lock);
  155. #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,36)
  156. static DECLARE_MUTEX(s_jpu_sem);
  157. #else
  158. static DEFINE_SEMAPHORE(s_jpu_sem);
  159. #endif
  160. static struct list_head s_jbp_head = LIST_HEAD_INIT(s_jbp_head);
  161. static struct list_head s_inst_list_head = LIST_HEAD_INIT(s_inst_list_head);
  162. #ifdef CONFIG_PM
  163. /* implement to power management functions */
  164. #endif
  165. #define NPT_BASE 0x0000
  166. #define NPT_REG_SIZE 0x300
  167. #define MJPEG_PIC_STATUS_REG(_inst_no) (NPT_BASE + (_inst_no*NPT_REG_SIZE) + 0x004)
  168. #define ReadJpuRegister(addr) *(volatile unsigned int *)(s_jpu_register.virt_addr + addr)
  169. #define WriteJpuRegister(addr, val) *(volatile unsigned int *)(s_jpu_register.virt_addr + addr) = (unsigned int)val
  170. #define WriteJpu(addr, val) *(volatile unsigned int *)(addr) = (unsigned int)val;
  171. static int jpu_alloc_dma_buffer(jpudrv_buffer_t *jb)
  172. {
  173. if (!jb)
  174. return -1;
  175. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  176. jb->phys_addr = (unsigned long long)jmem_alloc(&s_jmem, jb->size, 0);
  177. if ((unsigned long)jb->phys_addr == (unsigned long)-1) {
  178. printk(KERN_ERR "[JPUDRV] Physical memory allocation error size=%d\n", jb->size);
  179. return -1;
  180. }
  181. jb->base = (unsigned long)(s_video_memory.base + (jb->phys_addr - s_video_memory.phys_addr));
  182. #else
  183. jb->base = (unsigned long)dma_alloc_coherent(jpu_dev, PAGE_ALIGN(jb->size), (dma_addr_t *) (&jb->phys_addr), GFP_DMA | GFP_KERNEL);
  184. if ((void *)(jb->base) == NULL) {
  185. printk(KERN_ERR "[JPUDRV] Physical memory allocation error size=%d\n", jb->size);
  186. return -1;
  187. }
  188. starfive_flush_dcache(jb->phys_addr,PAGE_ALIGN(jb->size));
  189. #endif /* JPU_SUPPORT_RESERVED_VIDEO_MEMORY */
  190. return 0;
  191. }
  192. static void jpu_free_dma_buffer(jpudrv_buffer_t *jb)
  193. {
  194. if (!jb) {
  195. return;
  196. }
  197. if (jb->base)
  198. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  199. jmem_free(&s_jmem, jb->phys_addr, 0);
  200. #else
  201. dma_free_coherent(jpu_dev, PAGE_ALIGN(jb->size), (void *)jb->base, jb->phys_addr);
  202. #endif /* JPUR_SUPPORT_RESERVED_VIDEO_MEMORY */
  203. }
  204. static int jpu_free_instances(struct file *filp)
  205. {
  206. jpudrv_instance_list_t *vil, *n;
  207. jpudrv_instance_pool_t *vip;
  208. void *vip_base;
  209. int instance_pool_size_per_core;
  210. #if !defined(PTHREAD_MUTEX_ROBUST_NP)
  211. void *jdi_mutexes_base;
  212. const int PTHREAD_MUTEX_T_DESTROY_VALUE = 0xdead10cc;
  213. #endif
  214. DPRINTK("[JPUDRV] jpu_free_instances\n");
  215. instance_pool_size_per_core = (s_instance_pool.size/MAX_NUM_JPU_CORE); /* s_instance_pool.size assigned to the size of all core once call JDI_IOCTL_GET_INSTANCE_POOL by user. */
  216. list_for_each_entry_safe(vil, n, &s_inst_list_head, list)
  217. {
  218. if (vil->filp == filp) {
  219. vip_base = (void *)(s_instance_pool.base + instance_pool_size_per_core);
  220. DPRINTK("[JPUDRV] jpu_free_instances detect instance crash instIdx=%d, vip_base=%p, instance_pool_size_per_core=%d\n", (int)vil->inst_idx, vip_base, (int)instance_pool_size_per_core);
  221. vip = (jpudrv_instance_pool_t *)vip_base;
  222. if (vip) {
  223. memset(&vip->codecInstPool[vil->inst_idx], 0x00, 4); /* only first 4 byte is key point(inUse of CodecInst in jpuapi) to free the corresponding instance. */
  224. #if !defined(PTHREAD_MUTEX_ROBUST_NP)
  225. #define PTHREAD_MUTEX_T_HANDLE_SIZE 4
  226. jdi_mutexes_base = (vip_base + (instance_pool_size_per_core - PTHREAD_MUTEX_T_HANDLE_SIZE*4));
  227. DPRINTK("[JPUDRV] jpu_free_instances : force to destroy jdi_mutexes_base=%p in userspace \n", jdi_mutexes_base);
  228. if (jdi_mutexes_base) {
  229. int i;
  230. for (i = 0; i < 4; i++) {
  231. memcpy(jdi_mutexes_base, &PTHREAD_MUTEX_T_DESTROY_VALUE, PTHREAD_MUTEX_T_HANDLE_SIZE);
  232. jdi_mutexes_base += PTHREAD_MUTEX_T_HANDLE_SIZE;
  233. }
  234. }
  235. #endif
  236. }
  237. s_jpu_open_ref_count--;
  238. list_del(&vil->list);
  239. kfree(vil);
  240. }
  241. }
  242. return 1;
  243. }
  244. static int jpu_free_buffers(struct file *filp)
  245. {
  246. jpudrv_buffer_pool_t *pool, *n;
  247. jpudrv_buffer_t jb;
  248. DPRINTK("[JPUDRV] jpu_free_buffers\n");
  249. list_for_each_entry_safe(pool, n, &s_jbp_head, list)
  250. {
  251. if (pool->filp == filp) {
  252. jb = pool->jb;
  253. if (jb.base) {
  254. jpu_free_dma_buffer(&jb);
  255. list_del(&pool->list);
  256. kfree(pool);
  257. }
  258. }
  259. }
  260. return 0;
  261. }
  262. static irqreturn_t jpu_irq_handler(int irq, void *dev_id)
  263. {
  264. jpu_drv_context_t* dev = (jpu_drv_context_t *)dev_id;
  265. int i;
  266. u32 flag;
  267. DPRINTK("[JPUDRV][+]%s\n", __func__);
  268. #ifdef JPU_IRQ_CONTROL
  269. disable_irq_nosync(s_jpu_irq);
  270. #endif
  271. for (i=0; i<MAX_NUM_INSTANCE; i++) {
  272. flag = ReadJpuRegister(MJPEG_PIC_STATUS_REG(i));
  273. if (flag != 0) {
  274. break;
  275. }
  276. }
  277. dev->interrupt_reason[i] = flag;
  278. s_interrupt_flag[i] = 1;
  279. DPRINTK("[JPUDRV][%d] INTERRUPT FLAG: %08x, %08x\n", i, dev->interrupt_reason[i], MJPEG_PIC_STATUS_REG(i));
  280. if (dev->async_queue)
  281. kill_fasync(&dev->async_queue, SIGIO, POLL_IN); // notify the interrupt to userspace
  282. #ifndef JPU_IRQ_CONTROL
  283. WriteJpuRegister(MJPEG_PIC_STATUS_REG(i),flag); //clear interrut
  284. #endif
  285. wake_up_interruptible(&s_interrupt_wait_q[i]);
  286. DPRINTK("[JPUDRV][-]%s\n", __func__);
  287. return IRQ_HANDLED;
  288. }
  289. static int jpu_open(struct inode *inode, struct file *filp)
  290. {
  291. DPRINTK("[JPUDRV][+] %s\n", __func__);
  292. pm_runtime_get_sync(s_jpu_clk->dev);
  293. spin_lock(&s_jpu_lock);
  294. s_jpu_drv_context.open_count++;
  295. filp->private_data = (void *)(&s_jpu_drv_context);
  296. spin_unlock(&s_jpu_lock);
  297. DPRINTK("[JPUDRV][-] %s\n", __func__);
  298. return 0;
  299. }
  300. static long jpu_ioctl(struct file *filp, u_int cmd, u_long arg)
  301. {
  302. int ret = 0;
  303. switch (cmd)
  304. {
  305. case JDI_IOCTL_ALLOCATE_PHYSICAL_MEMORY:
  306. {
  307. jpudrv_buffer_pool_t *jbp;
  308. DPRINTK("[JPUDRV][+]JDI_IOCTL_ALLOCATE_PHYSICAL_MEMORY\n");
  309. if ((ret = down_interruptible(&s_jpu_sem)) == 0) {
  310. jbp = kzalloc(sizeof(jpudrv_buffer_pool_t), GFP_KERNEL);
  311. if (!jbp) {
  312. up(&s_jpu_sem);
  313. return -ENOMEM;
  314. }
  315. ret = copy_from_user(&(jbp->jb), (jpudrv_buffer_t *)arg, sizeof(jpudrv_buffer_t));
  316. if (ret)
  317. {
  318. kfree(jbp);
  319. up(&s_jpu_sem);
  320. return -EFAULT;
  321. }
  322. ret = jpu_alloc_dma_buffer(&(jbp->jb));
  323. if (ret == -1)
  324. {
  325. ret = -ENOMEM;
  326. kfree(jbp);
  327. up(&s_jpu_sem);
  328. break;
  329. }
  330. ret = copy_to_user((void __user *)arg, &(jbp->jb), sizeof(jpudrv_buffer_t));
  331. if (ret)
  332. {
  333. kfree(jbp);
  334. ret = -EFAULT;
  335. up(&s_jpu_sem);
  336. break;
  337. }
  338. jbp->filp = filp;
  339. spin_lock(&s_jpu_lock);
  340. list_add(&jbp->list, &s_jbp_head);
  341. spin_unlock(&s_jpu_lock);
  342. up(&s_jpu_sem);
  343. }
  344. DPRINTK("[JPUDRV][-]JDI_IOCTL_ALLOCATE_PHYSICAL_MEMORY\n");
  345. }
  346. break;
  347. case JDI_IOCTL_GET_PHYSICAL_MEMORY:
  348. {
  349. jpudrv_buffer_pool_t *jbp = NULL;
  350. void *user_address = NULL;
  351. struct task_struct *my_struct = NULL;
  352. struct mm_struct *mm = NULL;
  353. unsigned long address = 0;
  354. pgd_t *pgd = NULL;
  355. DPRINTK("[JPUDRV][+]JDI_IOCTL_GET_PHYSICAL_MEMORY\n");
  356. if ((ret = down_interruptible(&s_jpu_sem)) == 0) {
  357. jbp = kzalloc(sizeof(jpudrv_buffer_pool_t), GFP_KERNEL);
  358. if (!jbp) {
  359. up(&s_jpu_sem);
  360. return -ENOMEM;
  361. }
  362. ret = copy_from_user(&(jbp->jb), (jpudrv_buffer_t *)arg, sizeof(jpudrv_buffer_t));
  363. if (ret)
  364. {
  365. kfree(jbp);
  366. up(&s_jpu_sem);
  367. return -EFAULT;
  368. }
  369. user_address = (void *)jbp->jb.virt_addr;
  370. my_struct = get_current();
  371. mm = my_struct->mm;
  372. address = (unsigned long)user_address;
  373. pgd = pgd_offset(mm, address);
  374. if (!pgd_none(*pgd) && !pgd_bad(*pgd)) {
  375. p4d_t *p4d = p4d_offset(pgd, address);
  376. pud_t *pud = pud_offset(p4d, address);
  377. if (!pud_none(*pud) && !pud_bad(*pud)) {
  378. pmd_t *pmd = pmd_offset(pud, address);
  379. if (!pmd_none(*pmd) && !pmd_bad(*pmd)) {
  380. pte_t *pte = pte_offset_map(pmd, address);
  381. if (!pte_none(*pte)) {
  382. struct page *pg = pte_page(*pte);
  383. unsigned long phys = page_to_phys(pg);
  384. unsigned long virt = (unsigned long)phys_to_virt(phys);
  385. printk("phy address = %lx, virt = %lx\r\n", phys, virt);
  386. jbp->jb.phys_addr = phys;
  387. jbp->jb.base = virt;
  388. }
  389. pte_unmap(pte);
  390. }
  391. }
  392. }
  393. ret = copy_to_user((void __user *)arg, &(jbp->jb), sizeof(jpudrv_buffer_t));
  394. if (ret)
  395. {
  396. kfree(jbp);
  397. ret = -EFAULT;
  398. up(&s_jpu_sem);
  399. break;
  400. }
  401. kfree(jbp);
  402. up(&s_jpu_sem);
  403. }
  404. DPRINTK("[JPUDRV][-]JDI_IOCTL_GET_PHYSICAL_MEMORY\n");
  405. }
  406. break;
  407. case JDI_IOCTL_FREE_PHYSICALMEMORY:
  408. {
  409. jpudrv_buffer_pool_t *jbp, *n;
  410. jpudrv_buffer_t jb;
  411. DPRINTK("[JPUDRV][+]VDI_IOCTL_FREE_PHYSICALMEMORY\n");
  412. if ((ret = down_interruptible(&s_jpu_sem)) == 0) {
  413. ret = copy_from_user(&jb, (jpudrv_buffer_t *)arg, sizeof(jpudrv_buffer_t));
  414. if (ret) {
  415. up(&s_jpu_sem);
  416. return -EACCES;
  417. }
  418. if (jb.base)
  419. jpu_free_dma_buffer(&jb);
  420. spin_lock(&s_jpu_lock);
  421. list_for_each_entry_safe(jbp, n, &s_jbp_head, list) {
  422. if (jbp->jb.base == jb.base) {
  423. list_del(&jbp->list);
  424. kfree(jbp);
  425. break;
  426. }
  427. }
  428. spin_unlock(&s_jpu_lock);
  429. up(&s_jpu_sem);
  430. }
  431. DPRINTK("[JPUDRV][-]VDI_IOCTL_FREE_PHYSICALMEMORY\n");
  432. }
  433. break;
  434. case JDI_IOCTL_GET_RESERVED_VIDEO_MEMORY_INFO:
  435. {
  436. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  437. if (s_video_memory.base != 0) {
  438. ret = copy_to_user((void __user *)arg, &s_video_memory, sizeof(jpudrv_buffer_t));
  439. if (ret != 0)
  440. ret = -EFAULT;
  441. } else {
  442. ret = -EFAULT;
  443. }
  444. #endif /* JPU_SUPPORT_RESERVED_VIDEO_MEMORY */
  445. }
  446. break;
  447. case JDI_IOCTL_WAIT_INTERRUPT:
  448. {
  449. jpudrv_intr_info_t info;
  450. struct jpu_drv_context_t *dev = (struct jpu_drv_context_t *)filp->private_data;
  451. u32 instance_no;
  452. DPRINTK("[JPUDRV][+]JDI_IOCTL_WAIT_INTERRUPT\n");
  453. ret = copy_from_user(&info, (jpudrv_intr_info_t *)arg, sizeof(jpudrv_intr_info_t));
  454. if (ret != 0)
  455. return -EFAULT;
  456. instance_no = info.inst_idx;
  457. DPRINTK("[JPUDRV] INSTANCE NO: %d\n", instance_no);
  458. ret = wait_event_interruptible_timeout(s_interrupt_wait_q[instance_no], s_interrupt_flag[instance_no] != 0, msecs_to_jiffies(info.timeout));
  459. if (!ret) {
  460. DPRINTK("[JPUDRV] INSTANCE NO: %d ETIME\n", instance_no);
  461. ret = -ETIME;
  462. break;
  463. }
  464. if (signal_pending(current)) {
  465. ret = -ERESTARTSYS;
  466. DPRINTK("[JPUDRV] INSTANCE NO: %d ERESTARTSYS\n", instance_no);
  467. break;
  468. }
  469. DPRINTK("[JPUDRV] INST(%d) s_interrupt_flag(%d), reason(0x%08x)\n", instance_no, s_interrupt_flag[instance_no], dev->interrupt_reason[instance_no]);
  470. info.intr_reason = dev->interrupt_reason[instance_no];
  471. s_interrupt_flag[instance_no] = 0;
  472. dev->interrupt_reason[instance_no] = 0;
  473. ret = copy_to_user((void __user *)arg, &info, sizeof(jpudrv_intr_info_t));
  474. #ifdef JPU_IRQ_CONTROL
  475. enable_irq(s_jpu_irq);
  476. #endif
  477. DPRINTK("[VPUDRV][-]VDI_IOCTL_WAIT_INTERRUPT\n");
  478. if (ret != 0)
  479. return -EFAULT;
  480. }
  481. break;
  482. case JDI_IOCTL_SET_CLOCK_GATE:
  483. {
  484. u32 clkgate;
  485. if (get_user(clkgate, (u32 __user *) arg))
  486. return -EFAULT;
  487. #ifdef JPU_SUPPORT_CLOCK_CONTROL
  488. if (clkgate)
  489. jpu_clk_enable(s_jpu_clk);
  490. else
  491. jpu_clk_disable(s_jpu_clk);
  492. #endif /* JPU_SUPPORT_CLOCK_CONTROL */
  493. }
  494. break;
  495. case JDI_IOCTL_GET_INSTANCE_POOL:
  496. DPRINTK("[JPUDRV][+]JDI_IOCTL_GET_INSTANCE_POOL\n");
  497. if ((ret = down_interruptible(&s_jpu_sem)) == 0) {
  498. if (s_instance_pool.base != 0) {
  499. ret = copy_to_user((void __user *)arg, &s_instance_pool, sizeof(jpudrv_buffer_t));
  500. } else {
  501. ret = copy_from_user(&s_instance_pool, (jpudrv_buffer_t *)arg, sizeof(jpudrv_buffer_t));
  502. if (ret == 0) {
  503. s_instance_pool.size = PAGE_ALIGN(s_instance_pool.size);
  504. s_instance_pool.base = (unsigned long)vmalloc(s_instance_pool.size);
  505. s_instance_pool.phys_addr = s_instance_pool.base;
  506. if (s_instance_pool.base != 0) {
  507. memset((void *)s_instance_pool.base, 0x0, s_instance_pool.size); /*clearing memory*/
  508. ret = copy_to_user((void __user *)arg, &s_instance_pool, sizeof(jpudrv_buffer_t));
  509. if (ret == 0) {
  510. /* success to get memory for instance pool */
  511. up(&s_jpu_sem);
  512. break;
  513. }
  514. }
  515. ret = -EFAULT;
  516. }
  517. }
  518. up(&s_jpu_sem);
  519. }
  520. DPRINTK("[JPUDRV][-]JDI_IOCTL_GET_INSTANCE_POOL: %s base: %lx, size: %d\n",
  521. (ret==0 ? "OK" : "NG"), s_instance_pool.base, s_instance_pool.size);
  522. break;
  523. case JDI_IOCTL_OPEN_INSTANCE:
  524. {
  525. jpudrv_inst_info_t inst_info;
  526. if (copy_from_user(&inst_info, (jpudrv_inst_info_t *)arg, sizeof(jpudrv_inst_info_t)))
  527. return -EFAULT;
  528. spin_lock(&s_jpu_lock);
  529. s_jpu_open_ref_count++; /* flag just for that jpu is in opened or closed */
  530. inst_info.inst_open_count = s_jpu_open_ref_count;
  531. spin_unlock(&s_jpu_lock);
  532. if (copy_to_user((void __user *)arg, &inst_info, sizeof(jpudrv_inst_info_t))) {
  533. return -EFAULT;
  534. }
  535. DPRINTK("[JPUDRV] JDI_IOCTL_OPEN_INSTANCE inst_idx=%d, s_jpu_open_ref_count=%d, inst_open_count=%d\n",
  536. (int)inst_info.inst_idx, s_jpu_open_ref_count, inst_info.inst_open_count);
  537. }
  538. break;
  539. case JDI_IOCTL_CLOSE_INSTANCE:
  540. {
  541. jpudrv_inst_info_t inst_info;
  542. DPRINTK("[JPUDRV][+]JDI_IOCTL_CLOSE_INSTANCE\n");
  543. if (copy_from_user(&inst_info, (jpudrv_inst_info_t *)arg, sizeof(jpudrv_inst_info_t)))
  544. return -EFAULT;
  545. spin_lock(&s_jpu_lock);
  546. s_jpu_open_ref_count--; /* flag just for that jpu is in opened or closed */
  547. inst_info.inst_open_count = s_jpu_open_ref_count;
  548. spin_unlock(&s_jpu_lock);
  549. if (copy_to_user((void __user *)arg, &inst_info, sizeof(jpudrv_inst_info_t)))
  550. return -EFAULT;
  551. DPRINTK("[JPUDRV] JDI_IOCTL_CLOSE_INSTANCE inst_idx=%d, s_jpu_open_ref_count=%d, inst_open_count=%d\n",
  552. (int)inst_info.inst_idx, s_jpu_open_ref_count, inst_info.inst_open_count);
  553. }
  554. break;
  555. case JDI_IOCTL_GET_INSTANCE_NUM:
  556. {
  557. jpudrv_inst_info_t inst_info;
  558. DPRINTK("[JPUDRV][+]JDI_IOCTL_GET_INSTANCE_NUM\n");
  559. ret = copy_from_user(&inst_info, (jpudrv_inst_info_t *)arg, sizeof(jpudrv_inst_info_t));
  560. if (ret != 0)
  561. break;
  562. spin_lock(&s_jpu_lock);
  563. inst_info.inst_open_count = s_jpu_open_ref_count;
  564. spin_unlock(&s_jpu_lock);
  565. ret = copy_to_user((void __user *)arg, &inst_info, sizeof(jpudrv_inst_info_t));
  566. DPRINTK("[JPUDRV] JDI_IOCTL_GET_INSTANCE_NUM inst_idx=%d, open_count=%d\n", (int)inst_info.inst_idx, inst_info.inst_open_count);
  567. }
  568. break;
  569. case JDI_IOCTL_RESET:
  570. jpu_hw_reset();
  571. break;
  572. case JDI_IOCTL_GET_REGISTER_INFO:
  573. DPRINTK("[JPUDRV][+]JDI_IOCTL_GET_REGISTER_INFO\n");
  574. ret = copy_to_user((void __user *)arg, &s_jpu_register, sizeof(jpudrv_buffer_t));
  575. if (ret != 0)
  576. ret = -EFAULT;
  577. DPRINTK("[JPUDRV][-]JDI_IOCTL_GET_REGISTER_INFO s_jpu_register.phys_addr=0x%lx, s_jpu_register.virt_addr=0x%lx, s_jpu_register.size=%d\n",
  578. s_jpu_register.phys_addr , s_jpu_register.virt_addr, s_jpu_register.size);
  579. break;
  580. case JDI_IOCTL_FLUSH_DCACHE:
  581. {
  582. jpudrv_flush_cache_t cache_info;
  583. //DPRINTK("[JPUDRV][+]JDI_IOCTL_FLUSH_DCACHE\n");
  584. ret = copy_from_user(&cache_info, (jpudrv_flush_cache_t *)arg, sizeof(jpudrv_flush_cache_t));
  585. if (ret != 0)
  586. ret = -EFAULT;
  587. if(cache_info.flag)
  588. starfive_flush_dcache(cache_info.start,cache_info.size);
  589. //DPRINTK("[JPUDRV][-]JDI_IOCTL_FLUSH_DCACHE\n");
  590. break;
  591. }
  592. default:
  593. {
  594. printk(KERN_ERR "No such IOCTL, cmd is %d\n", cmd);
  595. }
  596. break;
  597. }
  598. return ret;
  599. }
  600. static ssize_t jpu_read(struct file *filp, char __user *buf, size_t len, loff_t *ppos)
  601. {
  602. return -1;
  603. }
  604. static ssize_t jpu_write(struct file *filp, const char __user *buf, size_t len, loff_t *ppos)
  605. {
  606. /* DPRINTK("[VPUDRV] vpu_write len=%d\n", (int)len); */
  607. if (!buf) {
  608. printk(KERN_ERR "[VPUDRV] vpu_write buf = NULL error \n");
  609. return -EFAULT;
  610. }
  611. return -1;
  612. }
  613. static int jpu_release(struct inode *inode, struct file *filp)
  614. {
  615. int ret = 0;
  616. u32 open_count;
  617. DPRINTK("[JPUDRV][+] jpu_release\n");
  618. if ((ret = down_interruptible(&s_jpu_sem)) == 0) {
  619. /* found and free the not handled buffer by user applications */
  620. jpu_free_buffers(filp);
  621. /* found and free the not closed instance by user applications */
  622. jpu_free_instances(filp);
  623. DPRINTK("[JPUDRV] open_count: %d\n", s_jpu_drv_context.open_count);
  624. spin_lock(&s_jpu_lock);
  625. s_jpu_drv_context.open_count--;
  626. open_count = s_jpu_drv_context.open_count;
  627. spin_unlock(&s_jpu_lock);
  628. if (open_count == 0) {
  629. if (s_instance_pool.base) {
  630. DPRINTK("[JPUDRV] free instance pool\n");
  631. vfree((const void *)s_instance_pool.base);
  632. s_instance_pool.base = 0;
  633. }
  634. }
  635. }
  636. up(&s_jpu_sem);
  637. DPRINTK("[JPUDRV][-] jpu_release\n");
  638. pm_runtime_put_sync(s_jpu_clk->dev);
  639. return 0;
  640. }
  641. static int jpu_fasync(int fd, struct file *filp, int mode)
  642. {
  643. struct jpu_drv_context_t *dev = (struct jpu_drv_context_t *)filp->private_data;
  644. return fasync_helper(fd, filp, mode, &dev->async_queue);
  645. }
  646. static int jpu_map_to_register(struct file *fp, struct vm_area_struct *vm)
  647. {
  648. unsigned long pfn;
  649. vm->vm_flags |= VM_IO | VM_RESERVED;
  650. vm->vm_page_prot = pgprot_noncached(vm->vm_page_prot);
  651. pfn = s_jpu_register.phys_addr >> PAGE_SHIFT;
  652. return remap_pfn_range(vm, vm->vm_start, pfn, vm->vm_end-vm->vm_start, vm->vm_page_prot) ? -EAGAIN : 0;
  653. }
  654. static int jpu_map_to_physical_memory(struct file *fp, struct vm_area_struct *vm)
  655. {
  656. vm->vm_flags |= VM_IO | VM_RESERVED;
  657. vm->vm_page_prot = pgprot_noncached(vm->vm_page_prot);
  658. return remap_pfn_range(vm, vm->vm_start, vm->vm_pgoff, vm->vm_end-vm->vm_start, vm->vm_page_prot) ? -EAGAIN : 0;
  659. }
  660. static int jpu_map_to_instance_pool_memory(struct file *fp, struct vm_area_struct *vm)
  661. {
  662. int ret;
  663. long length = vm->vm_end - vm->vm_start;
  664. unsigned long start = vm->vm_start;
  665. char *vmalloc_area_ptr = (char *)s_instance_pool.base;
  666. unsigned long pfn;
  667. vm->vm_flags |= VM_RESERVED;
  668. /* loop over all pages, map it page individually */
  669. while (length > 0) {
  670. pfn = vmalloc_to_pfn(vmalloc_area_ptr);
  671. if ((ret = remap_pfn_range(vm, start, pfn, PAGE_SIZE, PAGE_SHARED)) < 0) {
  672. return ret;
  673. }
  674. start += PAGE_SIZE;
  675. vmalloc_area_ptr += PAGE_SIZE;
  676. length -= PAGE_SIZE;
  677. }
  678. return 0;
  679. }
  680. /*!
  681. * @brief memory map interface for jpu file operation
  682. * @return 0 on success or negative error code on error
  683. */
  684. static int jpu_mmap(struct file *fp, struct vm_area_struct *vm)
  685. {
  686. if (vm->vm_pgoff == 0)
  687. return jpu_map_to_instance_pool_memory(fp, vm);
  688. if (vm->vm_pgoff == (s_jpu_register.phys_addr>>PAGE_SHIFT))
  689. return jpu_map_to_register(fp, vm);
  690. return jpu_map_to_physical_memory(fp, vm);
  691. }
  692. struct file_operations jpu_fops = {
  693. .owner = THIS_MODULE,
  694. .open = jpu_open,
  695. .read = jpu_read,
  696. .write = jpu_write,
  697. .unlocked_ioctl = jpu_ioctl,
  698. .release = jpu_release,
  699. .fasync = jpu_fasync,
  700. .mmap = jpu_mmap,
  701. };
  702. static int jpu_probe(struct platform_device *pdev)
  703. {
  704. int err = 0;
  705. struct resource *res = NULL;
  706. struct device *devices;
  707. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  708. struct resource res_cma;
  709. struct device_node *node;
  710. #endif
  711. DPRINTK("[JPUDRV] jpu_probe\n");
  712. if (pdev) {
  713. res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
  714. }
  715. if (res) {/* if platform driver is implemented */
  716. s_jpu_register.phys_addr = res->start;
  717. s_jpu_register.virt_addr = (unsigned long)ioremap(res->start, res->end - res->start);
  718. s_jpu_register.size = res->end - res->start;
  719. DPRINTK("[JPUDRV] : jpu base address get from platform driver physical base addr==0x%lx, virtual base=0x%lx\n", s_jpu_register.phys_addr , s_jpu_register.virt_addr);
  720. } else {
  721. s_jpu_register.phys_addr = JPU_REG_BASE_ADDR;
  722. s_jpu_register.virt_addr = (unsigned long)ioremap(s_jpu_register.phys_addr, JPU_REG_SIZE);
  723. s_jpu_register.size = JPU_REG_SIZE;
  724. DPRINTK("[JPUDRV] : jpu base address get from defined value physical base addr==0x%lx, virtual base=0x%lx\n", s_jpu_register.phys_addr, s_jpu_register.virt_addr);
  725. }
  726. if (pdev) {
  727. jpu_dev = &pdev->dev;
  728. //jpu_dev->dma_ops = NULL;
  729. dev_info(jpu_dev,"init device.\n");
  730. }
  731. /* get the major number of the character device */
  732. if ((alloc_chrdev_region(&s_jpu_devt, 0, 1, JPU_DEV_NAME)) < 0) {
  733. err = -EBUSY;
  734. printk(KERN_ERR "could not allocate major number\n");
  735. goto ERROR_PROVE_DEVICE;
  736. }
  737. s_jpu_major = MAJOR(s_jpu_devt);
  738. /* initialize the device structure and register the device with the kernel */
  739. cdev_init(&s_jpu_cdev, &jpu_fops);
  740. if ((cdev_add(&s_jpu_cdev, s_jpu_devt, 1)) < 0) {
  741. err = -EBUSY;
  742. printk(KERN_ERR "could not allocate chrdev\n");
  743. goto ERROR_PROVE_DEVICE;
  744. }
  745. s_jpu_class = class_create(THIS_MODULE, JPU_DEV_NAME);
  746. if (IS_ERR(s_jpu_class)) {
  747. dev_err(jpu_dev, "class creat error.\n");
  748. goto ERROR_CRART_CLASS;
  749. }
  750. devices = device_create(s_jpu_class, 0, MKDEV(s_jpu_major, 0),
  751. NULL, JPU_DEV_NAME);
  752. if (IS_ERR(devices)) {
  753. dev_err(jpu_dev, "device creat error.\n");
  754. goto ERROR_CREAT_DEVICE;
  755. }
  756. if (pdev)
  757. s_jpu_clk = jpu_clk_get(pdev);
  758. else
  759. s_jpu_clk = jpu_clk_get(NULL);
  760. if (!s_jpu_clk) {
  761. printk(KERN_ERR "[JPUDRV] : not support clock controller.\n");
  762. }
  763. else {
  764. DPRINTK("[JPUDRV] : get clock controller s_jpu_clk=%p\n", s_jpu_clk);
  765. }
  766. jpu_pmu_enable(s_jpu_clk->dev);
  767. jpu_clk_enable(s_jpu_clk);
  768. reset_control_deassert(s_jpu_clk->resets);
  769. #ifdef JPU_SUPPORT_ISR
  770. #ifdef JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  771. if(pdev)
  772. res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
  773. if (res) {/* if platform driver is implemented */
  774. s_jpu_irq = res->start;
  775. DPRINTK("[JPUDRV] : jpu irq number get from platform driver irq=0x%x\n", s_jpu_irq );
  776. } else {
  777. DPRINTK("[JPUDRV] : jpu irq number get from defined value irq=0x%x\n", s_jpu_irq );
  778. }
  779. #else
  780. DPRINTK("[JPUDRV] : jpu irq number get from defined value irq=0x%x\n", s_jpu_irq);
  781. #endif
  782. err = request_irq(s_jpu_irq, jpu_irq_handler, 0, "JPU_CODEC_IRQ", (void *)(&s_jpu_drv_context));
  783. if (err) {
  784. printk(KERN_ERR "[JPUDRV] : fail to register interrupt handler\n");
  785. goto ERROR_PROVE_DEVICE;
  786. }
  787. #endif
  788. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  789. node = of_parse_phandle(jpu_dev->of_node, "memory-region", 0);
  790. if(node){
  791. dev_info(jpu_dev, "Get mem form memory-region\n");
  792. of_address_to_resource(node, 0, &res_cma);
  793. s_video_memory.size = resource_size(&res_cma);
  794. s_video_memory.phys_addr = res_cma.start;
  795. }else{
  796. dev_info(jpu_dev, "Get mem form reserved memory failed.please check the dts file.\n");
  797. return 0;
  798. }
  799. s_video_memory.base = (unsigned long)ioremap(MEM2SYS(s_video_memory.phys_addr), PAGE_ALIGN(s_video_memory.size));
  800. if (!s_video_memory.base) {
  801. printk(KERN_ERR "[JPUDRV] : fail to remap video memory physical phys_addr=0x%lx, base=0x%lx, size=%d\n", MEM2SYS(s_video_memory.phys_addr), s_video_memory.base, s_video_memory.size);
  802. goto ERROR_PROVE_DEVICE;
  803. }
  804. if (jmem_init(&s_jmem, s_video_memory.phys_addr, s_video_memory.size) < 0) {
  805. printk(KERN_ERR "[JPUDRV] : fail to init vmem system\n");
  806. goto ERROR_PROVE_DEVICE;
  807. }
  808. DPRINTK("[JPUDRV] success to probe jpu device with reserved video memory phys_addr=0x%lx, base=0x%lx\n", s_video_memory.phys_addr, s_video_memory.base);
  809. #else
  810. DPRINTK("[JPUDRV] success to probe jpu device with non reserved video memory\n");
  811. #endif
  812. return 0;
  813. ERROR_CREAT_DEVICE:
  814. class_destroy(s_jpu_class);
  815. ERROR_CRART_CLASS:
  816. cdev_del(&s_jpu_cdev);
  817. ERROR_PROVE_DEVICE:
  818. if (s_jpu_major)
  819. unregister_chrdev_region(s_jpu_major, 1);
  820. if (s_jpu_register.virt_addr)
  821. iounmap((void *)s_jpu_register.virt_addr);
  822. return err;
  823. }
  824. static int jpu_remove(struct platform_device *pdev)
  825. {
  826. DPRINTK("[JPUDRV] jpu_remove\n");
  827. #ifdef JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  828. if (s_instance_pool.base) {
  829. vfree((const void *)s_instance_pool.base);
  830. s_instance_pool.base = 0;
  831. }
  832. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  833. if (s_video_memory.base) {
  834. iounmap((void *)s_video_memory.base);
  835. s_video_memory.base = 0;
  836. jmem_exit(&s_jmem);
  837. }
  838. #endif
  839. if (s_jpu_major > 0) {
  840. device_destroy(s_jpu_class, MKDEV(s_jpu_major, 0));
  841. class_destroy(s_jpu_class);
  842. cdev_del(&s_jpu_cdev);
  843. unregister_chrdev_region(s_jpu_devt, 1);
  844. s_jpu_major = 0;
  845. }
  846. #ifdef JPU_SUPPORT_ISR
  847. if (s_jpu_irq)
  848. free_irq(s_jpu_irq, &s_jpu_drv_context);
  849. #endif
  850. if (s_jpu_register.virt_addr)
  851. iounmap((void*)s_jpu_register.virt_addr);
  852. jpu_clk_put(s_jpu_clk);
  853. jpu_pmu_disable(&pdev->dev);
  854. #endif /* JPU_SUPPORT_PLATFORM_DRIVER_REGISTER */
  855. return 0;
  856. }
  857. #ifdef CONFIG_PM
  858. static int __maybe_unused jpu_runtime_suspend(struct device *dev)
  859. {
  860. reset_control_assert(s_jpu_clk->resets);
  861. jpu_clk_disable(s_jpu_clk);
  862. return 0;
  863. }
  864. static int __maybe_unused jpu_runtime_resume(struct device *dev)
  865. {
  866. jpu_clk_enable(s_jpu_clk);
  867. return reset_control_deassert(s_jpu_clk->resets);
  868. }
  869. #endif /* CONFIG_PM */
  870. #ifdef CONFIG_PM_SLEEP
  871. static int __maybe_unused jpu_suspend(struct device *dev)
  872. {
  873. return 0;
  874. }
  875. static int __maybe_unused jpu_resume(struct device *dev)
  876. {
  877. return 0;
  878. }
  879. #endif /* CONFIG_PM_SLEEP */
  880. static const struct dev_pm_ops cm_jpu_pm_ops = {
  881. SET_RUNTIME_PM_OPS(jpu_runtime_suspend,
  882. jpu_runtime_resume, NULL)
  883. //SET_SYSTEM_SLEEP_PM_OPS(jpu_suspend, jpu_resume)
  884. };
  885. #ifdef JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  886. static const struct of_device_id jpu_of_id_table[] = {
  887. { .compatible = "cm,codaj12-jpu-1" },
  888. { .compatible = "starfive,jpu" },
  889. {}
  890. };
  891. MODULE_DEVICE_TABLE(of, jpu_of_id_table);
  892. static struct platform_driver jpu_driver = {
  893. .driver = {
  894. .name = JPU_PLATFORM_DEVICE_NAME,
  895. .of_match_table = of_match_ptr(jpu_of_id_table),
  896. .pm = &cm_jpu_pm_ops,
  897. },
  898. .probe = jpu_probe,
  899. .remove = jpu_remove,
  900. };
  901. #endif /* JPU_SUPPORT_PLATFORM_DRIVER_REGISTER */
  902. static int __init jpu_init(void)
  903. {
  904. int res = 0;
  905. u32 i;
  906. DPRINTK("[JPUDRV] begin jpu_init\n");
  907. for (i=0; i<MAX_NUM_INSTANCE; i++) {
  908. init_waitqueue_head(&s_interrupt_wait_q[i]);
  909. }
  910. s_instance_pool.base = 0;
  911. #ifdef JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  912. res = platform_driver_register(&jpu_driver);
  913. #else
  914. res = jpu_probe(NULL);
  915. #endif /* JPU_SUPPORT_PLATFORM_DRIVER_REGISTER */
  916. DPRINTK("[JPUDRV] end jpu_init result=0x%x\n", res);
  917. return res;
  918. }
  919. static void __exit jpu_exit(void)
  920. {
  921. DPRINTK("[JPUDRV] [+]jpu_exit\n");
  922. #ifdef JPU_SUPPORT_PLATFORM_DRIVER_REGISTER
  923. platform_driver_unregister(&jpu_driver);
  924. #else /* JPU_SUPPORT_PLATFORM_DRIVER_REGISTER */
  925. #ifdef JPU_SUPPORT_CLOCK_CONTROL
  926. #else
  927. jpu_clk_disable(s_jpu_clk);
  928. #endif /* JPU_SUPPORT_CLOCK_CONTROL */
  929. jpu_clk_put(s_jpu_clk);
  930. if (s_instance_pool.base) {
  931. vfree((const void *)s_instance_pool.base);
  932. s_instance_pool.base = 0;
  933. }
  934. #ifdef JPU_SUPPORT_RESERVED_VIDEO_MEMORY
  935. if (s_video_memory.base) {
  936. iounmap((void *)s_video_memory.base);
  937. s_video_memory.base = 0;
  938. jmem_exit(&s_jmem);
  939. }
  940. #endif /* JPU_SUPPORT_RESERVED_VIDEO_MEMORY */
  941. if (s_jpu_major > 0) {
  942. device_destroy(s_jpu_class, MKDEV(s_jpu_major, 0));
  943. class_destroy(s_jpu_class);
  944. cdev_del(&s_jpu_cdev);
  945. unregister_chrdev_region(s_jpu_devt, 1);
  946. s_jpu_major = 0;
  947. }
  948. #ifdef JPU_SUPPORT_ISR
  949. if (s_jpu_irq)
  950. free_irq(s_jpu_irq, &s_jpu_drv_context);
  951. #endif /* JPU_SUPPORT_ISR */
  952. if (s_jpu_register.virt_addr) {
  953. iounmap((void *)s_jpu_register.virt_addr);
  954. s_jpu_register.virt_addr = 0x00;
  955. }
  956. #endif /* JPU_SUPPORT_PLATFORM_DRIVER_REGISTER */
  957. DPRINTK("[JPUDRV] [-]jpu_exit\n");
  958. return;
  959. }
  960. MODULE_AUTHOR("A customer using C&M JPU, Inc.");
  961. MODULE_DESCRIPTION("JPU linux driver");
  962. MODULE_LICENSE("Dual BSD/GPL");
  963. module_init(jpu_init);
  964. module_exit(jpu_exit);
  965. static int jpu_pmu_enable(struct device *dev)
  966. {
  967. pm_runtime_set_active(dev);
  968. pm_runtime_enable(dev);
  969. return 0;
  970. }
  971. static void jpu_pmu_disable(struct device *dev)
  972. {
  973. pm_runtime_disable(dev);
  974. pm_runtime_set_suspended(dev);
  975. }
  976. #ifndef STARFIVE_JPU_SUPPORT_CLOCK_CONTROL
  977. #define CLK_ENABLE_DATA 1
  978. #define CLK_DISABLE_DATA 0
  979. #define CLK_EN_SHIFT 31
  980. #define CLK_EN_MASK 0x80000000U
  981. #define SAIF_BD_APBS_BASE 0x13020000
  982. #define CODAJ12_CLK_AXI_CTRL 0x108U
  983. #define CODAJ12_CLK_APB_CTRL 0x110U
  984. #define CODAJ12_CLK_CORE_CTRL 0x10cU
  985. #define RSTGEN_SOFTWARE_RESET_ASSERT1 0x2FCU
  986. #define RSTGEN_SOFTWARE_RESET_STATUS1 0x30CU
  987. #define RSTN_AXI_MASK (0x1 << 12)
  988. #define RSTN_CORE_MASK (0x1 << 13)
  989. #define RSTN_APB_MASK (0x1 << 14)
  990. static __maybe_unused uint32_t saif_get_reg(
  991. const volatile void __iomem *addr,
  992. uint32_t shift, uint32_t mask)
  993. {
  994. u32 tmp;
  995. tmp = readl(addr);
  996. tmp = (tmp & mask) >> shift;
  997. return tmp;
  998. }
  999. static void saif_set_reg(volatile void __iomem *addr, uint32_t data,
  1000. uint32_t shift, uint32_t mask)
  1001. {
  1002. uint32_t tmp;
  1003. tmp = readl(addr);
  1004. tmp &= ~mask;
  1005. tmp |= (data << shift) & mask;
  1006. writel(tmp, addr);
  1007. }
  1008. static void saif_assert_rst(volatile void __iomem *addr,
  1009. const volatile void __iomem *addr_status, uint32_t mask)
  1010. {
  1011. uint32_t tmp;
  1012. tmp = readl(addr);
  1013. tmp |= mask;
  1014. writel(tmp, addr);
  1015. do {
  1016. tmp = readl(addr_status);
  1017. } while ((tmp & mask) != 0);
  1018. }
  1019. static void saif_clear_rst(volatile void __iomem *addr,
  1020. const volatile void __iomem *addr_status, uint32_t mask)
  1021. {
  1022. uint32_t tmp;
  1023. tmp = readl(addr);
  1024. tmp &= ~mask;
  1025. writel(tmp, addr);
  1026. do {
  1027. tmp = readl(addr_status);
  1028. } while ((tmp & mask) != mask);
  1029. }
  1030. static void jpu_clk_control(jpu_clk_t *clk, bool enable)
  1031. {
  1032. if (enable) {
  1033. /*enable*/
  1034. saif_set_reg(clk->apb_clk.en_ctrl, CLK_ENABLE_DATA, clk->en_shift, clk->en_mask);
  1035. saif_set_reg(clk->axi_clk.en_ctrl, CLK_ENABLE_DATA, clk->en_shift, clk->en_mask);
  1036. saif_set_reg(clk->core_clk.en_ctrl, CLK_ENABLE_DATA, clk->en_shift, clk->en_mask);
  1037. /*clr-reset*/
  1038. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->apb_clk.rst_mask);
  1039. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->axi_clk.rst_mask);
  1040. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->core_clk.rst_mask);
  1041. } else {
  1042. /*assert-reset*/
  1043. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->apb_clk.rst_mask);
  1044. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->axi_clk.rst_mask);
  1045. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->core_clk.rst_mask);
  1046. /*disable*/
  1047. saif_set_reg(clk->apb_clk.en_ctrl, CLK_DISABLE_DATA, clk->en_shift, clk->en_mask);
  1048. saif_set_reg(clk->axi_clk.en_ctrl, CLK_DISABLE_DATA, clk->en_shift, clk->en_mask);
  1049. saif_set_reg(clk->core_clk.en_ctrl, CLK_DISABLE_DATA, clk->en_shift, clk->en_mask);
  1050. }
  1051. }
  1052. static void jpu_clk_reset(jpu_clk_t *clk)
  1053. {
  1054. /*assert-reset*/
  1055. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->apb_clk.rst_mask);
  1056. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->axi_clk.rst_mask);
  1057. saif_assert_rst(clk->rst_ctrl, clk->rst_status, clk->core_clk.rst_mask);
  1058. /*clr-reset*/
  1059. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->apb_clk.rst_mask);
  1060. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->axi_clk.rst_mask);
  1061. saif_clear_rst(clk->rst_ctrl, clk->rst_status, clk->core_clk.rst_mask);
  1062. }
  1063. int jpu_hw_reset(void)
  1064. {
  1065. if (!s_jpu_clk)
  1066. return -1;
  1067. jpu_clk_reset(s_jpu_clk);
  1068. DPRINTK("[VPUDRV] reset vpu hardware. \n");
  1069. return 0;
  1070. }
  1071. static int jpu_of_clk_get(struct platform_device *pdev, jpu_clk_t *jpu_clk)
  1072. {
  1073. if (!pdev)
  1074. return -ENXIO;
  1075. jpu_clk->clkgen = ioremap(SAIF_BD_APBS_BASE, 0x400);
  1076. if (IS_ERR(jpu_clk->clkgen)) {
  1077. dev_err(&pdev->dev, "ioremap clkgen failed.\n");
  1078. return PTR_ERR(jpu_clk->clkgen);
  1079. }
  1080. /* clkgen define */
  1081. jpu_clk->axi_clk.en_ctrl = jpu_clk->clkgen + CODAJ12_CLK_AXI_CTRL;
  1082. jpu_clk->apb_clk.en_ctrl = jpu_clk->clkgen + CODAJ12_CLK_APB_CTRL;
  1083. jpu_clk->core_clk.en_ctrl = jpu_clk->clkgen + CODAJ12_CLK_CORE_CTRL;
  1084. jpu_clk->en_mask = CLK_EN_MASK;
  1085. jpu_clk->en_shift = CLK_EN_SHIFT;
  1086. /* rstgen define */
  1087. jpu_clk->rst_ctrl = jpu_clk->clkgen + RSTGEN_SOFTWARE_RESET_ASSERT1;
  1088. jpu_clk->rst_status = jpu_clk->clkgen + RSTGEN_SOFTWARE_RESET_STATUS1;
  1089. jpu_clk->axi_clk.rst_mask = RSTN_AXI_MASK;
  1090. jpu_clk->apb_clk.rst_mask = RSTN_APB_MASK;
  1091. jpu_clk->core_clk.rst_mask = RSTN_CORE_MASK;
  1092. return 0;
  1093. }
  1094. static jpu_clk_t *jpu_clk_get(struct platform_device *pdev)
  1095. {
  1096. jpu_clk_t *jpu_clk;
  1097. jpu_clk = devm_kzalloc(&pdev->dev, sizeof(*jpu_clk), GFP_KERNEL);
  1098. if (!jpu_clk)
  1099. return NULL;
  1100. if (jpu_of_clk_get(pdev, jpu_clk))
  1101. goto err_get_clk;
  1102. return jpu_clk;
  1103. err_get_clk:
  1104. devm_kfree(&pdev->dev, jpu_clk);
  1105. return NULL;
  1106. }
  1107. static void jpu_clk_put(jpu_clk_t *clk)
  1108. {
  1109. if (clk->clkgen) {
  1110. iounmap(clk->clkgen);
  1111. clk->clkgen = NULL;
  1112. }
  1113. }
  1114. static int jpu_clk_enable(jpu_clk_t *clk)
  1115. {
  1116. if (clk == NULL || IS_ERR(clk))
  1117. return -1;
  1118. jpu_pmu_enable(clk->dev);
  1119. jpu_clk_control(clk, true);
  1120. DPRINTK("[VPUDRV] vpu_clk_enable\n");
  1121. return 0;
  1122. }
  1123. static void jpu_clk_disable(jpu_clk_t *clk)
  1124. {
  1125. if (clk == NULL || IS_ERR(clk))
  1126. return;
  1127. jpu_clk_control(clk, false);
  1128. jpu_pmu_disable(clk->dev);
  1129. DPRINTK("[VPUDRV] vpu_clk_disable\n");
  1130. }
  1131. #else /* STARFIVE_JPU_SUPPORT_CLOCK_CONTROL */
  1132. static int jpu_hw_reset(void)
  1133. {
  1134. return reset_control_reset(s_jpu_clk->resets);
  1135. }
  1136. static int jpu_of_clk_get(struct platform_device *pdev, jpu_clk_t *jpu_clk)
  1137. {
  1138. struct device *dev = &pdev->dev;
  1139. int ret;
  1140. jpu_clk->dev = dev;
  1141. jpu_clk->clks = jpu_clks;
  1142. jpu_clk->nr_clks = ARRAY_SIZE(jpu_clks);
  1143. jpu_clk->resets = devm_reset_control_array_get_shared(dev);
  1144. if (IS_ERR(jpu_clk->resets)) {
  1145. ret = PTR_ERR(jpu_clk->resets);
  1146. dev_err(dev, "faied to get jpu reset controls\n");
  1147. }
  1148. ret = devm_clk_bulk_get(dev, jpu_clk->nr_clks, jpu_clk->clks);
  1149. if (ret)
  1150. dev_err(dev, "faied to get jpu clk controls\n");
  1151. return 0;
  1152. }
  1153. static jpu_clk_t *jpu_clk_get(struct platform_device *pdev)
  1154. {
  1155. jpu_clk_t *jpu_clk;
  1156. if (!pdev)
  1157. return NULL;
  1158. jpu_clk = devm_kzalloc(&pdev->dev, sizeof(*jpu_clk), GFP_KERNEL);
  1159. if (!jpu_clk)
  1160. return NULL;
  1161. if (jpu_of_clk_get(pdev, jpu_clk))
  1162. goto err_of_clk_get;
  1163. return jpu_clk;
  1164. err_of_clk_get:
  1165. devm_kfree(&pdev->dev, jpu_clk);
  1166. return NULL;
  1167. }
  1168. static void jpu_clk_put(jpu_clk_t *clk)
  1169. {
  1170. clk_bulk_put(clk->nr_clks, clk->clks);
  1171. }
  1172. static int jpu_clk_enable(jpu_clk_t *clk)
  1173. {
  1174. int ret;
  1175. ret = clk_bulk_prepare_enable(clk->nr_clks, clk->clks);
  1176. if (ret)
  1177. dev_err(clk->dev, "enable clk error.\n");
  1178. ret = reset_control_deassert(clk->resets);
  1179. if (ret)
  1180. dev_err(clk->dev, "deassert jpu error.\n");
  1181. DPRINTK("[VPUDRV] jpu_clk_enable\n");
  1182. return ret;
  1183. }
  1184. static void jpu_clk_disable(jpu_clk_t *clk)
  1185. {
  1186. int ret;
  1187. ret = reset_control_assert(clk->resets);
  1188. if (ret)
  1189. dev_err(clk->dev, "assert jpu error.\n");
  1190. clk_bulk_disable_unprepare(clk->nr_clks, clk->clks);
  1191. }
  1192. #endif /* STARFIVE_JPU_SUPPORT_CLOCK_CONTROL */