k3-ringacc.c 31 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * TI K3 NAVSS Ring Accelerator subsystem driver
  4. *
  5. * Copyright (C) 2019 Texas Instruments Incorporated - http://www.ti.com
  6. */
  7. #include <linux/dma-mapping.h>
  8. #include <linux/io.h>
  9. #include <linux/init.h>
  10. #include <linux/of.h>
  11. #include <linux/platform_device.h>
  12. #include <linux/sys_soc.h>
  13. #include <linux/soc/ti/k3-ringacc.h>
  14. #include <linux/soc/ti/ti_sci_protocol.h>
  15. #include <linux/soc/ti/ti_sci_inta_msi.h>
  16. #include <linux/of_irq.h>
  17. #include <linux/irqdomain.h>
  18. static LIST_HEAD(k3_ringacc_list);
  19. static DEFINE_MUTEX(k3_ringacc_list_lock);
  20. #define K3_RINGACC_CFG_RING_SIZE_ELCNT_MASK GENMASK(19, 0)
  21. /**
  22. * struct k3_ring_rt_regs - The RA realtime Control/Status Registers region
  23. *
  24. * @resv_16: Reserved
  25. * @db: Ring Doorbell Register
  26. * @resv_4: Reserved
  27. * @occ: Ring Occupancy Register
  28. * @indx: Ring Current Index Register
  29. * @hwocc: Ring Hardware Occupancy Register
  30. * @hwindx: Ring Hardware Current Index Register
  31. */
  32. struct k3_ring_rt_regs {
  33. u32 resv_16[4];
  34. u32 db;
  35. u32 resv_4[1];
  36. u32 occ;
  37. u32 indx;
  38. u32 hwocc;
  39. u32 hwindx;
  40. };
  41. #define K3_RINGACC_RT_REGS_STEP 0x1000
  42. /**
  43. * struct k3_ring_fifo_regs - The Ring Accelerator Queues Registers region
  44. *
  45. * @head_data: Ring Head Entry Data Registers
  46. * @tail_data: Ring Tail Entry Data Registers
  47. * @peek_head_data: Ring Peek Head Entry Data Regs
  48. * @peek_tail_data: Ring Peek Tail Entry Data Regs
  49. */
  50. struct k3_ring_fifo_regs {
  51. u32 head_data[128];
  52. u32 tail_data[128];
  53. u32 peek_head_data[128];
  54. u32 peek_tail_data[128];
  55. };
  56. /**
  57. * struct k3_ringacc_proxy_gcfg_regs - RA Proxy Global Config MMIO Region
  58. *
  59. * @revision: Revision Register
  60. * @config: Config Register
  61. */
  62. struct k3_ringacc_proxy_gcfg_regs {
  63. u32 revision;
  64. u32 config;
  65. };
  66. #define K3_RINGACC_PROXY_CFG_THREADS_MASK GENMASK(15, 0)
  67. /**
  68. * struct k3_ringacc_proxy_target_regs - Proxy Datapath MMIO Region
  69. *
  70. * @control: Proxy Control Register
  71. * @status: Proxy Status Register
  72. * @resv_512: Reserved
  73. * @data: Proxy Data Register
  74. */
  75. struct k3_ringacc_proxy_target_regs {
  76. u32 control;
  77. u32 status;
  78. u8 resv_512[504];
  79. u32 data[128];
  80. };
  81. #define K3_RINGACC_PROXY_TARGET_STEP 0x1000
  82. #define K3_RINGACC_PROXY_NOT_USED (-1)
  83. enum k3_ringacc_proxy_access_mode {
  84. PROXY_ACCESS_MODE_HEAD = 0,
  85. PROXY_ACCESS_MODE_TAIL = 1,
  86. PROXY_ACCESS_MODE_PEEK_HEAD = 2,
  87. PROXY_ACCESS_MODE_PEEK_TAIL = 3,
  88. };
  89. #define K3_RINGACC_FIFO_WINDOW_SIZE_BYTES (512U)
  90. #define K3_RINGACC_FIFO_REGS_STEP 0x1000
  91. #define K3_RINGACC_MAX_DB_RING_CNT (127U)
  92. struct k3_ring_ops {
  93. int (*push_tail)(struct k3_ring *ring, void *elm);
  94. int (*push_head)(struct k3_ring *ring, void *elm);
  95. int (*pop_tail)(struct k3_ring *ring, void *elm);
  96. int (*pop_head)(struct k3_ring *ring, void *elm);
  97. };
  98. /**
  99. * struct k3_ring_state - Internal state tracking structure
  100. *
  101. * @free: Number of free entries
  102. * @occ: Occupancy
  103. * @windex: Write index
  104. * @rindex: Read index
  105. */
  106. struct k3_ring_state {
  107. u32 free;
  108. u32 occ;
  109. u32 windex;
  110. u32 rindex;
  111. };
  112. /**
  113. * struct k3_ring - RA Ring descriptor
  114. *
  115. * @rt: Ring control/status registers
  116. * @fifos: Ring queues registers
  117. * @proxy: Ring Proxy Datapath registers
  118. * @ring_mem_dma: Ring buffer dma address
  119. * @ring_mem_virt: Ring buffer virt address
  120. * @ops: Ring operations
  121. * @size: Ring size in elements
  122. * @elm_size: Size of the ring element
  123. * @mode: Ring mode
  124. * @flags: flags
  125. * @ring_id: Ring Id
  126. * @parent: Pointer on struct @k3_ringacc
  127. * @use_count: Use count for shared rings
  128. * @proxy_id: RA Ring Proxy Id (only if @K3_RINGACC_RING_USE_PROXY)
  129. */
  130. struct k3_ring {
  131. struct k3_ring_rt_regs __iomem *rt;
  132. struct k3_ring_fifo_regs __iomem *fifos;
  133. struct k3_ringacc_proxy_target_regs __iomem *proxy;
  134. dma_addr_t ring_mem_dma;
  135. void *ring_mem_virt;
  136. struct k3_ring_ops *ops;
  137. u32 size;
  138. enum k3_ring_size elm_size;
  139. enum k3_ring_mode mode;
  140. u32 flags;
  141. #define K3_RING_FLAG_BUSY BIT(1)
  142. #define K3_RING_FLAG_SHARED BIT(2)
  143. struct k3_ring_state state;
  144. u32 ring_id;
  145. struct k3_ringacc *parent;
  146. u32 use_count;
  147. int proxy_id;
  148. };
  149. struct k3_ringacc_ops {
  150. int (*init)(struct platform_device *pdev, struct k3_ringacc *ringacc);
  151. };
  152. /**
  153. * struct k3_ringacc - Rings accelerator descriptor
  154. *
  155. * @dev: pointer on RA device
  156. * @proxy_gcfg: RA proxy global config registers
  157. * @proxy_target_base: RA proxy datapath region
  158. * @num_rings: number of ring in RA
  159. * @rings_inuse: bitfield for ring usage tracking
  160. * @rm_gp_range: general purpose rings range from tisci
  161. * @dma_ring_reset_quirk: DMA reset w/a enable
  162. * @num_proxies: number of RA proxies
  163. * @proxy_inuse: bitfield for proxy usage tracking
  164. * @rings: array of rings descriptors (struct @k3_ring)
  165. * @list: list of RAs in the system
  166. * @req_lock: protect rings allocation
  167. * @tisci: pointer ti-sci handle
  168. * @tisci_ring_ops: ti-sci rings ops
  169. * @tisci_dev_id: ti-sci device id
  170. * @ops: SoC specific ringacc operation
  171. */
  172. struct k3_ringacc {
  173. struct device *dev;
  174. struct k3_ringacc_proxy_gcfg_regs __iomem *proxy_gcfg;
  175. void __iomem *proxy_target_base;
  176. u32 num_rings; /* number of rings in Ringacc module */
  177. unsigned long *rings_inuse;
  178. struct ti_sci_resource *rm_gp_range;
  179. bool dma_ring_reset_quirk;
  180. u32 num_proxies;
  181. unsigned long *proxy_inuse;
  182. struct k3_ring *rings;
  183. struct list_head list;
  184. struct mutex req_lock; /* protect rings allocation */
  185. const struct ti_sci_handle *tisci;
  186. const struct ti_sci_rm_ringacc_ops *tisci_ring_ops;
  187. u32 tisci_dev_id;
  188. const struct k3_ringacc_ops *ops;
  189. };
  190. /**
  191. * struct k3_ringacc - Rings accelerator SoC data
  192. *
  193. * @dma_ring_reset_quirk: DMA reset w/a enable
  194. */
  195. struct k3_ringacc_soc_data {
  196. unsigned dma_ring_reset_quirk:1;
  197. };
  198. static long k3_ringacc_ring_get_fifo_pos(struct k3_ring *ring)
  199. {
  200. return K3_RINGACC_FIFO_WINDOW_SIZE_BYTES -
  201. (4 << ring->elm_size);
  202. }
  203. static void *k3_ringacc_get_elm_addr(struct k3_ring *ring, u32 idx)
  204. {
  205. return (ring->ring_mem_virt + idx * (4 << ring->elm_size));
  206. }
  207. static int k3_ringacc_ring_push_mem(struct k3_ring *ring, void *elem);
  208. static int k3_ringacc_ring_pop_mem(struct k3_ring *ring, void *elem);
  209. static struct k3_ring_ops k3_ring_mode_ring_ops = {
  210. .push_tail = k3_ringacc_ring_push_mem,
  211. .pop_head = k3_ringacc_ring_pop_mem,
  212. };
  213. static int k3_ringacc_ring_push_io(struct k3_ring *ring, void *elem);
  214. static int k3_ringacc_ring_pop_io(struct k3_ring *ring, void *elem);
  215. static int k3_ringacc_ring_push_head_io(struct k3_ring *ring, void *elem);
  216. static int k3_ringacc_ring_pop_tail_io(struct k3_ring *ring, void *elem);
  217. static struct k3_ring_ops k3_ring_mode_msg_ops = {
  218. .push_tail = k3_ringacc_ring_push_io,
  219. .push_head = k3_ringacc_ring_push_head_io,
  220. .pop_tail = k3_ringacc_ring_pop_tail_io,
  221. .pop_head = k3_ringacc_ring_pop_io,
  222. };
  223. static int k3_ringacc_ring_push_head_proxy(struct k3_ring *ring, void *elem);
  224. static int k3_ringacc_ring_push_tail_proxy(struct k3_ring *ring, void *elem);
  225. static int k3_ringacc_ring_pop_head_proxy(struct k3_ring *ring, void *elem);
  226. static int k3_ringacc_ring_pop_tail_proxy(struct k3_ring *ring, void *elem);
  227. static struct k3_ring_ops k3_ring_mode_proxy_ops = {
  228. .push_tail = k3_ringacc_ring_push_tail_proxy,
  229. .push_head = k3_ringacc_ring_push_head_proxy,
  230. .pop_tail = k3_ringacc_ring_pop_tail_proxy,
  231. .pop_head = k3_ringacc_ring_pop_head_proxy,
  232. };
  233. static void k3_ringacc_ring_dump(struct k3_ring *ring)
  234. {
  235. struct device *dev = ring->parent->dev;
  236. dev_dbg(dev, "dump ring: %d\n", ring->ring_id);
  237. dev_dbg(dev, "dump mem virt %p, dma %pad\n", ring->ring_mem_virt,
  238. &ring->ring_mem_dma);
  239. dev_dbg(dev, "dump elmsize %d, size %d, mode %d, proxy_id %d\n",
  240. ring->elm_size, ring->size, ring->mode, ring->proxy_id);
  241. dev_dbg(dev, "dump flags %08X\n", ring->flags);
  242. dev_dbg(dev, "dump ring_rt_regs: db%08x\n", readl(&ring->rt->db));
  243. dev_dbg(dev, "dump occ%08x\n", readl(&ring->rt->occ));
  244. dev_dbg(dev, "dump indx%08x\n", readl(&ring->rt->indx));
  245. dev_dbg(dev, "dump hwocc%08x\n", readl(&ring->rt->hwocc));
  246. dev_dbg(dev, "dump hwindx%08x\n", readl(&ring->rt->hwindx));
  247. if (ring->ring_mem_virt)
  248. print_hex_dump_debug("dump ring_mem_virt ", DUMP_PREFIX_NONE,
  249. 16, 1, ring->ring_mem_virt, 16 * 8, false);
  250. }
  251. struct k3_ring *k3_ringacc_request_ring(struct k3_ringacc *ringacc,
  252. int id, u32 flags)
  253. {
  254. int proxy_id = K3_RINGACC_PROXY_NOT_USED;
  255. mutex_lock(&ringacc->req_lock);
  256. if (id == K3_RINGACC_RING_ID_ANY) {
  257. /* Request for any general purpose ring */
  258. struct ti_sci_resource_desc *gp_rings =
  259. &ringacc->rm_gp_range->desc[0];
  260. unsigned long size;
  261. size = gp_rings->start + gp_rings->num;
  262. id = find_next_zero_bit(ringacc->rings_inuse, size,
  263. gp_rings->start);
  264. if (id == size)
  265. goto error;
  266. } else if (id < 0) {
  267. goto error;
  268. }
  269. if (test_bit(id, ringacc->rings_inuse) &&
  270. !(ringacc->rings[id].flags & K3_RING_FLAG_SHARED))
  271. goto error;
  272. else if (ringacc->rings[id].flags & K3_RING_FLAG_SHARED)
  273. goto out;
  274. if (flags & K3_RINGACC_RING_USE_PROXY) {
  275. proxy_id = find_next_zero_bit(ringacc->proxy_inuse,
  276. ringacc->num_proxies, 0);
  277. if (proxy_id == ringacc->num_proxies)
  278. goto error;
  279. }
  280. if (proxy_id != K3_RINGACC_PROXY_NOT_USED) {
  281. set_bit(proxy_id, ringacc->proxy_inuse);
  282. ringacc->rings[id].proxy_id = proxy_id;
  283. dev_dbg(ringacc->dev, "Giving ring#%d proxy#%d\n", id,
  284. proxy_id);
  285. } else {
  286. dev_dbg(ringacc->dev, "Giving ring#%d\n", id);
  287. }
  288. set_bit(id, ringacc->rings_inuse);
  289. out:
  290. ringacc->rings[id].use_count++;
  291. mutex_unlock(&ringacc->req_lock);
  292. return &ringacc->rings[id];
  293. error:
  294. mutex_unlock(&ringacc->req_lock);
  295. return NULL;
  296. }
  297. EXPORT_SYMBOL_GPL(k3_ringacc_request_ring);
  298. int k3_ringacc_request_rings_pair(struct k3_ringacc *ringacc,
  299. int fwd_id, int compl_id,
  300. struct k3_ring **fwd_ring,
  301. struct k3_ring **compl_ring)
  302. {
  303. int ret = 0;
  304. if (!fwd_ring || !compl_ring)
  305. return -EINVAL;
  306. *fwd_ring = k3_ringacc_request_ring(ringacc, fwd_id, 0);
  307. if (!(*fwd_ring))
  308. return -ENODEV;
  309. *compl_ring = k3_ringacc_request_ring(ringacc, compl_id, 0);
  310. if (!(*compl_ring)) {
  311. k3_ringacc_ring_free(*fwd_ring);
  312. ret = -ENODEV;
  313. }
  314. return ret;
  315. }
  316. EXPORT_SYMBOL_GPL(k3_ringacc_request_rings_pair);
  317. static void k3_ringacc_ring_reset_sci(struct k3_ring *ring)
  318. {
  319. struct k3_ringacc *ringacc = ring->parent;
  320. int ret;
  321. ret = ringacc->tisci_ring_ops->config(
  322. ringacc->tisci,
  323. TI_SCI_MSG_VALUE_RM_RING_COUNT_VALID,
  324. ringacc->tisci_dev_id,
  325. ring->ring_id,
  326. 0,
  327. 0,
  328. ring->size,
  329. 0,
  330. 0,
  331. 0);
  332. if (ret)
  333. dev_err(ringacc->dev, "TISCI reset ring fail (%d) ring_idx %d\n",
  334. ret, ring->ring_id);
  335. }
  336. void k3_ringacc_ring_reset(struct k3_ring *ring)
  337. {
  338. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  339. return;
  340. memset(&ring->state, 0, sizeof(ring->state));
  341. k3_ringacc_ring_reset_sci(ring);
  342. }
  343. EXPORT_SYMBOL_GPL(k3_ringacc_ring_reset);
  344. static void k3_ringacc_ring_reconfig_qmode_sci(struct k3_ring *ring,
  345. enum k3_ring_mode mode)
  346. {
  347. struct k3_ringacc *ringacc = ring->parent;
  348. int ret;
  349. ret = ringacc->tisci_ring_ops->config(
  350. ringacc->tisci,
  351. TI_SCI_MSG_VALUE_RM_RING_MODE_VALID,
  352. ringacc->tisci_dev_id,
  353. ring->ring_id,
  354. 0,
  355. 0,
  356. 0,
  357. mode,
  358. 0,
  359. 0);
  360. if (ret)
  361. dev_err(ringacc->dev, "TISCI reconf qmode fail (%d) ring_idx %d\n",
  362. ret, ring->ring_id);
  363. }
  364. void k3_ringacc_ring_reset_dma(struct k3_ring *ring, u32 occ)
  365. {
  366. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  367. return;
  368. if (!ring->parent->dma_ring_reset_quirk)
  369. goto reset;
  370. if (!occ)
  371. occ = readl(&ring->rt->occ);
  372. if (occ) {
  373. u32 db_ring_cnt, db_ring_cnt_cur;
  374. dev_dbg(ring->parent->dev, "%s %u occ: %u\n", __func__,
  375. ring->ring_id, occ);
  376. /* TI-SCI ring reset */
  377. k3_ringacc_ring_reset_sci(ring);
  378. /*
  379. * Setup the ring in ring/doorbell mode (if not already in this
  380. * mode)
  381. */
  382. if (ring->mode != K3_RINGACC_RING_MODE_RING)
  383. k3_ringacc_ring_reconfig_qmode_sci(
  384. ring, K3_RINGACC_RING_MODE_RING);
  385. /*
  386. * Ring the doorbell 2**22 – ringOcc times.
  387. * This will wrap the internal UDMAP ring state occupancy
  388. * counter (which is 21-bits wide) to 0.
  389. */
  390. db_ring_cnt = (1U << 22) - occ;
  391. while (db_ring_cnt != 0) {
  392. /*
  393. * Ring the doorbell with the maximum count each
  394. * iteration if possible to minimize the total
  395. * of writes
  396. */
  397. if (db_ring_cnt > K3_RINGACC_MAX_DB_RING_CNT)
  398. db_ring_cnt_cur = K3_RINGACC_MAX_DB_RING_CNT;
  399. else
  400. db_ring_cnt_cur = db_ring_cnt;
  401. writel(db_ring_cnt_cur, &ring->rt->db);
  402. db_ring_cnt -= db_ring_cnt_cur;
  403. }
  404. /* Restore the original ring mode (if not ring mode) */
  405. if (ring->mode != K3_RINGACC_RING_MODE_RING)
  406. k3_ringacc_ring_reconfig_qmode_sci(ring, ring->mode);
  407. }
  408. reset:
  409. /* Reset the ring */
  410. k3_ringacc_ring_reset(ring);
  411. }
  412. EXPORT_SYMBOL_GPL(k3_ringacc_ring_reset_dma);
  413. static void k3_ringacc_ring_free_sci(struct k3_ring *ring)
  414. {
  415. struct k3_ringacc *ringacc = ring->parent;
  416. int ret;
  417. ret = ringacc->tisci_ring_ops->config(
  418. ringacc->tisci,
  419. TI_SCI_MSG_VALUE_RM_ALL_NO_ORDER,
  420. ringacc->tisci_dev_id,
  421. ring->ring_id,
  422. 0,
  423. 0,
  424. 0,
  425. 0,
  426. 0,
  427. 0);
  428. if (ret)
  429. dev_err(ringacc->dev, "TISCI ring free fail (%d) ring_idx %d\n",
  430. ret, ring->ring_id);
  431. }
  432. int k3_ringacc_ring_free(struct k3_ring *ring)
  433. {
  434. struct k3_ringacc *ringacc;
  435. if (!ring)
  436. return -EINVAL;
  437. ringacc = ring->parent;
  438. dev_dbg(ring->parent->dev, "flags: 0x%08x\n", ring->flags);
  439. if (!test_bit(ring->ring_id, ringacc->rings_inuse))
  440. return -EINVAL;
  441. mutex_lock(&ringacc->req_lock);
  442. if (--ring->use_count)
  443. goto out;
  444. if (!(ring->flags & K3_RING_FLAG_BUSY))
  445. goto no_init;
  446. k3_ringacc_ring_free_sci(ring);
  447. dma_free_coherent(ringacc->dev,
  448. ring->size * (4 << ring->elm_size),
  449. ring->ring_mem_virt, ring->ring_mem_dma);
  450. ring->flags = 0;
  451. ring->ops = NULL;
  452. if (ring->proxy_id != K3_RINGACC_PROXY_NOT_USED) {
  453. clear_bit(ring->proxy_id, ringacc->proxy_inuse);
  454. ring->proxy = NULL;
  455. ring->proxy_id = K3_RINGACC_PROXY_NOT_USED;
  456. }
  457. no_init:
  458. clear_bit(ring->ring_id, ringacc->rings_inuse);
  459. out:
  460. mutex_unlock(&ringacc->req_lock);
  461. return 0;
  462. }
  463. EXPORT_SYMBOL_GPL(k3_ringacc_ring_free);
  464. u32 k3_ringacc_get_ring_id(struct k3_ring *ring)
  465. {
  466. if (!ring)
  467. return -EINVAL;
  468. return ring->ring_id;
  469. }
  470. EXPORT_SYMBOL_GPL(k3_ringacc_get_ring_id);
  471. u32 k3_ringacc_get_tisci_dev_id(struct k3_ring *ring)
  472. {
  473. if (!ring)
  474. return -EINVAL;
  475. return ring->parent->tisci_dev_id;
  476. }
  477. EXPORT_SYMBOL_GPL(k3_ringacc_get_tisci_dev_id);
  478. int k3_ringacc_get_ring_irq_num(struct k3_ring *ring)
  479. {
  480. int irq_num;
  481. if (!ring)
  482. return -EINVAL;
  483. irq_num = ti_sci_inta_msi_get_virq(ring->parent->dev, ring->ring_id);
  484. if (irq_num <= 0)
  485. irq_num = -EINVAL;
  486. return irq_num;
  487. }
  488. EXPORT_SYMBOL_GPL(k3_ringacc_get_ring_irq_num);
  489. static int k3_ringacc_ring_cfg_sci(struct k3_ring *ring)
  490. {
  491. struct k3_ringacc *ringacc = ring->parent;
  492. u32 ring_idx;
  493. int ret;
  494. if (!ringacc->tisci)
  495. return -EINVAL;
  496. ring_idx = ring->ring_id;
  497. ret = ringacc->tisci_ring_ops->config(
  498. ringacc->tisci,
  499. TI_SCI_MSG_VALUE_RM_ALL_NO_ORDER,
  500. ringacc->tisci_dev_id,
  501. ring_idx,
  502. lower_32_bits(ring->ring_mem_dma),
  503. upper_32_bits(ring->ring_mem_dma),
  504. ring->size,
  505. ring->mode,
  506. ring->elm_size,
  507. 0);
  508. if (ret)
  509. dev_err(ringacc->dev, "TISCI config ring fail (%d) ring_idx %d\n",
  510. ret, ring_idx);
  511. return ret;
  512. }
  513. int k3_ringacc_ring_cfg(struct k3_ring *ring, struct k3_ring_cfg *cfg)
  514. {
  515. struct k3_ringacc *ringacc;
  516. int ret = 0;
  517. if (!ring || !cfg)
  518. return -EINVAL;
  519. ringacc = ring->parent;
  520. if (cfg->elm_size > K3_RINGACC_RING_ELSIZE_256 ||
  521. cfg->mode >= K3_RINGACC_RING_MODE_INVALID ||
  522. cfg->size & ~K3_RINGACC_CFG_RING_SIZE_ELCNT_MASK ||
  523. !test_bit(ring->ring_id, ringacc->rings_inuse))
  524. return -EINVAL;
  525. if (cfg->mode == K3_RINGACC_RING_MODE_MESSAGE &&
  526. ring->proxy_id == K3_RINGACC_PROXY_NOT_USED &&
  527. cfg->elm_size > K3_RINGACC_RING_ELSIZE_8) {
  528. dev_err(ringacc->dev,
  529. "Message mode must use proxy for %u element size\n",
  530. 4 << ring->elm_size);
  531. return -EINVAL;
  532. }
  533. /*
  534. * In case of shared ring only the first user (master user) can
  535. * configure the ring. The sequence should be by the client:
  536. * ring = k3_ringacc_request_ring(ringacc, ring_id, 0); # master user
  537. * k3_ringacc_ring_cfg(ring, cfg); # master configuration
  538. * k3_ringacc_request_ring(ringacc, ring_id, K3_RING_FLAG_SHARED);
  539. * k3_ringacc_request_ring(ringacc, ring_id, K3_RING_FLAG_SHARED);
  540. */
  541. if (ring->use_count != 1)
  542. return 0;
  543. ring->size = cfg->size;
  544. ring->elm_size = cfg->elm_size;
  545. ring->mode = cfg->mode;
  546. memset(&ring->state, 0, sizeof(ring->state));
  547. if (ring->proxy_id != K3_RINGACC_PROXY_NOT_USED)
  548. ring->proxy = ringacc->proxy_target_base +
  549. ring->proxy_id * K3_RINGACC_PROXY_TARGET_STEP;
  550. switch (ring->mode) {
  551. case K3_RINGACC_RING_MODE_RING:
  552. ring->ops = &k3_ring_mode_ring_ops;
  553. break;
  554. case K3_RINGACC_RING_MODE_MESSAGE:
  555. if (ring->proxy)
  556. ring->ops = &k3_ring_mode_proxy_ops;
  557. else
  558. ring->ops = &k3_ring_mode_msg_ops;
  559. break;
  560. default:
  561. ring->ops = NULL;
  562. ret = -EINVAL;
  563. goto err_free_proxy;
  564. }
  565. ring->ring_mem_virt = dma_alloc_coherent(ringacc->dev,
  566. ring->size * (4 << ring->elm_size),
  567. &ring->ring_mem_dma, GFP_KERNEL);
  568. if (!ring->ring_mem_virt) {
  569. dev_err(ringacc->dev, "Failed to alloc ring mem\n");
  570. ret = -ENOMEM;
  571. goto err_free_ops;
  572. }
  573. ret = k3_ringacc_ring_cfg_sci(ring);
  574. if (ret)
  575. goto err_free_mem;
  576. ring->flags |= K3_RING_FLAG_BUSY;
  577. ring->flags |= (cfg->flags & K3_RINGACC_RING_SHARED) ?
  578. K3_RING_FLAG_SHARED : 0;
  579. k3_ringacc_ring_dump(ring);
  580. return 0;
  581. err_free_mem:
  582. dma_free_coherent(ringacc->dev,
  583. ring->size * (4 << ring->elm_size),
  584. ring->ring_mem_virt,
  585. ring->ring_mem_dma);
  586. err_free_ops:
  587. ring->ops = NULL;
  588. err_free_proxy:
  589. ring->proxy = NULL;
  590. return ret;
  591. }
  592. EXPORT_SYMBOL_GPL(k3_ringacc_ring_cfg);
  593. u32 k3_ringacc_ring_get_size(struct k3_ring *ring)
  594. {
  595. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  596. return -EINVAL;
  597. return ring->size;
  598. }
  599. EXPORT_SYMBOL_GPL(k3_ringacc_ring_get_size);
  600. u32 k3_ringacc_ring_get_free(struct k3_ring *ring)
  601. {
  602. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  603. return -EINVAL;
  604. if (!ring->state.free)
  605. ring->state.free = ring->size - readl(&ring->rt->occ);
  606. return ring->state.free;
  607. }
  608. EXPORT_SYMBOL_GPL(k3_ringacc_ring_get_free);
  609. u32 k3_ringacc_ring_get_occ(struct k3_ring *ring)
  610. {
  611. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  612. return -EINVAL;
  613. return readl(&ring->rt->occ);
  614. }
  615. EXPORT_SYMBOL_GPL(k3_ringacc_ring_get_occ);
  616. u32 k3_ringacc_ring_is_full(struct k3_ring *ring)
  617. {
  618. return !k3_ringacc_ring_get_free(ring);
  619. }
  620. EXPORT_SYMBOL_GPL(k3_ringacc_ring_is_full);
  621. enum k3_ringacc_access_mode {
  622. K3_RINGACC_ACCESS_MODE_PUSH_HEAD,
  623. K3_RINGACC_ACCESS_MODE_POP_HEAD,
  624. K3_RINGACC_ACCESS_MODE_PUSH_TAIL,
  625. K3_RINGACC_ACCESS_MODE_POP_TAIL,
  626. K3_RINGACC_ACCESS_MODE_PEEK_HEAD,
  627. K3_RINGACC_ACCESS_MODE_PEEK_TAIL,
  628. };
  629. #define K3_RINGACC_PROXY_MODE(x) (((x) & 0x3) << 16)
  630. #define K3_RINGACC_PROXY_ELSIZE(x) (((x) & 0x7) << 24)
  631. static int k3_ringacc_ring_cfg_proxy(struct k3_ring *ring,
  632. enum k3_ringacc_proxy_access_mode mode)
  633. {
  634. u32 val;
  635. val = ring->ring_id;
  636. val |= K3_RINGACC_PROXY_MODE(mode);
  637. val |= K3_RINGACC_PROXY_ELSIZE(ring->elm_size);
  638. writel(val, &ring->proxy->control);
  639. return 0;
  640. }
  641. static int k3_ringacc_ring_access_proxy(struct k3_ring *ring, void *elem,
  642. enum k3_ringacc_access_mode access_mode)
  643. {
  644. void __iomem *ptr;
  645. ptr = (void __iomem *)&ring->proxy->data;
  646. switch (access_mode) {
  647. case K3_RINGACC_ACCESS_MODE_PUSH_HEAD:
  648. case K3_RINGACC_ACCESS_MODE_POP_HEAD:
  649. k3_ringacc_ring_cfg_proxy(ring, PROXY_ACCESS_MODE_HEAD);
  650. break;
  651. case K3_RINGACC_ACCESS_MODE_PUSH_TAIL:
  652. case K3_RINGACC_ACCESS_MODE_POP_TAIL:
  653. k3_ringacc_ring_cfg_proxy(ring, PROXY_ACCESS_MODE_TAIL);
  654. break;
  655. default:
  656. return -EINVAL;
  657. }
  658. ptr += k3_ringacc_ring_get_fifo_pos(ring);
  659. switch (access_mode) {
  660. case K3_RINGACC_ACCESS_MODE_POP_HEAD:
  661. case K3_RINGACC_ACCESS_MODE_POP_TAIL:
  662. dev_dbg(ring->parent->dev,
  663. "proxy:memcpy_fromio(x): --> ptr(%p), mode:%d\n", ptr,
  664. access_mode);
  665. memcpy_fromio(elem, ptr, (4 << ring->elm_size));
  666. ring->state.occ--;
  667. break;
  668. case K3_RINGACC_ACCESS_MODE_PUSH_TAIL:
  669. case K3_RINGACC_ACCESS_MODE_PUSH_HEAD:
  670. dev_dbg(ring->parent->dev,
  671. "proxy:memcpy_toio(x): --> ptr(%p), mode:%d\n", ptr,
  672. access_mode);
  673. memcpy_toio(ptr, elem, (4 << ring->elm_size));
  674. ring->state.free--;
  675. break;
  676. default:
  677. return -EINVAL;
  678. }
  679. dev_dbg(ring->parent->dev, "proxy: free%d occ%d\n", ring->state.free,
  680. ring->state.occ);
  681. return 0;
  682. }
  683. static int k3_ringacc_ring_push_head_proxy(struct k3_ring *ring, void *elem)
  684. {
  685. return k3_ringacc_ring_access_proxy(ring, elem,
  686. K3_RINGACC_ACCESS_MODE_PUSH_HEAD);
  687. }
  688. static int k3_ringacc_ring_push_tail_proxy(struct k3_ring *ring, void *elem)
  689. {
  690. return k3_ringacc_ring_access_proxy(ring, elem,
  691. K3_RINGACC_ACCESS_MODE_PUSH_TAIL);
  692. }
  693. static int k3_ringacc_ring_pop_head_proxy(struct k3_ring *ring, void *elem)
  694. {
  695. return k3_ringacc_ring_access_proxy(ring, elem,
  696. K3_RINGACC_ACCESS_MODE_POP_HEAD);
  697. }
  698. static int k3_ringacc_ring_pop_tail_proxy(struct k3_ring *ring, void *elem)
  699. {
  700. return k3_ringacc_ring_access_proxy(ring, elem,
  701. K3_RINGACC_ACCESS_MODE_POP_HEAD);
  702. }
  703. static int k3_ringacc_ring_access_io(struct k3_ring *ring, void *elem,
  704. enum k3_ringacc_access_mode access_mode)
  705. {
  706. void __iomem *ptr;
  707. switch (access_mode) {
  708. case K3_RINGACC_ACCESS_MODE_PUSH_HEAD:
  709. case K3_RINGACC_ACCESS_MODE_POP_HEAD:
  710. ptr = (void __iomem *)&ring->fifos->head_data;
  711. break;
  712. case K3_RINGACC_ACCESS_MODE_PUSH_TAIL:
  713. case K3_RINGACC_ACCESS_MODE_POP_TAIL:
  714. ptr = (void __iomem *)&ring->fifos->tail_data;
  715. break;
  716. default:
  717. return -EINVAL;
  718. }
  719. ptr += k3_ringacc_ring_get_fifo_pos(ring);
  720. switch (access_mode) {
  721. case K3_RINGACC_ACCESS_MODE_POP_HEAD:
  722. case K3_RINGACC_ACCESS_MODE_POP_TAIL:
  723. dev_dbg(ring->parent->dev,
  724. "memcpy_fromio(x): --> ptr(%p), mode:%d\n", ptr,
  725. access_mode);
  726. memcpy_fromio(elem, ptr, (4 << ring->elm_size));
  727. ring->state.occ--;
  728. break;
  729. case K3_RINGACC_ACCESS_MODE_PUSH_TAIL:
  730. case K3_RINGACC_ACCESS_MODE_PUSH_HEAD:
  731. dev_dbg(ring->parent->dev,
  732. "memcpy_toio(x): --> ptr(%p), mode:%d\n", ptr,
  733. access_mode);
  734. memcpy_toio(ptr, elem, (4 << ring->elm_size));
  735. ring->state.free--;
  736. break;
  737. default:
  738. return -EINVAL;
  739. }
  740. dev_dbg(ring->parent->dev, "free%d index%d occ%d index%d\n",
  741. ring->state.free, ring->state.windex, ring->state.occ,
  742. ring->state.rindex);
  743. return 0;
  744. }
  745. static int k3_ringacc_ring_push_head_io(struct k3_ring *ring, void *elem)
  746. {
  747. return k3_ringacc_ring_access_io(ring, elem,
  748. K3_RINGACC_ACCESS_MODE_PUSH_HEAD);
  749. }
  750. static int k3_ringacc_ring_push_io(struct k3_ring *ring, void *elem)
  751. {
  752. return k3_ringacc_ring_access_io(ring, elem,
  753. K3_RINGACC_ACCESS_MODE_PUSH_TAIL);
  754. }
  755. static int k3_ringacc_ring_pop_io(struct k3_ring *ring, void *elem)
  756. {
  757. return k3_ringacc_ring_access_io(ring, elem,
  758. K3_RINGACC_ACCESS_MODE_POP_HEAD);
  759. }
  760. static int k3_ringacc_ring_pop_tail_io(struct k3_ring *ring, void *elem)
  761. {
  762. return k3_ringacc_ring_access_io(ring, elem,
  763. K3_RINGACC_ACCESS_MODE_POP_HEAD);
  764. }
  765. static int k3_ringacc_ring_push_mem(struct k3_ring *ring, void *elem)
  766. {
  767. void *elem_ptr;
  768. elem_ptr = k3_ringacc_get_elm_addr(ring, ring->state.windex);
  769. memcpy(elem_ptr, elem, (4 << ring->elm_size));
  770. ring->state.windex = (ring->state.windex + 1) % ring->size;
  771. ring->state.free--;
  772. writel(1, &ring->rt->db);
  773. dev_dbg(ring->parent->dev, "ring_push_mem: free%d index%d\n",
  774. ring->state.free, ring->state.windex);
  775. return 0;
  776. }
  777. static int k3_ringacc_ring_pop_mem(struct k3_ring *ring, void *elem)
  778. {
  779. void *elem_ptr;
  780. elem_ptr = k3_ringacc_get_elm_addr(ring, ring->state.rindex);
  781. memcpy(elem, elem_ptr, (4 << ring->elm_size));
  782. ring->state.rindex = (ring->state.rindex + 1) % ring->size;
  783. ring->state.occ--;
  784. writel(-1, &ring->rt->db);
  785. dev_dbg(ring->parent->dev, "ring_pop_mem: occ%d index%d pos_ptr%p\n",
  786. ring->state.occ, ring->state.rindex, elem_ptr);
  787. return 0;
  788. }
  789. int k3_ringacc_ring_push(struct k3_ring *ring, void *elem)
  790. {
  791. int ret = -EOPNOTSUPP;
  792. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  793. return -EINVAL;
  794. dev_dbg(ring->parent->dev, "ring_push: free%d index%d\n",
  795. ring->state.free, ring->state.windex);
  796. if (k3_ringacc_ring_is_full(ring))
  797. return -ENOMEM;
  798. if (ring->ops && ring->ops->push_tail)
  799. ret = ring->ops->push_tail(ring, elem);
  800. return ret;
  801. }
  802. EXPORT_SYMBOL_GPL(k3_ringacc_ring_push);
  803. int k3_ringacc_ring_push_head(struct k3_ring *ring, void *elem)
  804. {
  805. int ret = -EOPNOTSUPP;
  806. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  807. return -EINVAL;
  808. dev_dbg(ring->parent->dev, "ring_push_head: free%d index%d\n",
  809. ring->state.free, ring->state.windex);
  810. if (k3_ringacc_ring_is_full(ring))
  811. return -ENOMEM;
  812. if (ring->ops && ring->ops->push_head)
  813. ret = ring->ops->push_head(ring, elem);
  814. return ret;
  815. }
  816. EXPORT_SYMBOL_GPL(k3_ringacc_ring_push_head);
  817. int k3_ringacc_ring_pop(struct k3_ring *ring, void *elem)
  818. {
  819. int ret = -EOPNOTSUPP;
  820. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  821. return -EINVAL;
  822. if (!ring->state.occ)
  823. ring->state.occ = k3_ringacc_ring_get_occ(ring);
  824. dev_dbg(ring->parent->dev, "ring_pop: occ%d index%d\n", ring->state.occ,
  825. ring->state.rindex);
  826. if (!ring->state.occ)
  827. return -ENODATA;
  828. if (ring->ops && ring->ops->pop_head)
  829. ret = ring->ops->pop_head(ring, elem);
  830. return ret;
  831. }
  832. EXPORT_SYMBOL_GPL(k3_ringacc_ring_pop);
  833. int k3_ringacc_ring_pop_tail(struct k3_ring *ring, void *elem)
  834. {
  835. int ret = -EOPNOTSUPP;
  836. if (!ring || !(ring->flags & K3_RING_FLAG_BUSY))
  837. return -EINVAL;
  838. if (!ring->state.occ)
  839. ring->state.occ = k3_ringacc_ring_get_occ(ring);
  840. dev_dbg(ring->parent->dev, "ring_pop_tail: occ%d index%d\n",
  841. ring->state.occ, ring->state.rindex);
  842. if (!ring->state.occ)
  843. return -ENODATA;
  844. if (ring->ops && ring->ops->pop_tail)
  845. ret = ring->ops->pop_tail(ring, elem);
  846. return ret;
  847. }
  848. EXPORT_SYMBOL_GPL(k3_ringacc_ring_pop_tail);
  849. struct k3_ringacc *of_k3_ringacc_get_by_phandle(struct device_node *np,
  850. const char *property)
  851. {
  852. struct device_node *ringacc_np;
  853. struct k3_ringacc *ringacc = ERR_PTR(-EPROBE_DEFER);
  854. struct k3_ringacc *entry;
  855. ringacc_np = of_parse_phandle(np, property, 0);
  856. if (!ringacc_np)
  857. return ERR_PTR(-ENODEV);
  858. mutex_lock(&k3_ringacc_list_lock);
  859. list_for_each_entry(entry, &k3_ringacc_list, list)
  860. if (entry->dev->of_node == ringacc_np) {
  861. ringacc = entry;
  862. break;
  863. }
  864. mutex_unlock(&k3_ringacc_list_lock);
  865. of_node_put(ringacc_np);
  866. return ringacc;
  867. }
  868. EXPORT_SYMBOL_GPL(of_k3_ringacc_get_by_phandle);
  869. static int k3_ringacc_probe_dt(struct k3_ringacc *ringacc)
  870. {
  871. struct device_node *node = ringacc->dev->of_node;
  872. struct device *dev = ringacc->dev;
  873. struct platform_device *pdev = to_platform_device(dev);
  874. int ret;
  875. if (!node) {
  876. dev_err(dev, "device tree info unavailable\n");
  877. return -ENODEV;
  878. }
  879. ret = of_property_read_u32(node, "ti,num-rings", &ringacc->num_rings);
  880. if (ret) {
  881. dev_err(dev, "ti,num-rings read failure %d\n", ret);
  882. return ret;
  883. }
  884. ringacc->tisci = ti_sci_get_by_phandle(node, "ti,sci");
  885. if (IS_ERR(ringacc->tisci)) {
  886. ret = PTR_ERR(ringacc->tisci);
  887. if (ret != -EPROBE_DEFER)
  888. dev_err(dev, "ti,sci read fail %d\n", ret);
  889. ringacc->tisci = NULL;
  890. return ret;
  891. }
  892. ret = of_property_read_u32(node, "ti,sci-dev-id",
  893. &ringacc->tisci_dev_id);
  894. if (ret) {
  895. dev_err(dev, "ti,sci-dev-id read fail %d\n", ret);
  896. return ret;
  897. }
  898. pdev->id = ringacc->tisci_dev_id;
  899. ringacc->rm_gp_range = devm_ti_sci_get_of_resource(ringacc->tisci, dev,
  900. ringacc->tisci_dev_id,
  901. "ti,sci-rm-range-gp-rings");
  902. if (IS_ERR(ringacc->rm_gp_range)) {
  903. dev_err(dev, "Failed to allocate MSI interrupts\n");
  904. return PTR_ERR(ringacc->rm_gp_range);
  905. }
  906. return ti_sci_inta_msi_domain_alloc_irqs(ringacc->dev,
  907. ringacc->rm_gp_range);
  908. }
  909. static const struct k3_ringacc_soc_data k3_ringacc_soc_data_sr1 = {
  910. .dma_ring_reset_quirk = 1,
  911. };
  912. static const struct soc_device_attribute k3_ringacc_socinfo[] = {
  913. { .family = "AM65X",
  914. .revision = "SR1.0",
  915. .data = &k3_ringacc_soc_data_sr1
  916. },
  917. {/* sentinel */}
  918. };
  919. static int k3_ringacc_init(struct platform_device *pdev,
  920. struct k3_ringacc *ringacc)
  921. {
  922. const struct soc_device_attribute *soc;
  923. void __iomem *base_fifo, *base_rt;
  924. struct device *dev = &pdev->dev;
  925. struct resource *res;
  926. int ret, i;
  927. dev->msi_domain = of_msi_get_domain(dev, dev->of_node,
  928. DOMAIN_BUS_TI_SCI_INTA_MSI);
  929. if (!dev->msi_domain) {
  930. dev_err(dev, "Failed to get MSI domain\n");
  931. return -EPROBE_DEFER;
  932. }
  933. ret = k3_ringacc_probe_dt(ringacc);
  934. if (ret)
  935. return ret;
  936. soc = soc_device_match(k3_ringacc_socinfo);
  937. if (soc && soc->data) {
  938. const struct k3_ringacc_soc_data *soc_data = soc->data;
  939. ringacc->dma_ring_reset_quirk = soc_data->dma_ring_reset_quirk;
  940. }
  941. res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "rt");
  942. base_rt = devm_ioremap_resource(dev, res);
  943. if (IS_ERR(base_rt))
  944. return PTR_ERR(base_rt);
  945. res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "fifos");
  946. base_fifo = devm_ioremap_resource(dev, res);
  947. if (IS_ERR(base_fifo))
  948. return PTR_ERR(base_fifo);
  949. res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "proxy_gcfg");
  950. ringacc->proxy_gcfg = devm_ioremap_resource(dev, res);
  951. if (IS_ERR(ringacc->proxy_gcfg))
  952. return PTR_ERR(ringacc->proxy_gcfg);
  953. res = platform_get_resource_byname(pdev, IORESOURCE_MEM,
  954. "proxy_target");
  955. ringacc->proxy_target_base = devm_ioremap_resource(dev, res);
  956. if (IS_ERR(ringacc->proxy_target_base))
  957. return PTR_ERR(ringacc->proxy_target_base);
  958. ringacc->num_proxies = readl(&ringacc->proxy_gcfg->config) &
  959. K3_RINGACC_PROXY_CFG_THREADS_MASK;
  960. ringacc->rings = devm_kzalloc(dev,
  961. sizeof(*ringacc->rings) *
  962. ringacc->num_rings,
  963. GFP_KERNEL);
  964. ringacc->rings_inuse = devm_kcalloc(dev,
  965. BITS_TO_LONGS(ringacc->num_rings),
  966. sizeof(unsigned long), GFP_KERNEL);
  967. ringacc->proxy_inuse = devm_kcalloc(dev,
  968. BITS_TO_LONGS(ringacc->num_proxies),
  969. sizeof(unsigned long), GFP_KERNEL);
  970. if (!ringacc->rings || !ringacc->rings_inuse || !ringacc->proxy_inuse)
  971. return -ENOMEM;
  972. for (i = 0; i < ringacc->num_rings; i++) {
  973. ringacc->rings[i].rt = base_rt +
  974. K3_RINGACC_RT_REGS_STEP * i;
  975. ringacc->rings[i].fifos = base_fifo +
  976. K3_RINGACC_FIFO_REGS_STEP * i;
  977. ringacc->rings[i].parent = ringacc;
  978. ringacc->rings[i].ring_id = i;
  979. ringacc->rings[i].proxy_id = K3_RINGACC_PROXY_NOT_USED;
  980. }
  981. ringacc->tisci_ring_ops = &ringacc->tisci->ops.rm_ring_ops;
  982. dev_info(dev, "Ring Accelerator probed rings:%u, gp-rings[%u,%u] sci-dev-id:%u\n",
  983. ringacc->num_rings,
  984. ringacc->rm_gp_range->desc[0].start,
  985. ringacc->rm_gp_range->desc[0].num,
  986. ringacc->tisci_dev_id);
  987. dev_info(dev, "dma-ring-reset-quirk: %s\n",
  988. ringacc->dma_ring_reset_quirk ? "enabled" : "disabled");
  989. dev_info(dev, "RA Proxy rev. %08x, num_proxies:%u\n",
  990. readl(&ringacc->proxy_gcfg->revision), ringacc->num_proxies);
  991. return 0;
  992. }
  993. struct ringacc_match_data {
  994. struct k3_ringacc_ops ops;
  995. };
  996. static struct ringacc_match_data k3_ringacc_data = {
  997. .ops = {
  998. .init = k3_ringacc_init,
  999. },
  1000. };
  1001. /* Match table for of_platform binding */
  1002. static const struct of_device_id k3_ringacc_of_match[] = {
  1003. { .compatible = "ti,am654-navss-ringacc", .data = &k3_ringacc_data, },
  1004. {},
  1005. };
  1006. static int k3_ringacc_probe(struct platform_device *pdev)
  1007. {
  1008. const struct ringacc_match_data *match_data;
  1009. const struct of_device_id *match;
  1010. struct device *dev = &pdev->dev;
  1011. struct k3_ringacc *ringacc;
  1012. int ret;
  1013. match = of_match_node(k3_ringacc_of_match, dev->of_node);
  1014. if (!match)
  1015. return -ENODEV;
  1016. match_data = match->data;
  1017. ringacc = devm_kzalloc(dev, sizeof(*ringacc), GFP_KERNEL);
  1018. if (!ringacc)
  1019. return -ENOMEM;
  1020. ringacc->dev = dev;
  1021. mutex_init(&ringacc->req_lock);
  1022. ringacc->ops = &match_data->ops;
  1023. ret = ringacc->ops->init(pdev, ringacc);
  1024. if (ret)
  1025. return ret;
  1026. dev_set_drvdata(dev, ringacc);
  1027. mutex_lock(&k3_ringacc_list_lock);
  1028. list_add_tail(&ringacc->list, &k3_ringacc_list);
  1029. mutex_unlock(&k3_ringacc_list_lock);
  1030. return 0;
  1031. }
  1032. static struct platform_driver k3_ringacc_driver = {
  1033. .probe = k3_ringacc_probe,
  1034. .driver = {
  1035. .name = "k3-ringacc",
  1036. .of_match_table = k3_ringacc_of_match,
  1037. .suppress_bind_attrs = true,
  1038. },
  1039. };
  1040. builtin_platform_driver(k3_ringacc_driver);