v1.c 8.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351
  1. /*
  2. * Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
  3. *
  4. * Permission is hereby granted, free of charge, to any person obtaining a
  5. * copy of this software and associated documentation files (the "Software"),
  6. * to deal in the Software without restriction, including without limitation
  7. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  8. * and/or sell copies of the Software, and to permit persons to whom the
  9. * Software is furnished to do so, subject to the following conditions:
  10. *
  11. * The above copyright notice and this permission notice shall be included in
  12. * all copies or substantial portions of the Software.
  13. *
  14. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  17. * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  18. * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
  19. * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
  20. * DEALINGS IN THE SOFTWARE.
  21. */
  22. #include "priv.h"
  23. #include <core/gpuobj.h>
  24. #include <core/memory.h>
  25. #include <subdev/timer.h>
  26. static void
  27. nvkm_falcon_v1_load_imem(struct nvkm_falcon *falcon, void *data, u32 start,
  28. u32 size, u16 tag, u8 port, bool secure)
  29. {
  30. u8 rem = size % 4;
  31. u32 reg;
  32. int i;
  33. size -= rem;
  34. reg = start | BIT(24) | (secure ? BIT(28) : 0);
  35. nvkm_falcon_wr32(falcon, 0x180 + (port * 16), reg);
  36. for (i = 0; i < size / 4; i++) {
  37. /* write new tag every 256B */
  38. if ((i & 0x3f) == 0)
  39. nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
  40. nvkm_falcon_wr32(falcon, 0x184 + (port * 16), ((u32 *)data)[i]);
  41. }
  42. /*
  43. * If size is not a multiple of 4, mask the last work to ensure garbage
  44. * does not get written
  45. */
  46. if (rem) {
  47. u32 extra = ((u32 *)data)[i];
  48. /* write new tag every 256B */
  49. if ((i & 0x3f) == 0)
  50. nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
  51. nvkm_falcon_wr32(falcon, 0x184 + (port * 16),
  52. extra & (BIT(rem * 8) - 1));
  53. ++i;
  54. }
  55. /* code must be padded to 0x40 words */
  56. for (; i & 0x3f; i++)
  57. nvkm_falcon_wr32(falcon, 0x184 + (port * 16), 0);
  58. }
  59. static void
  60. nvkm_falcon_v1_load_emem(struct nvkm_falcon *falcon, void *data, u32 start,
  61. u32 size, u8 port)
  62. {
  63. u8 rem = size % 4;
  64. int i;
  65. size -= rem;
  66. nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 24));
  67. for (i = 0; i < size / 4; i++)
  68. nvkm_falcon_wr32(falcon, 0xac4 + (port * 8), ((u32 *)data)[i]);
  69. /*
  70. * If size is not a multiple of 4, mask the last word to ensure garbage
  71. * does not get written
  72. */
  73. if (rem) {
  74. u32 extra = ((u32 *)data)[i];
  75. nvkm_falcon_wr32(falcon, 0xac4 + (port * 8),
  76. extra & (BIT(rem * 8) - 1));
  77. }
  78. }
  79. static const u32 EMEM_START_ADDR = 0x1000000;
  80. static void
  81. nvkm_falcon_v1_load_dmem(struct nvkm_falcon *falcon, void *data, u32 start,
  82. u32 size, u8 port)
  83. {
  84. u8 rem = size % 4;
  85. int i;
  86. if (start >= EMEM_START_ADDR && falcon->has_emem)
  87. return nvkm_falcon_v1_load_emem(falcon, data,
  88. start - EMEM_START_ADDR, size,
  89. port);
  90. size -= rem;
  91. nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 24));
  92. for (i = 0; i < size / 4; i++)
  93. nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8), ((u32 *)data)[i]);
  94. /*
  95. * If size is not a multiple of 4, mask the last word to ensure garbage
  96. * does not get written
  97. */
  98. if (rem) {
  99. u32 extra = ((u32 *)data)[i];
  100. nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8),
  101. extra & (BIT(rem * 8) - 1));
  102. }
  103. }
  104. static void
  105. nvkm_falcon_v1_read_emem(struct nvkm_falcon *falcon, u32 start, u32 size,
  106. u8 port, void *data)
  107. {
  108. u8 rem = size % 4;
  109. int i;
  110. size -= rem;
  111. nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 25));
  112. for (i = 0; i < size / 4; i++)
  113. ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
  114. /*
  115. * If size is not a multiple of 4, mask the last word to ensure garbage
  116. * does not get read
  117. */
  118. if (rem) {
  119. u32 extra = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
  120. for (i = size; i < size + rem; i++) {
  121. ((u8 *)data)[i] = (u8)(extra & 0xff);
  122. extra >>= 8;
  123. }
  124. }
  125. }
  126. static void
  127. nvkm_falcon_v1_read_dmem(struct nvkm_falcon *falcon, u32 start, u32 size,
  128. u8 port, void *data)
  129. {
  130. u8 rem = size % 4;
  131. int i;
  132. if (start >= EMEM_START_ADDR && falcon->has_emem)
  133. return nvkm_falcon_v1_read_emem(falcon, start - EMEM_START_ADDR,
  134. size, port, data);
  135. size -= rem;
  136. nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 25));
  137. for (i = 0; i < size / 4; i++)
  138. ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
  139. /*
  140. * If size is not a multiple of 4, mask the last word to ensure garbage
  141. * does not get read
  142. */
  143. if (rem) {
  144. u32 extra = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
  145. for (i = size; i < size + rem; i++) {
  146. ((u8 *)data)[i] = (u8)(extra & 0xff);
  147. extra >>= 8;
  148. }
  149. }
  150. }
  151. static void
  152. nvkm_falcon_v1_bind_context(struct nvkm_falcon *falcon, struct nvkm_memory *ctx)
  153. {
  154. u32 inst_loc;
  155. u32 fbif;
  156. /* disable instance block binding */
  157. if (ctx == NULL) {
  158. nvkm_falcon_wr32(falcon, 0x10c, 0x0);
  159. return;
  160. }
  161. switch (falcon->owner->index) {
  162. case NVKM_ENGINE_NVENC0:
  163. case NVKM_ENGINE_NVENC1:
  164. case NVKM_ENGINE_NVENC2:
  165. fbif = 0x800;
  166. break;
  167. case NVKM_SUBDEV_PMU:
  168. fbif = 0xe00;
  169. break;
  170. default:
  171. fbif = 0x600;
  172. break;
  173. }
  174. nvkm_falcon_wr32(falcon, 0x10c, 0x1);
  175. /* setup apertures - virtual */
  176. nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_UCODE, 0x4);
  177. nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_VIRT, 0x0);
  178. /* setup apertures - physical */
  179. nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_VID, 0x4);
  180. nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_COH, 0x5);
  181. nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_NCOH, 0x6);
  182. /* Set context */
  183. switch (nvkm_memory_target(ctx)) {
  184. case NVKM_MEM_TARGET_VRAM: inst_loc = 0; break;
  185. case NVKM_MEM_TARGET_HOST: inst_loc = 2; break;
  186. case NVKM_MEM_TARGET_NCOH: inst_loc = 3; break;
  187. default:
  188. WARN_ON(1);
  189. return;
  190. }
  191. /* Enable context */
  192. nvkm_falcon_mask(falcon, 0x048, 0x1, 0x1);
  193. nvkm_falcon_wr32(falcon, 0x054,
  194. ((nvkm_memory_addr(ctx) >> 12) & 0xfffffff) |
  195. (inst_loc << 28) | (1 << 30));
  196. nvkm_falcon_mask(falcon, 0x090, 0x10000, 0x10000);
  197. nvkm_falcon_mask(falcon, 0x0a4, 0x8, 0x8);
  198. }
  199. static void
  200. nvkm_falcon_v1_set_start_addr(struct nvkm_falcon *falcon, u32 start_addr)
  201. {
  202. nvkm_falcon_wr32(falcon, 0x104, start_addr);
  203. }
  204. static void
  205. nvkm_falcon_v1_start(struct nvkm_falcon *falcon)
  206. {
  207. u32 reg = nvkm_falcon_rd32(falcon, 0x100);
  208. if (reg & BIT(6))
  209. nvkm_falcon_wr32(falcon, 0x130, 0x2);
  210. else
  211. nvkm_falcon_wr32(falcon, 0x100, 0x2);
  212. }
  213. static int
  214. nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
  215. {
  216. struct nvkm_device *device = falcon->owner->device;
  217. int ret;
  218. ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
  219. if (ret < 0)
  220. return ret;
  221. return 0;
  222. }
  223. static int
  224. nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon *falcon, u32 mask)
  225. {
  226. struct nvkm_device *device = falcon->owner->device;
  227. int ret;
  228. /* clear interrupt(s) */
  229. nvkm_falcon_mask(falcon, 0x004, mask, mask);
  230. /* wait until interrupts are cleared */
  231. ret = nvkm_wait_msec(device, 10, falcon->addr + 0x008, mask, 0x0);
  232. if (ret < 0)
  233. return ret;
  234. return 0;
  235. }
  236. static int
  237. falcon_v1_wait_idle(struct nvkm_falcon *falcon)
  238. {
  239. struct nvkm_device *device = falcon->owner->device;
  240. int ret;
  241. ret = nvkm_wait_msec(device, 10, falcon->addr + 0x04c, 0xffff, 0x0);
  242. if (ret < 0)
  243. return ret;
  244. return 0;
  245. }
  246. static int
  247. nvkm_falcon_v1_enable(struct nvkm_falcon *falcon)
  248. {
  249. struct nvkm_device *device = falcon->owner->device;
  250. int ret;
  251. ret = nvkm_wait_msec(device, 10, falcon->addr + 0x10c, 0x6, 0x0);
  252. if (ret < 0) {
  253. nvkm_error(falcon->user, "Falcon mem scrubbing timeout\n");
  254. return ret;
  255. }
  256. ret = falcon_v1_wait_idle(falcon);
  257. if (ret)
  258. return ret;
  259. /* enable IRQs */
  260. nvkm_falcon_wr32(falcon, 0x010, 0xff);
  261. return 0;
  262. }
  263. static void
  264. nvkm_falcon_v1_disable(struct nvkm_falcon *falcon)
  265. {
  266. /* disable IRQs and wait for any previous code to complete */
  267. nvkm_falcon_wr32(falcon, 0x014, 0xff);
  268. falcon_v1_wait_idle(falcon);
  269. }
  270. static const struct nvkm_falcon_func
  271. nvkm_falcon_v1 = {
  272. .load_imem = nvkm_falcon_v1_load_imem,
  273. .load_dmem = nvkm_falcon_v1_load_dmem,
  274. .read_dmem = nvkm_falcon_v1_read_dmem,
  275. .bind_context = nvkm_falcon_v1_bind_context,
  276. .start = nvkm_falcon_v1_start,
  277. .wait_for_halt = nvkm_falcon_v1_wait_for_halt,
  278. .clear_interrupt = nvkm_falcon_v1_clear_interrupt,
  279. .enable = nvkm_falcon_v1_enable,
  280. .disable = nvkm_falcon_v1_disable,
  281. .set_start_addr = nvkm_falcon_v1_set_start_addr,
  282. };
  283. int
  284. nvkm_falcon_v1_new(struct nvkm_subdev *owner, const char *name, u32 addr,
  285. struct nvkm_falcon **pfalcon)
  286. {
  287. struct nvkm_falcon *falcon;
  288. if (!(falcon = *pfalcon = kzalloc(sizeof(*falcon), GFP_KERNEL)))
  289. return -ENOMEM;
  290. nvkm_falcon_ctor(&nvkm_falcon_v1, owner, name, addr, falcon);
  291. return 0;
  292. }