123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351 |
- /*
- * Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
- *
- * Permission is hereby granted, free of charge, to any person obtaining a
- * copy of this software and associated documentation files (the "Software"),
- * to deal in the Software without restriction, including without limitation
- * the rights to use, copy, modify, merge, publish, distribute, sublicense,
- * and/or sell copies of the Software, and to permit persons to whom the
- * Software is furnished to do so, subject to the following conditions:
- *
- * The above copyright notice and this permission notice shall be included in
- * all copies or substantial portions of the Software.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
- * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
- * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
- * DEALINGS IN THE SOFTWARE.
- */
- #include "priv.h"
- #include <core/gpuobj.h>
- #include <core/memory.h>
- #include <subdev/timer.h>
- static void
- nvkm_falcon_v1_load_imem(struct nvkm_falcon *falcon, void *data, u32 start,
- u32 size, u16 tag, u8 port, bool secure)
- {
- u8 rem = size % 4;
- u32 reg;
- int i;
- size -= rem;
- reg = start | BIT(24) | (secure ? BIT(28) : 0);
- nvkm_falcon_wr32(falcon, 0x180 + (port * 16), reg);
- for (i = 0; i < size / 4; i++) {
- /* write new tag every 256B */
- if ((i & 0x3f) == 0)
- nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
- nvkm_falcon_wr32(falcon, 0x184 + (port * 16), ((u32 *)data)[i]);
- }
- /*
- * If size is not a multiple of 4, mask the last work to ensure garbage
- * does not get written
- */
- if (rem) {
- u32 extra = ((u32 *)data)[i];
- /* write new tag every 256B */
- if ((i & 0x3f) == 0)
- nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
- nvkm_falcon_wr32(falcon, 0x184 + (port * 16),
- extra & (BIT(rem * 8) - 1));
- ++i;
- }
- /* code must be padded to 0x40 words */
- for (; i & 0x3f; i++)
- nvkm_falcon_wr32(falcon, 0x184 + (port * 16), 0);
- }
- static void
- nvkm_falcon_v1_load_emem(struct nvkm_falcon *falcon, void *data, u32 start,
- u32 size, u8 port)
- {
- u8 rem = size % 4;
- int i;
- size -= rem;
- nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 24));
- for (i = 0; i < size / 4; i++)
- nvkm_falcon_wr32(falcon, 0xac4 + (port * 8), ((u32 *)data)[i]);
- /*
- * If size is not a multiple of 4, mask the last word to ensure garbage
- * does not get written
- */
- if (rem) {
- u32 extra = ((u32 *)data)[i];
- nvkm_falcon_wr32(falcon, 0xac4 + (port * 8),
- extra & (BIT(rem * 8) - 1));
- }
- }
- static const u32 EMEM_START_ADDR = 0x1000000;
- static void
- nvkm_falcon_v1_load_dmem(struct nvkm_falcon *falcon, void *data, u32 start,
- u32 size, u8 port)
- {
- u8 rem = size % 4;
- int i;
- if (start >= EMEM_START_ADDR && falcon->has_emem)
- return nvkm_falcon_v1_load_emem(falcon, data,
- start - EMEM_START_ADDR, size,
- port);
- size -= rem;
- nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 24));
- for (i = 0; i < size / 4; i++)
- nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8), ((u32 *)data)[i]);
- /*
- * If size is not a multiple of 4, mask the last word to ensure garbage
- * does not get written
- */
- if (rem) {
- u32 extra = ((u32 *)data)[i];
- nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8),
- extra & (BIT(rem * 8) - 1));
- }
- }
- static void
- nvkm_falcon_v1_read_emem(struct nvkm_falcon *falcon, u32 start, u32 size,
- u8 port, void *data)
- {
- u8 rem = size % 4;
- int i;
- size -= rem;
- nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 25));
- for (i = 0; i < size / 4; i++)
- ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
- /*
- * If size is not a multiple of 4, mask the last word to ensure garbage
- * does not get read
- */
- if (rem) {
- u32 extra = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
- for (i = size; i < size + rem; i++) {
- ((u8 *)data)[i] = (u8)(extra & 0xff);
- extra >>= 8;
- }
- }
- }
- static void
- nvkm_falcon_v1_read_dmem(struct nvkm_falcon *falcon, u32 start, u32 size,
- u8 port, void *data)
- {
- u8 rem = size % 4;
- int i;
- if (start >= EMEM_START_ADDR && falcon->has_emem)
- return nvkm_falcon_v1_read_emem(falcon, start - EMEM_START_ADDR,
- size, port, data);
- size -= rem;
- nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 25));
- for (i = 0; i < size / 4; i++)
- ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
- /*
- * If size is not a multiple of 4, mask the last word to ensure garbage
- * does not get read
- */
- if (rem) {
- u32 extra = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
- for (i = size; i < size + rem; i++) {
- ((u8 *)data)[i] = (u8)(extra & 0xff);
- extra >>= 8;
- }
- }
- }
- static void
- nvkm_falcon_v1_bind_context(struct nvkm_falcon *falcon, struct nvkm_memory *ctx)
- {
- u32 inst_loc;
- u32 fbif;
- /* disable instance block binding */
- if (ctx == NULL) {
- nvkm_falcon_wr32(falcon, 0x10c, 0x0);
- return;
- }
- switch (falcon->owner->index) {
- case NVKM_ENGINE_NVENC0:
- case NVKM_ENGINE_NVENC1:
- case NVKM_ENGINE_NVENC2:
- fbif = 0x800;
- break;
- case NVKM_SUBDEV_PMU:
- fbif = 0xe00;
- break;
- default:
- fbif = 0x600;
- break;
- }
- nvkm_falcon_wr32(falcon, 0x10c, 0x1);
- /* setup apertures - virtual */
- nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_UCODE, 0x4);
- nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_VIRT, 0x0);
- /* setup apertures - physical */
- nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_VID, 0x4);
- nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_COH, 0x5);
- nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_NCOH, 0x6);
- /* Set context */
- switch (nvkm_memory_target(ctx)) {
- case NVKM_MEM_TARGET_VRAM: inst_loc = 0; break;
- case NVKM_MEM_TARGET_HOST: inst_loc = 2; break;
- case NVKM_MEM_TARGET_NCOH: inst_loc = 3; break;
- default:
- WARN_ON(1);
- return;
- }
- /* Enable context */
- nvkm_falcon_mask(falcon, 0x048, 0x1, 0x1);
- nvkm_falcon_wr32(falcon, 0x054,
- ((nvkm_memory_addr(ctx) >> 12) & 0xfffffff) |
- (inst_loc << 28) | (1 << 30));
- nvkm_falcon_mask(falcon, 0x090, 0x10000, 0x10000);
- nvkm_falcon_mask(falcon, 0x0a4, 0x8, 0x8);
- }
- static void
- nvkm_falcon_v1_set_start_addr(struct nvkm_falcon *falcon, u32 start_addr)
- {
- nvkm_falcon_wr32(falcon, 0x104, start_addr);
- }
- static void
- nvkm_falcon_v1_start(struct nvkm_falcon *falcon)
- {
- u32 reg = nvkm_falcon_rd32(falcon, 0x100);
- if (reg & BIT(6))
- nvkm_falcon_wr32(falcon, 0x130, 0x2);
- else
- nvkm_falcon_wr32(falcon, 0x100, 0x2);
- }
- static int
- nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
- {
- struct nvkm_device *device = falcon->owner->device;
- int ret;
- ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
- if (ret < 0)
- return ret;
- return 0;
- }
- static int
- nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon *falcon, u32 mask)
- {
- struct nvkm_device *device = falcon->owner->device;
- int ret;
- /* clear interrupt(s) */
- nvkm_falcon_mask(falcon, 0x004, mask, mask);
- /* wait until interrupts are cleared */
- ret = nvkm_wait_msec(device, 10, falcon->addr + 0x008, mask, 0x0);
- if (ret < 0)
- return ret;
- return 0;
- }
- static int
- falcon_v1_wait_idle(struct nvkm_falcon *falcon)
- {
- struct nvkm_device *device = falcon->owner->device;
- int ret;
- ret = nvkm_wait_msec(device, 10, falcon->addr + 0x04c, 0xffff, 0x0);
- if (ret < 0)
- return ret;
- return 0;
- }
- static int
- nvkm_falcon_v1_enable(struct nvkm_falcon *falcon)
- {
- struct nvkm_device *device = falcon->owner->device;
- int ret;
- ret = nvkm_wait_msec(device, 10, falcon->addr + 0x10c, 0x6, 0x0);
- if (ret < 0) {
- nvkm_error(falcon->user, "Falcon mem scrubbing timeout\n");
- return ret;
- }
- ret = falcon_v1_wait_idle(falcon);
- if (ret)
- return ret;
- /* enable IRQs */
- nvkm_falcon_wr32(falcon, 0x010, 0xff);
- return 0;
- }
- static void
- nvkm_falcon_v1_disable(struct nvkm_falcon *falcon)
- {
- /* disable IRQs and wait for any previous code to complete */
- nvkm_falcon_wr32(falcon, 0x014, 0xff);
- falcon_v1_wait_idle(falcon);
- }
- static const struct nvkm_falcon_func
- nvkm_falcon_v1 = {
- .load_imem = nvkm_falcon_v1_load_imem,
- .load_dmem = nvkm_falcon_v1_load_dmem,
- .read_dmem = nvkm_falcon_v1_read_dmem,
- .bind_context = nvkm_falcon_v1_bind_context,
- .start = nvkm_falcon_v1_start,
- .wait_for_halt = nvkm_falcon_v1_wait_for_halt,
- .clear_interrupt = nvkm_falcon_v1_clear_interrupt,
- .enable = nvkm_falcon_v1_enable,
- .disable = nvkm_falcon_v1_disable,
- .set_start_addr = nvkm_falcon_v1_set_start_addr,
- };
- int
- nvkm_falcon_v1_new(struct nvkm_subdev *owner, const char *name, u32 addr,
- struct nvkm_falcon **pfalcon)
- {
- struct nvkm_falcon *falcon;
- if (!(falcon = *pfalcon = kzalloc(sizeof(*falcon), GFP_KERNEL)))
- return -ENOMEM;
- nvkm_falcon_ctor(&nvkm_falcon_v1, owner, name, addr, falcon);
- return 0;
- }
|