]> git.karo-electronics.de Git - karo-tx-linux.git/blobdiff - drivers/gpu/drm/nouveau/nouveau_bo.c
drm/nouveau: s/mem/reg/ for struct ttm_mem_reg variables
[karo-tx-linux.git] / drivers / gpu / drm / nouveau / nouveau_bo.c
index 1b8d61b591c2daea7bf17a33f093c7574916e14b..bbcfd278eacb99c8707c39e5cbf8d8cc46be3191 100644 (file)
@@ -655,20 +655,20 @@ nve0_bo_move_init(struct nouveau_channel *chan, u32 handle)
 
 static int
 nve0_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
+       struct nvkm_mem *mem = old_reg->mm_node;
        int ret = RING_SPACE(chan, 10);
        if (ret == 0) {
                BEGIN_NVC0(chan, NvSubCopy, 0x0400, 8);
-               OUT_RING  (chan, upper_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, upper_32_bits(node->vma[1].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[1].offset));
+               OUT_RING  (chan, upper_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, upper_32_bits(mem->vma[1].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[1].offset));
                OUT_RING  (chan, PAGE_SIZE);
                OUT_RING  (chan, PAGE_SIZE);
                OUT_RING  (chan, PAGE_SIZE);
-               OUT_RING  (chan, new_mem->num_pages);
+               OUT_RING  (chan, new_reg->num_pages);
                BEGIN_IMC0(chan, NvSubCopy, 0x0300, 0x0386);
        }
        return ret;
@@ -687,15 +687,15 @@ nvc0_bo_move_init(struct nouveau_channel *chan, u32 handle)
 
 static int
 nvc0_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
-       u64 src_offset = node->vma[0].offset;
-       u64 dst_offset = node->vma[1].offset;
-       u32 page_count = new_mem->num_pages;
+       struct nvkm_mem *mem = old_reg->mm_node;
+       u64 src_offset = mem->vma[0].offset;
+       u64 dst_offset = mem->vma[1].offset;
+       u32 page_count = new_reg->num_pages;
        int ret;
 
-       page_count = new_mem->num_pages;
+       page_count = new_reg->num_pages;
        while (page_count) {
                int line_count = (page_count > 8191) ? 8191 : page_count;
 
@@ -725,15 +725,15 @@ nvc0_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
 
 static int
 nvc0_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
-       u64 src_offset = node->vma[0].offset;
-       u64 dst_offset = node->vma[1].offset;
-       u32 page_count = new_mem->num_pages;
+       struct nvkm_mem *mem = old_reg->mm_node;
+       u64 src_offset = mem->vma[0].offset;
+       u64 dst_offset = mem->vma[1].offset;
+       u32 page_count = new_reg->num_pages;
        int ret;
 
-       page_count = new_mem->num_pages;
+       page_count = new_reg->num_pages;
        while (page_count) {
                int line_count = (page_count > 2047) ? 2047 : page_count;
 
@@ -764,15 +764,15 @@ nvc0_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
 
 static int
 nva3_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
-       u64 src_offset = node->vma[0].offset;
-       u64 dst_offset = node->vma[1].offset;
-       u32 page_count = new_mem->num_pages;
+       struct nvkm_mem *mem = old_reg->mm_node;
+       u64 src_offset = mem->vma[0].offset;
+       u64 dst_offset = mem->vma[1].offset;
+       u32 page_count = new_reg->num_pages;
        int ret;
 
-       page_count = new_mem->num_pages;
+       page_count = new_reg->num_pages;
        while (page_count) {
                int line_count = (page_count > 8191) ? 8191 : page_count;
 
@@ -802,35 +802,35 @@ nva3_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
 
 static int
 nv98_bo_move_exec(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
+       struct nvkm_mem *mem = old_reg->mm_node;
        int ret = RING_SPACE(chan, 7);
        if (ret == 0) {
                BEGIN_NV04(chan, NvSubCopy, 0x0320, 6);
-               OUT_RING  (chan, upper_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, upper_32_bits(node->vma[1].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[1].offset));
+               OUT_RING  (chan, upper_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, upper_32_bits(mem->vma[1].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[1].offset));
                OUT_RING  (chan, 0x00000000 /* COPY */);
-               OUT_RING  (chan, new_mem->num_pages << PAGE_SHIFT);
+               OUT_RING  (chan, new_reg->num_pages << PAGE_SHIFT);
        }
        return ret;
 }
 
 static int
 nv84_bo_move_exec(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
+       struct nvkm_mem *mem = old_reg->mm_node;
        int ret = RING_SPACE(chan, 7);
        if (ret == 0) {
                BEGIN_NV04(chan, NvSubCopy, 0x0304, 6);
-               OUT_RING  (chan, new_mem->num_pages << PAGE_SHIFT);
-               OUT_RING  (chan, upper_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[0].offset));
-               OUT_RING  (chan, upper_32_bits(node->vma[1].offset));
-               OUT_RING  (chan, lower_32_bits(node->vma[1].offset));
+               OUT_RING  (chan, new_reg->num_pages << PAGE_SHIFT);
+               OUT_RING  (chan, upper_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[0].offset));
+               OUT_RING  (chan, upper_32_bits(mem->vma[1].offset));
+               OUT_RING  (chan, lower_32_bits(mem->vma[1].offset));
                OUT_RING  (chan, 0x00000000 /* MODE_COPY, QUERY_NONE */);
        }
        return ret;
@@ -854,14 +854,14 @@ nv50_bo_move_init(struct nouveau_channel *chan, u32 handle)
 
 static int
 nv50_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       struct nvkm_mem *node = old_mem->mm_node;
-       u64 length = (new_mem->num_pages << PAGE_SHIFT);
-       u64 src_offset = node->vma[0].offset;
-       u64 dst_offset = node->vma[1].offset;
-       int src_tiled = !!node->memtype;
-       int dst_tiled = !!((struct nvkm_mem *)new_mem->mm_node)->memtype;
+       struct nvkm_mem *mem = old_reg->mm_node;
+       u64 length = (new_reg->num_pages << PAGE_SHIFT);
+       u64 src_offset = mem->vma[0].offset;
+       u64 dst_offset = mem->vma[1].offset;
+       int src_tiled = !!mem->memtype;
+       int dst_tiled = !!((struct nvkm_mem *)new_reg->mm_node)->memtype;
        int ret;
 
        while (length) {
@@ -941,20 +941,20 @@ nv04_bo_move_init(struct nouveau_channel *chan, u32 handle)
 
 static inline uint32_t
 nouveau_bo_mem_ctxdma(struct ttm_buffer_object *bo,
-                     struct nouveau_channel *chan, struct ttm_mem_reg *mem)
+                     struct nouveau_channel *chan, struct ttm_mem_reg *reg)
 {
-       if (mem->mem_type == TTM_PL_TT)
+       if (reg->mem_type == TTM_PL_TT)
                return NvDmaTT;
        return chan->vram.handle;
 }
 
 static int
 nv04_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
-                 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem)
+                 struct ttm_mem_reg *old_reg, struct ttm_mem_reg *new_reg)
 {
-       u32 src_offset = old_mem->start << PAGE_SHIFT;
-       u32 dst_offset = new_mem->start << PAGE_SHIFT;
-       u32 page_count = new_mem->num_pages;
+       u32 src_offset = old_reg->start << PAGE_SHIFT;
+       u32 dst_offset = new_reg->start << PAGE_SHIFT;
+       u32 page_count = new_reg->num_pages;
        int ret;
 
        ret = RING_SPACE(chan, 3);
@@ -962,10 +962,10 @@ nv04_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
                return ret;
 
        BEGIN_NV04(chan, NvSubCopy, NV_MEMORY_TO_MEMORY_FORMAT_DMA_SOURCE, 2);
-       OUT_RING  (chan, nouveau_bo_mem_ctxdma(bo, chan, old_mem));
-       OUT_RING  (chan, nouveau_bo_mem_ctxdma(bo, chan, new_mem));
+       OUT_RING  (chan, nouveau_bo_mem_ctxdma(bo, chan, old_reg));
+       OUT_RING  (chan, nouveau_bo_mem_ctxdma(bo, chan, new_reg));
 
-       page_count = new_mem->num_pages;
+       page_count = new_reg->num_pages;
        while (page_count) {
                int line_count = (page_count > 2047) ? 2047 : page_count;
 
@@ -996,33 +996,33 @@ nv04_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo,
 
 static int
 nouveau_bo_move_prep(struct nouveau_drm *drm, struct ttm_buffer_object *bo,
-                    struct ttm_mem_reg *mem)
+                    struct ttm_mem_reg *reg)
 {
-       struct nvkm_mem *old_node = bo->mem.mm_node;
-       struct nvkm_mem *new_node = mem->mm_node;
-       u64 size = (u64)mem->num_pages << PAGE_SHIFT;
+       struct nvkm_mem *old_mem = bo->mem.mm_node;
+       struct nvkm_mem *new_mem = reg->mm_node;
+       u64 size = (u64)reg->num_pages << PAGE_SHIFT;
        int ret;
 
-       ret = nvkm_vm_get(drm->client.vm, size, old_node->page_shift,
-                         NV_MEM_ACCESS_RW, &old_node->vma[0]);
+       ret = nvkm_vm_get(drm->client.vm, size, old_mem->page_shift,
+                         NV_MEM_ACCESS_RW, &old_mem->vma[0]);
        if (ret)
                return ret;
 
-       ret = nvkm_vm_get(drm->client.vm, size, new_node->page_shift,
-                         NV_MEM_ACCESS_RW, &old_node->vma[1]);
+       ret = nvkm_vm_get(drm->client.vm, size, new_mem->page_shift,
+                         NV_MEM_ACCESS_RW, &old_mem->vma[1]);
        if (ret) {
-               nvkm_vm_put(&old_node->vma[0]);
+               nvkm_vm_put(&old_mem->vma[0]);
                return ret;
        }
 
-       nvkm_vm_map(&old_node->vma[0], old_node);
-       nvkm_vm_map(&old_node->vma[1], new_node);
+       nvkm_vm_map(&old_mem->vma[0], old_mem);
+       nvkm_vm_map(&old_mem->vma[1], new_mem);
        return 0;
 }
 
 static int
 nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
-                    bool no_wait_gpu, struct ttm_mem_reg *new_mem)
+                    bool no_wait_gpu, struct ttm_mem_reg *new_reg)
 {
        struct nouveau_drm *drm = nouveau_bdev(bo->bdev);
        struct nouveau_channel *chan = drm->ttm.chan;
@@ -1035,7 +1035,7 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
         * destroyed the ttm_mem_reg
         */
        if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) {
-               ret = nouveau_bo_move_prep(drm, bo, new_mem);
+               ret = nouveau_bo_move_prep(drm, bo, new_reg);
                if (ret)
                        return ret;
        }
@@ -1043,14 +1043,14 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
        mutex_lock_nested(&cli->mutex, SINGLE_DEPTH_NESTING);
        ret = nouveau_fence_sync(nouveau_bo(bo), chan, true, intr);
        if (ret == 0) {
-               ret = drm->ttm.move(chan, bo, &bo->mem, new_mem);
+               ret = drm->ttm.move(chan, bo, &bo->mem, new_reg);
                if (ret == 0) {
                        ret = nouveau_fence_new(chan, false, &fence);
                        if (ret == 0) {
                                ret = ttm_bo_move_accel_cleanup(bo,
                                                                &fence->base,
                                                                evict,
-                                                               new_mem);
+                                                               new_reg);
                                nouveau_fence_unref(&fence);
                        }
                }
@@ -1125,7 +1125,7 @@ nouveau_bo_move_init(struct nouveau_drm *drm)
 
 static int
 nouveau_bo_move_flipd(struct ttm_buffer_object *bo, bool evict, bool intr,
-                     bool no_wait_gpu, struct ttm_mem_reg *new_mem)
+                     bool no_wait_gpu, struct ttm_mem_reg *new_reg)
 {
        struct ttm_place placement_memtype = {
                .fpfn = 0,
@@ -1133,35 +1133,35 @@ nouveau_bo_move_flipd(struct ttm_buffer_object *bo, bool evict, bool intr,
                .flags = TTM_PL_FLAG_TT | TTM_PL_MASK_CACHING
        };
        struct ttm_placement placement;
-       struct ttm_mem_reg tmp_mem;
+       struct ttm_mem_reg tmp_reg;
        int ret;
 
        placement.num_placement = placement.num_busy_placement = 1;
        placement.placement = placement.busy_placement = &placement_memtype;
 
-       tmp_mem = *new_mem;
-       tmp_mem.mm_node = NULL;
-       ret = ttm_bo_mem_space(bo, &placement, &tmp_mem, intr, no_wait_gpu);
+       tmp_reg = *new_reg;
+       tmp_reg.mm_node = NULL;
+       ret = ttm_bo_mem_space(bo, &placement, &tmp_reg, intr, no_wait_gpu);
        if (ret)
                return ret;
 
-       ret = ttm_tt_bind(bo->ttm, &tmp_mem);
+       ret = ttm_tt_bind(bo->ttm, &tmp_reg);
        if (ret)
                goto out;
 
-       ret = nouveau_bo_move_m2mf(bo, true, intr, no_wait_gpu, &tmp_mem);
+       ret = nouveau_bo_move_m2mf(bo, true, intr, no_wait_gpu, &tmp_reg);
        if (ret)
                goto out;
 
-       ret = ttm_bo_move_ttm(bo, intr, no_wait_gpu, new_mem);
+       ret = ttm_bo_move_ttm(bo, intr, no_wait_gpu, new_reg);
 out:
-       ttm_bo_mem_put(bo, &tmp_mem);
+       ttm_bo_mem_put(bo, &tmp_reg);
        return ret;
 }
 
 static int
 nouveau_bo_move_flips(struct ttm_buffer_object *bo, bool evict, bool intr,
-                     bool no_wait_gpu, struct ttm_mem_reg *new_mem)
+                     bool no_wait_gpu, struct ttm_mem_reg *new_reg)
 {
        struct ttm_place placement_memtype = {
                .fpfn = 0,
@@ -1169,34 +1169,34 @@ nouveau_bo_move_flips(struct ttm_buffer_object *bo, bool evict, bool intr,
                .flags = TTM_PL_FLAG_TT | TTM_PL_MASK_CACHING
        };
        struct ttm_placement placement;
-       struct ttm_mem_reg tmp_mem;
+       struct ttm_mem_reg tmp_reg;
        int ret;
 
        placement.num_placement = placement.num_busy_placement = 1;
        placement.placement = placement.busy_placement = &placement_memtype;
 
-       tmp_mem = *new_mem;
-       tmp_mem.mm_node = NULL;
-       ret = ttm_bo_mem_space(bo, &placement, &tmp_mem, intr, no_wait_gpu);
+       tmp_reg = *new_reg;
+       tmp_reg.mm_node = NULL;
+       ret = ttm_bo_mem_space(bo, &placement, &tmp_reg, intr, no_wait_gpu);
        if (ret)
                return ret;
 
-       ret = ttm_bo_move_ttm(bo, intr, no_wait_gpu, &tmp_mem);
+       ret = ttm_bo_move_ttm(bo, intr, no_wait_gpu, &tmp_reg);
        if (ret)
                goto out;
 
-       ret = nouveau_bo_move_m2mf(bo, true, intr, no_wait_gpu, new_mem);
+       ret = nouveau_bo_move_m2mf(bo, true, intr, no_wait_gpu, new_reg);
        if (ret)
                goto out;
 
 out:
-       ttm_bo_mem_put(bo, &tmp_mem);
+       ttm_bo_mem_put(bo, &tmp_reg);
        return ret;
 }
 
 static void
 nouveau_bo_move_ntfy(struct ttm_buffer_object *bo, bool evict,
-                    struct ttm_mem_reg *new_mem)
+                    struct ttm_mem_reg *new_reg)
 {
        struct nouveau_bo *nvbo = nouveau_bo(bo);
        struct nvkm_vma *vma;
@@ -1206,10 +1206,10 @@ nouveau_bo_move_ntfy(struct ttm_buffer_object *bo, bool evict,
                return;
 
        list_for_each_entry(vma, &nvbo->vma_list, head) {
-               if (new_mem && new_mem->mem_type != TTM_PL_SYSTEM &&
-                             (new_mem->mem_type == TTM_PL_VRAM ||
+               if (new_reg && new_reg->mem_type != TTM_PL_SYSTEM &&
+                             (new_reg->mem_type == TTM_PL_VRAM ||
                               nvbo->page_shift != vma->vm->mmu->lpg_shift)) {
-                       nvkm_vm_map(vma, new_mem->mm_node);
+                       nvkm_vm_map(vma, new_reg->mm_node);
                } else {
                        WARN_ON(ttm_bo_wait(bo, false, false));
                        nvkm_vm_unmap(vma);
@@ -1218,20 +1218,20 @@ nouveau_bo_move_ntfy(struct ttm_buffer_object *bo, bool evict,
 }
 
 static int
-nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem,
+nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_reg,
                   struct nouveau_drm_tile **new_tile)
 {
        struct nouveau_drm *drm = nouveau_bdev(bo->bdev);
        struct drm_device *dev = drm->dev;
        struct nouveau_bo *nvbo = nouveau_bo(bo);
-       u64 offset = new_mem->start << PAGE_SHIFT;
+       u64 offset = new_reg->start << PAGE_SHIFT;
 
        *new_tile = NULL;
-       if (new_mem->mem_type != TTM_PL_VRAM)
+       if (new_reg->mem_type != TTM_PL_VRAM)
                return 0;
 
        if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_CELSIUS) {
-               *new_tile = nv10_bo_set_tiling(dev, offset, new_mem->size,
+               *new_tile = nv10_bo_set_tiling(dev, offset, new_reg->size,
                                                nvbo->tile_mode,
                                                nvbo->tile_flags);
        }
@@ -1254,11 +1254,11 @@ nouveau_bo_vm_cleanup(struct ttm_buffer_object *bo,
 
 static int
 nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, bool intr,
-               bool no_wait_gpu, struct ttm_mem_reg *new_mem)
+               bool no_wait_gpu, struct ttm_mem_reg *new_reg)
 {
        struct nouveau_drm *drm = nouveau_bdev(bo->bdev);
        struct nouveau_bo *nvbo = nouveau_bo(bo);
-       struct ttm_mem_reg *old_mem = &bo->mem;
+       struct ttm_mem_reg *old_reg = &bo->mem;
        struct nouveau_drm_tile *new_tile = NULL;
        int ret = 0;
 
@@ -1270,30 +1270,30 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, bool intr,
                NV_WARN(drm, "Moving pinned object %p!\n", nvbo);
 
        if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
-               ret = nouveau_bo_vm_bind(bo, new_mem, &new_tile);
+               ret = nouveau_bo_vm_bind(bo, new_reg, &new_tile);
                if (ret)
                        return ret;
        }
 
        /* Fake bo copy. */
-       if (old_mem->mem_type == TTM_PL_SYSTEM && !bo->ttm) {
+       if (old_reg->mem_type == TTM_PL_SYSTEM && !bo->ttm) {
                BUG_ON(bo->mem.mm_node != NULL);
-               bo->mem = *new_mem;
-               new_mem->mm_node = NULL;
+               bo->mem = *new_reg;
+               new_reg->mm_node = NULL;
                goto out;
        }
 
        /* Hardware assisted copy. */
        if (drm->ttm.move) {
-               if (new_mem->mem_type == TTM_PL_SYSTEM)
+               if (new_reg->mem_type == TTM_PL_SYSTEM)
                        ret = nouveau_bo_move_flipd(bo, evict, intr,
-                                                   no_wait_gpu, new_mem);
-               else if (old_mem->mem_type == TTM_PL_SYSTEM)
+                                                   no_wait_gpu, new_reg);
+               else if (old_reg->mem_type == TTM_PL_SYSTEM)
                        ret = nouveau_bo_move_flips(bo, evict, intr,
-                                                   no_wait_gpu, new_mem);
+                                                   no_wait_gpu, new_reg);
                else
                        ret = nouveau_bo_move_m2mf(bo, evict, intr,
-                                                  no_wait_gpu, new_mem);
+                                                  no_wait_gpu, new_reg);
                if (!ret)
                        goto out;
        }
@@ -1301,7 +1301,7 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, bool intr,
        /* Fallback to software copy. */
        ret = ttm_bo_wait(bo, intr, no_wait_gpu);
        if (ret == 0)
-               ret = ttm_bo_move_memcpy(bo, intr, no_wait_gpu, new_mem);
+               ret = ttm_bo_move_memcpy(bo, intr, no_wait_gpu, new_reg);
 
 out:
        if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
@@ -1324,54 +1324,54 @@ nouveau_bo_verify_access(struct ttm_buffer_object *bo, struct file *filp)
 }
 
 static int
-nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
+nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *reg)
 {
-       struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
+       struct ttm_mem_type_manager *man = &bdev->man[reg->mem_type];
        struct nouveau_drm *drm = nouveau_bdev(bdev);
        struct nvkm_device *device = nvxx_device(&drm->client.device);
-       struct nvkm_mem *node = mem->mm_node;
+       struct nvkm_mem *mem = reg->mm_node;
        int ret;
 
-       mem->bus.addr = NULL;
-       mem->bus.offset = 0;
-       mem->bus.size = mem->num_pages << PAGE_SHIFT;
-       mem->bus.base = 0;
-       mem->bus.is_iomem = false;
+       reg->bus.addr = NULL;
+       reg->bus.offset = 0;
+       reg->bus.size = reg->num_pages << PAGE_SHIFT;
+       reg->bus.base = 0;
+       reg->bus.is_iomem = false;
        if (!(man->flags & TTM_MEMTYPE_FLAG_MAPPABLE))
                return -EINVAL;
-       switch (mem->mem_type) {
+       switch (reg->mem_type) {
        case TTM_PL_SYSTEM:
                /* System memory */
                return 0;
        case TTM_PL_TT:
 #if IS_ENABLED(CONFIG_AGP)
                if (drm->agp.bridge) {
-                       mem->bus.offset = mem->start << PAGE_SHIFT;
-                       mem->bus.base = drm->agp.base;
-                       mem->bus.is_iomem = !drm->agp.cma;
+                       reg->bus.offset = reg->start << PAGE_SHIFT;
+                       reg->bus.base = drm->agp.base;
+                       reg->bus.is_iomem = !drm->agp.cma;
                }
 #endif
-               if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA || !node->memtype)
+               if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA || !mem->memtype)
                        /* untiled */
                        break;
                /* fallthrough, tiled memory */
        case TTM_PL_VRAM:
-               mem->bus.offset = mem->start << PAGE_SHIFT;
-               mem->bus.base = device->func->resource_addr(device, 1);
-               mem->bus.is_iomem = true;
+               reg->bus.offset = reg->start << PAGE_SHIFT;
+               reg->bus.base = device->func->resource_addr(device, 1);
+               reg->bus.is_iomem = true;
                if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_TESLA) {
                        struct nvkm_bar *bar = nvxx_bar(&drm->client.device);
                        int page_shift = 12;
                        if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_FERMI)
-                               page_shift = node->page_shift;
+                               page_shift = mem->page_shift;
 
-                       ret = nvkm_bar_umap(bar, node->size << 12, page_shift,
-                                           &node->bar_vma);
+                       ret = nvkm_bar_umap(bar, mem->size << 12, page_shift,
+                                           &mem->bar_vma);
                        if (ret)
                                return ret;
 
-                       nvkm_vm_map(&node->bar_vma, node);
-                       mem->bus.offset = node->bar_vma.offset;
+                       nvkm_vm_map(&mem->bar_vma, mem);
+                       reg->bus.offset = mem->bar_vma.offset;
                }
                break;
        default:
@@ -1381,15 +1381,15 @@ nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
 }
 
 static void
-nouveau_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
+nouveau_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *reg)
 {
-       struct nvkm_mem *node = mem->mm_node;
+       struct nvkm_mem *mem = reg->mm_node;
 
-       if (!node->bar_vma.node)
+       if (!mem->bar_vma.node)
                return;
 
-       nvkm_vm_unmap(&node->bar_vma);
-       nvkm_vm_put(&node->bar_vma);
+       nvkm_vm_unmap(&mem->bar_vma);
+       nvkm_vm_put(&mem->bar_vma);
 }
 
 static int