Message ID | 5051FA51.9010207@vodafone.de (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Christian, > as I already noted in our internal discussion, the first step to > hierarchical page table support should be to cleanup the set_page interface. > Please see the attached patch, it does exactly this. I suggest that you > rebase on it and try to don't touch the chipset specific code in ni.c any > more (if you find that you need to change something there move that into > this cleanup patch instead). > Currently we alloc all PTs in one call and they lay at fixed distance from each other. Using this knowledge we can setup all PDEs of BO in one pass. I suggest the following changes to the cleanup patch: * add BLOCK_SIZE bitfield to bo_va->flags (RADEON_VM_FLAGS) * shift RADEON_GPU_PAGE_SIZE by BLOCK_SIZE obtained from flags here: void cayman_vm_set_page(struct radeon_device *rdev, uint64_t pe, uint64_t addr, unsigned count, uint32_t flags) { + unsigned block_size = RADEON_EXTRACT_BLOCKSIZE(flags) struct radeon_ring *ring = &rdev->ring[rdev->asic->vm.pt_ring_index]; uint32_t r600_flags = cayman_vm_page_flags(rdev, flags); } else if (flags & RADEON_VM_PAGE_VALID) { value = addr; - addr += RADEON_GPU_PAGE_SIZE; + addr += RADEON_GPU_PAGE_SIZE << block_size; } In future when PTt's may be scattered around due to their on-demand allocation we may devise a method to supply list of their addresses to this code (array or callback function, whatever). What do you think? > Additionally to that see the notes below: > > > On 13.09.2012 16:13, Dmitry Cherkasov wrote: >> >> PDE/PTE update code uses CP ring for memory writes. >> All page table entries are preallocated for now in alloc_pt(). >> >> It is made as whole because it's hard to divide it to several patches >> that compile and doesn't break anything being applied separately. >> >> Tested on cayman card. >> >> Signed-off-by: Dmitry Cherkasov <Dmitrii.Cherkasov@amd.com> >> --- >> I couldn't test in on SI card, so would be happy if someone could check it >> there. >> >> drivers/gpu/drm/radeon/ni.c | 64 ++++++++++++++++++------- >> drivers/gpu/drm/radeon/radeon.h | 79 >> +++++++++++++++++++++++++++--- >> drivers/gpu/drm/radeon/radeon_asic.c | 1 + >> drivers/gpu/drm/radeon/radeon_asic.h | 13 +++-- >> drivers/gpu/drm/radeon/radeon_gart.c | 88 >> ++++++++++++++++++++++++++++++---- >> drivers/gpu/drm/radeon/si.c | 4 +- >> 6 files changed, 210 insertions(+), 39 deletions(-) >> >> diff --git a/drivers/gpu/drm/radeon/ni.c b/drivers/gpu/drm/radeon/ni.c >> index b238216..31d9027 100644 >> --- a/drivers/gpu/drm/radeon/ni.c >> +++ b/drivers/gpu/drm/radeon/ni.c >> @@ -782,7 +782,7 @@ static int cayman_pcie_gart_enable(struct >> radeon_device *rdev) >> (u32)(rdev->dummy_page.addr >> 12)); >> WREG32(VM_CONTEXT1_CNTL2, 0); >> WREG32(VM_CONTEXT1_CNTL, 0); >> - WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) | >> + WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(1) | >> RANGE_PROTECTION_FAULT_ENABLE_DEFAULT); >> cayman_pcie_gart_tlb_flush(rdev); >> @@ -1517,37 +1517,65 @@ uint32_t cayman_vm_page_flags(struct radeon_device >> *rdev, uint32_t flags) >> return r600_flags; >> } >> -/** >> - * cayman_vm_set_page - update the page tables using the CP >> - * >> - * @rdev: radeon_device pointer >> - * >> - * Update the page tables using the CP (cayman-si). >> - */ > > It's probably just a rebase artifact, please never remove those > documentation lines. Instead fix them if the need arise. > > >> -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, >> - unsigned pfn, struct ttm_mem_reg *mem, >> - unsigned npages, uint32_t flags) >> +void cayman_vm_set_page(struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + uint64_t pte_gpu_addr, >> + uint64_t mem_pfn_offset, >> + unsigned count, uint32_t flags) >> { >> struct radeon_ring *ring = >> &rdev->ring[rdev->asic->vm.pt_ring_index]; >> - uint64_t addr, pt = vm->pt_gpu_addr + pfn * 8; >> + uint64_t addr; >> int i; >> addr = flags = cayman_vm_page_flags(rdev, flags); >> - radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + npages * >> 2)); >> - radeon_ring_write(ring, pt & 0xffffffff); >> - radeon_ring_write(ring, (pt >> 32) & 0xff); >> - for (i = 0; i < npages; ++i) { >> + radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + count * 2)); >> + radeon_ring_write(ring, pte_gpu_addr & 0xffffffff); >> + radeon_ring_write(ring, (pte_gpu_addr >> 32) & 0xff); >> + for (i = 0; i < count; ++i) { >> if (mem) { >> - addr = radeon_vm_get_addr(rdev, mem, i); >> + addr = radeon_vm_get_addr(rdev, mem, >> + i + mem_pfn_offset); >> addr = addr & 0xFFFFFFFFFFFFF000ULL; >> addr |= flags; >> } >> + >> + radeon_ring_write(ring, addr & 0xffffffff); >> + radeon_ring_write(ring, (addr >> 32) & 0xffffffff); >> + } >> +} >> + >> +void cayman_vm_set_pdes(struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + struct radeon_vm *vm, >> + uint64_t pde_index, >> + unsigned pde_count, uint32_t flags) >> +{ >> + struct radeon_ring *ring = >> &rdev->ring[rdev->asic->vm.pt_ring_index]; >> + uint64_t addr; >> + int i; >> + uint64_t pde_gpu_addr = RADEON_BASE_GPU_ADDR(vm) + >> + PDE_OFFSET(rdev, pde_index); >> + >> + addr = 0; >> + >> + radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + pde_count * >> 2)); >> + radeon_ring_write(ring, pde_gpu_addr & 0xffffffff); >> + radeon_ring_write(ring, (pde_gpu_addr >> 32) & 0xff); >> + >> + for (i = 0; i < pde_count; ++i) { >> + addr = RADEON_BASE_GPU_ADDR(vm) + >> + PTE_OFFSET(rdev, i + pde_index, 0); >> + addr = addr & 0xFFFFFFFFFFFFF000ULL; >> + addr |= flags; >> + >> radeon_ring_write(ring, addr & 0xffffffff); >> radeon_ring_write(ring, (addr >> 32) & 0xffffffff); >> } >> } >> + >> + > > Splitting that into two functions should be made unnecessary by the > interface cleanup patch. > > >> /** >> * cayman_vm_flush - vm flush using the CP >> * >> @@ -1571,7 +1599,7 @@ void cayman_vm_flush(struct radeon_device *rdev, >> struct radeon_ib *ib) >> radeon_ring_write(ring, vm->last_pfn); >> radeon_ring_write(ring, PACKET0(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + >> (vm->id << 2), 0)); >> - radeon_ring_write(ring, vm->pt_gpu_addr >> 12); >> + radeon_ring_write(ring, vm->pd_gpu_addr >> 12); >> /* flush hdp cache */ >> radeon_ring_write(ring, PACKET0(HDP_MEM_COHERENCY_FLUSH_CNTL, 0)); >> diff --git a/drivers/gpu/drm/radeon/radeon.h >> b/drivers/gpu/drm/radeon/radeon.h >> index 4d67f0f..062896f 100644 >> --- a/drivers/gpu/drm/radeon/radeon.h >> +++ b/drivers/gpu/drm/radeon/radeon.h >> @@ -655,8 +655,8 @@ struct radeon_vm { >> struct list_head va; >> unsigned id; >> unsigned last_pfn; >> - u64 pt_gpu_addr; >> - u64 *pt; >> + u64 pd_gpu_addr; >> + u64 __iomem *pd_addr; > > We probably never need the CPU address of the page directory, except for > initial clearing it. > > >> struct radeon_sa_bo *sa_bo; >> struct mutex mutex; >> /* last fence for cs using this vm */ >> @@ -679,6 +679,57 @@ struct radeon_vm_manager { >> bool enabled; >> }; >> + >> + >> +/* We consider the case where BLOCK_SIZE is 0 */ >> +/* So PDE is 19 bits long, PTE is 9 and OFFSET is 12 */ >> +#define RADEON_BLOCK_SIZE 0 >> + >> +/* By default there are 512 entries in Page Table */ >> +#define RADEON_DEFAULT_PTE_COUNT (1 << 9) >> + >> +/* number of PTEs in Page Table */ >> +#define RADEON_PTE_COUNT (RADEON_DEFAULT_PTE_COUNT << RADEON_BLOCK_SIZE) >> + >> +/* Get last PDE number containing nth PTE */ >> +#define RADEON_GET_LAST_PDE_FOR_PFN(_n) ((_n) / RADEON_PTE_COUNT) >> + >> +/* Get PTE number to containing nth pfn */ >> +#define RADEON_GET_PTE_FOR_PFN(_n) ((_n) % RADEON_PTE_COUNT) >> + >> +/* Number of PDE tables to cover n PTEs */ >> +#define RADEON_PDE_COUNT_FOR_N_PAGES(_n) \ >> + (((_n) + RADEON_PTE_COUNT - 1) / RADEON_PTE_COUNT) >> + >> +/* Number of PDE tables to cover max_pfn (maximum number of PTEs) */ >> +#define RADEON_TOTAL_PDE_COUNT(rdev) \ >> + RADEON_PDE_COUNT_FOR_N_PAGES(rdev->vm_manager.max_pfn) >> + >> +#define PTE_SIZE 8 >> +#define PDE_SIZE 8 >> + >> +/* offset for nth PDE starting from beginning of PDE table */ >> +#define PDE_OFFSET(_rdev, _npde) ((_npde) * PDE_SIZE) >> + >> +#define PT_OFFSET(_rdev) \ >> + (RADEON_GPU_PAGE_ALIGN(RADEON_TOTAL_PDE_COUNT(rdev) * PDE_SIZE)) >> + >> +/* offset for nth PTE of mth PDE starting from beginning of PDE table */ >> +#define PTE_OFFSET(_rdev, _npde, _npte) \ >> + (PT_OFFSET(_rdev) + \ >> + (_npde) * RADEON_PTE_COUNT * PTE_SIZE + \ >> + (_npte) * PTE_SIZE) > > All defines should have RADEON_ prefix, also I'm not 100% sure why we need > that as a global define, only the vm code should deal with that. > > >> + >> +/* cpu address of gpuvm page table */ >> +#define RADEON_BASE_CPU_ADDR(_vm) \ >> + radeon_sa_bo_cpu_addr(vm->sa_bo) >> + >> +/* gpu address of gpuvm page table */ >> +#define RADEON_BASE_GPU_ADDR(_vm) \ >> + radeon_sa_bo_gpu_addr(vm->sa_bo) >> + >> +#define RADEON_PDE_FLAG_VALID 1 >> + >> /* >> * file private structure >> */ >> @@ -1141,9 +1192,20 @@ struct radeon_asic { >> void (*fini)(struct radeon_device *rdev); >> u32 pt_ring_index; >> - void (*set_page)(struct radeon_device *rdev, struct >> radeon_vm *vm, >> - unsigned pfn, struct ttm_mem_reg *mem, >> - unsigned npages, uint32_t flags); >> + void (*set_page)( >> + struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + uint64_t pte_gpu_addr, uint64_t mem_pfn_offset, >> + unsigned count, uint32_t flags >> +); >> + >> + void (*set_pdes)( >> + struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + struct radeon_vm *vm, >> + uint64_t pde_index, >> + unsigned pde_count, uint32_t flags); >> + >> } vm; >> /* ring specific callbacks */ >> struct { >> @@ -1755,7 +1817,12 @@ void radeon_ring_write(struct radeon_ring *ring, >> uint32_t v); >> #define radeon_gart_set_page(rdev, i, p) >> (rdev)->asic->gart.set_page((rdev), (i), (p)) >> #define radeon_asic_vm_init(rdev) (rdev)->asic->vm.init((rdev)) >> #define radeon_asic_vm_fini(rdev) (rdev)->asic->vm.fini((rdev)) >> -#define radeon_asic_vm_set_page(rdev, v, pfn, mem, npages, flags) >> (rdev)->asic->vm.set_page((rdev), (v), (pfn), (mem), (npages), (flags)) >> +#define radeon_asic_vm_set_page(rdev, mem, start_pte, start_pfn, count, >> flags) \ >> +((rdev)->asic->vm.set_page((rdev), (mem), (start_pte), (start_pfn), \ >> + (count), (flags))) >> +#define radeon_asic_vm_set_pdes(rdev, mem, vm, pde_index, pde_count, >> flags) \ >> +((rdev)->asic->vm.set_pdes((rdev), (mem), (vm), (pde_index), \ >> + (pde_count), (flags))) >> #define radeon_ring_start(rdev, r, cp) >> (rdev)->asic->ring[(r)].ring_start((rdev), (cp)) >> #define radeon_ring_test(rdev, r, cp) >> (rdev)->asic->ring[(r)].ring_test((rdev), (cp)) >> #define radeon_ib_test(rdev, r, cp) >> (rdev)->asic->ring[(r)].ib_test((rdev), (cp)) >> diff --git a/drivers/gpu/drm/radeon/radeon_asic.c >> b/drivers/gpu/drm/radeon/radeon_asic.c >> index 2f7adea..0df6a55 100644 >> --- a/drivers/gpu/drm/radeon/radeon_asic.c >> +++ b/drivers/gpu/drm/radeon/radeon_asic.c >> @@ -1377,6 +1377,7 @@ static struct radeon_asic cayman_asic = { >> .fini = &cayman_vm_fini, >> .pt_ring_index = RADEON_RING_TYPE_GFX_INDEX, >> .set_page = &cayman_vm_set_page, >> + .set_pdes = &cayman_vm_set_pdes, >> }, >> .ring = { >> [RADEON_RING_TYPE_GFX_INDEX] = { >> diff --git a/drivers/gpu/drm/radeon/radeon_asic.h >> b/drivers/gpu/drm/radeon/radeon_asic.h >> index 29b3d05..7a694c0 100644 >> --- a/drivers/gpu/drm/radeon/radeon_asic.h >> +++ b/drivers/gpu/drm/radeon/radeon_asic.h >> @@ -442,9 +442,16 @@ int cayman_vm_init(struct radeon_device *rdev); >> void cayman_vm_fini(struct radeon_device *rdev); >> void cayman_vm_flush(struct radeon_device *rdev, struct radeon_ib *ib); >> uint32_t cayman_vm_page_flags(struct radeon_device *rdev, uint32_t >> flags); >> -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, >> - unsigned pfn, struct ttm_mem_reg *mem, >> - unsigned npages, uint32_t flags); >> +void cayman_vm_set_page(struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + uint64_t pte_gpu_addr, uint64_t mem_pfn_offset, >> + unsigned count, uint32_t flags); >> +void cayman_vm_set_pdes(struct radeon_device *rdev, >> + struct ttm_mem_reg *mem, >> + struct radeon_vm *vm, >> + uint64_t pde_index, >> + unsigned pde_count, uint32_t flags); >> + >> int evergreen_ib_parse(struct radeon_device *rdev, struct radeon_ib >> *ib); >> /* DCE6 - SI */ >> diff --git a/drivers/gpu/drm/radeon/radeon_gart.c >> b/drivers/gpu/drm/radeon/radeon_gart.c >> index 2f28ff3..f9bda6e 100644 >> --- a/drivers/gpu/drm/radeon/radeon_gart.c >> +++ b/drivers/gpu/drm/radeon/radeon_gart.c >> @@ -490,7 +490,6 @@ static void radeon_vm_free_pt(struct radeon_device >> *rdev, >> list_del_init(&vm->list); >> radeon_sa_bo_free(rdev, &vm->sa_bo, vm->fence); >> - vm->pt = NULL; >> list_for_each_entry(bo_va, &vm->va, vm_list) { >> bo_va->valid = false; >> @@ -547,6 +546,10 @@ int radeon_vm_alloc_pt(struct radeon_device *rdev, >> struct radeon_vm *vm) >> struct radeon_vm *vm_evict; >> int r; >> + int gpuvm_tables_sz = >> + RADEON_GPU_PAGE_ALIGN(RADEON_TOTAL_PDE_COUNT(rdev) * PDE_SIZE) >> + >> + vm->last_pfn * PTE_SIZE; >> + >> if (vm == NULL) { >> return -EINVAL; >> } >> @@ -560,7 +563,7 @@ int radeon_vm_alloc_pt(struct radeon_device *rdev, >> struct radeon_vm *vm) >> retry: >> r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, >> &vm->sa_bo, >> - RADEON_GPU_PAGE_ALIGN(vm->last_pfn * 8), >> + gpuvm_tables_sz, >> RADEON_GPU_PAGE_SIZE, false); >> if (r == -ENOMEM) { >> if (list_empty(&rdev->vm_manager.lru_vm)) { >> @@ -576,9 +579,13 @@ retry: >> return r; >> } >> - vm->pt = radeon_sa_bo_cpu_addr(vm->sa_bo); >> - vm->pt_gpu_addr = radeon_sa_bo_gpu_addr(vm->sa_bo); >> - memset(vm->pt, 0, RADEON_GPU_PAGE_ALIGN(vm->last_pfn * 8)); >> + DRM_INFO("radeon: reserved %d kb pd & pt tables\n", >> + gpuvm_tables_sz/1024); >> + >> + vm->pd_addr = radeon_sa_bo_cpu_addr(vm->sa_bo); >> + vm->pd_gpu_addr = radeon_sa_bo_gpu_addr(vm->sa_bo); >> + memset(vm->pd_addr, 0, gpuvm_tables_sz); >> + >> list_add_tail(&vm->list, &rdev->vm_manager.lru_vm); >> return radeon_vm_bo_update_pte(rdev, vm, rdev->ring_tmp_bo.bo, >> @@ -861,6 +868,63 @@ u64 radeon_vm_get_addr(struct radeon_device *rdev, >> } >> /** >> + * radeon_vm_bo_set_pages - update PDE and PTE for pfn >> + * >> + * @rdev: radeon_device pointer >> + * @first_pfn: first pfn in bo address space to start mapping from >> + * @mem: ttm mem >> + * @vm: requested vm >> + * @bo: radeon buffer object >> + * >> + * Update page directory and table entries for pfn (cayman+). >> + */ >> + >> +int radeon_vm_bo_set_pages(struct radeon_device *rdev, struct radeon_vm >> *vm, >> + unsigned first_pfn, struct ttm_mem_reg *mem, >> + unsigned npages, uint32_t flags) > > Shouldn't that be a static function? > > >> +{ >> + u64 pde_num, pte_num, start_pde, pde_count = 0; >> + >> + unsigned pt_total, count, pfn_idx; >> + unsigned last_pfn = first_pfn + npages, pfns_to_pt_edge, >> pfns_to_end; >> + unsigned mem_pfn_offset; >> + >> + pt_total = RADEON_PDE_COUNT_FOR_N_PAGES(npages); >> + count = RADEON_PTE_COUNT; >> + >> + pfn_idx = first_pfn; >> + >> + for (mem_pfn_offset = 0; mem_pfn_offset < npages;) { >> + >> + pfns_to_end = last_pfn - pfn_idx; >> + pfns_to_pt_edge = RADEON_PTE_COUNT - >> + (pfn_idx % RADEON_PTE_COUNT); >> + >> + count = pfns_to_pt_edge < pfns_to_end ? >> + pfns_to_pt_edge : pfns_to_end; >> + >> + pde_num = RADEON_GET_LAST_PDE_FOR_PFN(pfn_idx); >> + pte_num = RADEON_GET_PTE_FOR_PFN(pfn_idx); >> + >> + radeon_asic_vm_set_page(rdev, mem, >> + vm->pd_gpu_addr + >> + PTE_OFFSET(rdev, pde_num, >> pte_num), >> + mem_pfn_offset, count, flags); >> + pfn_idx += count; >> + mem_pfn_offset += count; >> + >> + pde_count++; >> + } >> + >> + start_pde = RADEON_GET_LAST_PDE_FOR_PFN(first_pfn); >> + >> + radeon_asic_vm_set_pdes(rdev, mem, vm, >> + start_pde, pde_count, >> RADEON_PDE_FLAG_VALID); >> + >> + return 0; >> +} >> + >> +/** >> * radeon_vm_bo_update_pte - map a bo into the vm page table >> * >> * @rdev: radeon_device pointer >> @@ -882,7 +946,7 @@ int radeon_vm_bo_update_pte(struct radeon_device >> *rdev, >> struct radeon_ring *ring = &rdev->ring[ridx]; >> struct radeon_semaphore *sem = NULL; >> struct radeon_bo_va *bo_va; >> - unsigned ngpu_pages, ndw; >> + unsigned ngpu_pages, ndw, npdes; >> uint64_t pfn; >> int r; >> @@ -935,10 +999,12 @@ int radeon_vm_bo_update_pte(struct radeon_device >> *rdev, >> } >> } >> - /* estimate number of dw needed */ >> + npdes = (RADEON_PDE_COUNT_FOR_N_PAGES(pfn + ngpu_pages) - >> + RADEON_GET_LAST_PDE_FOR_PFN(pfn)); >> + >> ndw = 32; >> - ndw += (ngpu_pages >> 12) * 3; >> - ndw += ngpu_pages * 2; >> + ndw += ngpu_pages * 2 + 3 * npdes; >> + ndw += npdes * 2 + 3; >> r = radeon_ring_lock(rdev, ring, ndw); >> if (r) { >> @@ -950,7 +1016,7 @@ int radeon_vm_bo_update_pte(struct radeon_device >> *rdev, >> radeon_fence_note_sync(vm->fence, ridx); >> } >> - radeon_asic_vm_set_page(rdev, vm, pfn, mem, ngpu_pages, >> bo_va->flags); >> + radeon_vm_bo_set_pages(rdev, vm, pfn, mem, ngpu_pages, >> bo_va->flags); >> radeon_fence_unref(&vm->fence); >> r = radeon_fence_emit(rdev, &vm->fence, ridx); >> @@ -958,9 +1024,11 @@ int radeon_vm_bo_update_pte(struct radeon_device >> *rdev, >> radeon_ring_unlock_undo(rdev, ring); >> return r; >> } >> + >> radeon_ring_unlock_commit(rdev, ring); >> radeon_semaphore_free(rdev, &sem, vm->fence); >> radeon_fence_unref(&vm->last_flush); >> + >> return 0; >> } >> diff --git a/drivers/gpu/drm/radeon/si.c b/drivers/gpu/drm/radeon/si.c >> index 2a5c337..156c994 100644 >> --- a/drivers/gpu/drm/radeon/si.c >> +++ b/drivers/gpu/drm/radeon/si.c >> @@ -2426,7 +2426,7 @@ static int si_pcie_gart_enable(struct radeon_device >> *rdev) >> WREG32(VM_CONTEXT1_PROTECTION_FAULT_DEFAULT_ADDR, >> (u32)(rdev->dummy_page.addr >> 12)); >> WREG32(VM_CONTEXT1_CNTL2, 0); >> - WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) | >> + WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(1) | >> RANGE_PROTECTION_FAULT_ENABLE_DEFAULT); >> si_pcie_gart_tlb_flush(rdev); >> @@ -2804,7 +2804,7 @@ void si_vm_flush(struct radeon_device *rdev, struct >> radeon_ib *ib) >> radeon_ring_write(ring, >> PACKET0(VM_CONTEXT8_PAGE_TABLE_BASE_ADDR >> + ((vm->id - 8) << 2), >> 0)); >> } >> - radeon_ring_write(ring, vm->pt_gpu_addr >> 12); >> + radeon_ring_write(ring, vm->pd_gpu_addr >> 12); >> /* flush hdp cache */ >> radeon_ring_write(ring, PACKET0(HDP_MEM_COHERENCY_FLUSH_CNTL, 0)); > > > Cheers, > Christian. >
Dropping David and kernel mailing list for now, they are probably not immediately interested in such specific discussion. On 13.09.2012 18:52, Dmitry Cherkassov wrote: > Christian, > >> as I already noted in our internal discussion, the first step to >> hierarchical page table support should be to cleanup the set_page interface. >> Please see the attached patch, it does exactly this. I suggest that you >> rebase on it and try to don't touch the chipset specific code in ni.c any >> more (if you find that you need to change something there move that into >> this cleanup patch instead). >> > Currently we alloc all PTs in one call and they lay at fixed distance > from each other. > Using this knowledge we can setup all PDEs of BO in one pass. > > I suggest the following changes to the cleanup patch: > * add BLOCK_SIZE bitfield to bo_va->flags (RADEON_VM_FLAGS) > * shift RADEON_GPU_PAGE_SIZE by BLOCK_SIZE obtained from flags here: > > void cayman_vm_set_page(struct radeon_device *rdev, uint64_t pe, > uint64_t addr, unsigned count, uint32_t flags) > { > + unsigned block_size = RADEON_EXTRACT_BLOCKSIZE(flags) > struct radeon_ring *ring = &rdev->ring[rdev->asic->vm.pt_ring_index]; > uint32_t r600_flags = cayman_vm_page_flags(rdev, flags); > > > } else if (flags & RADEON_VM_PAGE_VALID) { > value = addr; > - addr += RADEON_GPU_PAGE_SIZE; > + addr += RADEON_GPU_PAGE_SIZE << block_size; > } I would rather prefer a separate "incr" parameter that gets added to addr on each loop. When we keep in mind how we want to implement this in the future that interface seems to match. > In future when PTt's may be scattered around due to their on-demand > allocation we may devise > a method to supply list of their addresses to this code (array or > callback function, whatever). > > What do you think? Even with the smallest blocksize possible a single page directory entry covers 2MB virtual memory space, so calling set_page (maybe we should also call it set_pages instead) multiple times doesn't seems to be so much of an overhead. When we really chose to optimize that case also I would indeed split that into two functions, but not distinct by pde/pte, but instead by set_scattered_pages and set_linear_pages. With set_scattered_pages taking a list of addresses and set_linear_pages taking just count, addr and increment..... But I'm really unsure if we should do that, especially on a first implementation. Cheers, Christian. > > >> Additionally to that see the notes below: >> >> >> On 13.09.2012 16:13, Dmitry Cherkasov wrote: >>> PDE/PTE update code uses CP ring for memory writes. >>> All page table entries are preallocated for now in alloc_pt(). >>> >>> It is made as whole because it's hard to divide it to several patches >>> that compile and doesn't break anything being applied separately. >>> >>> Tested on cayman card. >>> >>> Signed-off-by: Dmitry Cherkasov <Dmitrii.Cherkasov@amd.com> >>> --- >>> I couldn't test in on SI card, so would be happy if someone could check it >>> there. >>> >>> drivers/gpu/drm/radeon/ni.c | 64 ++++++++++++++++++------- >>> drivers/gpu/drm/radeon/radeon.h | 79 >>> +++++++++++++++++++++++++++--- >>> drivers/gpu/drm/radeon/radeon_asic.c | 1 + >>> drivers/gpu/drm/radeon/radeon_asic.h | 13 +++-- >>> drivers/gpu/drm/radeon/radeon_gart.c | 88 >>> ++++++++++++++++++++++++++++++---- >>> drivers/gpu/drm/radeon/si.c | 4 +- >>> 6 files changed, 210 insertions(+), 39 deletions(-) >>> >>> diff --git a/drivers/gpu/drm/radeon/ni.c b/drivers/gpu/drm/radeon/ni.c >>> index b238216..31d9027 100644 >>> --- a/drivers/gpu/drm/radeon/ni.c >>> +++ b/drivers/gpu/drm/radeon/ni.c >>> @@ -782,7 +782,7 @@ static int cayman_pcie_gart_enable(struct >>> radeon_device *rdev) >>> (u32)(rdev->dummy_page.addr >> 12)); >>> WREG32(VM_CONTEXT1_CNTL2, 0); >>> WREG32(VM_CONTEXT1_CNTL, 0); >>> - WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) | >>> + WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(1) | >>> RANGE_PROTECTION_FAULT_ENABLE_DEFAULT); >>> cayman_pcie_gart_tlb_flush(rdev); >>> @@ -1517,37 +1517,65 @@ uint32_t cayman_vm_page_flags(struct radeon_device >>> *rdev, uint32_t flags) >>> return r600_flags; >>> } >>> -/** >>> - * cayman_vm_set_page - update the page tables using the CP >>> - * >>> - * @rdev: radeon_device pointer >>> - * >>> - * Update the page tables using the CP (cayman-si). >>> - */ >> It's probably just a rebase artifact, please never remove those >> documentation lines. Instead fix them if the need arise. >> >> >>> -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, >>> - unsigned pfn, struct ttm_mem_reg *mem, >>> - unsigned npages, uint32_t flags) >>> +void cayman_vm_set_page(struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + uint64_t pte_gpu_addr, >>> + uint64_t mem_pfn_offset, >>> + unsigned count, uint32_t flags) >>> { >>> struct radeon_ring *ring = >>> &rdev->ring[rdev->asic->vm.pt_ring_index]; >>> - uint64_t addr, pt = vm->pt_gpu_addr + pfn * 8; >>> + uint64_t addr; >>> int i; >>> addr = flags = cayman_vm_page_flags(rdev, flags); >>> - radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + npages * >>> 2)); >>> - radeon_ring_write(ring, pt & 0xffffffff); >>> - radeon_ring_write(ring, (pt >> 32) & 0xff); >>> - for (i = 0; i < npages; ++i) { >>> + radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + count * 2)); >>> + radeon_ring_write(ring, pte_gpu_addr & 0xffffffff); >>> + radeon_ring_write(ring, (pte_gpu_addr >> 32) & 0xff); >>> + for (i = 0; i < count; ++i) { >>> if (mem) { >>> - addr = radeon_vm_get_addr(rdev, mem, i); >>> + addr = radeon_vm_get_addr(rdev, mem, >>> + i + mem_pfn_offset); >>> addr = addr & 0xFFFFFFFFFFFFF000ULL; >>> addr |= flags; >>> } >>> + >>> + radeon_ring_write(ring, addr & 0xffffffff); >>> + radeon_ring_write(ring, (addr >> 32) & 0xffffffff); >>> + } >>> +} >>> + >>> +void cayman_vm_set_pdes(struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + struct radeon_vm *vm, >>> + uint64_t pde_index, >>> + unsigned pde_count, uint32_t flags) >>> +{ >>> + struct radeon_ring *ring = >>> &rdev->ring[rdev->asic->vm.pt_ring_index]; >>> + uint64_t addr; >>> + int i; >>> + uint64_t pde_gpu_addr = RADEON_BASE_GPU_ADDR(vm) + >>> + PDE_OFFSET(rdev, pde_index); >>> + >>> + addr = 0; >>> + >>> + radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + pde_count * >>> 2)); >>> + radeon_ring_write(ring, pde_gpu_addr & 0xffffffff); >>> + radeon_ring_write(ring, (pde_gpu_addr >> 32) & 0xff); >>> + >>> + for (i = 0; i < pde_count; ++i) { >>> + addr = RADEON_BASE_GPU_ADDR(vm) + >>> + PTE_OFFSET(rdev, i + pde_index, 0); >>> + addr = addr & 0xFFFFFFFFFFFFF000ULL; >>> + addr |= flags; >>> + >>> radeon_ring_write(ring, addr & 0xffffffff); >>> radeon_ring_write(ring, (addr >> 32) & 0xffffffff); >>> } >>> } >>> + >>> + >> Splitting that into two functions should be made unnecessary by the >> interface cleanup patch. >> >> >>> /** >>> * cayman_vm_flush - vm flush using the CP >>> * >>> @@ -1571,7 +1599,7 @@ void cayman_vm_flush(struct radeon_device *rdev, >>> struct radeon_ib *ib) >>> radeon_ring_write(ring, vm->last_pfn); >>> radeon_ring_write(ring, PACKET0(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + >>> (vm->id << 2), 0)); >>> - radeon_ring_write(ring, vm->pt_gpu_addr >> 12); >>> + radeon_ring_write(ring, vm->pd_gpu_addr >> 12); >>> /* flush hdp cache */ >>> radeon_ring_write(ring, PACKET0(HDP_MEM_COHERENCY_FLUSH_CNTL, 0)); >>> diff --git a/drivers/gpu/drm/radeon/radeon.h >>> b/drivers/gpu/drm/radeon/radeon.h >>> index 4d67f0f..062896f 100644 >>> --- a/drivers/gpu/drm/radeon/radeon.h >>> +++ b/drivers/gpu/drm/radeon/radeon.h >>> @@ -655,8 +655,8 @@ struct radeon_vm { >>> struct list_head va; >>> unsigned id; >>> unsigned last_pfn; >>> - u64 pt_gpu_addr; >>> - u64 *pt; >>> + u64 pd_gpu_addr; >>> + u64 __iomem *pd_addr; >> We probably never need the CPU address of the page directory, except for >> initial clearing it. >> >> >>> struct radeon_sa_bo *sa_bo; >>> struct mutex mutex; >>> /* last fence for cs using this vm */ >>> @@ -679,6 +679,57 @@ struct radeon_vm_manager { >>> bool enabled; >>> }; >>> + >>> + >>> +/* We consider the case where BLOCK_SIZE is 0 */ >>> +/* So PDE is 19 bits long, PTE is 9 and OFFSET is 12 */ >>> +#define RADEON_BLOCK_SIZE 0 >>> + >>> +/* By default there are 512 entries in Page Table */ >>> +#define RADEON_DEFAULT_PTE_COUNT (1 << 9) >>> + >>> +/* number of PTEs in Page Table */ >>> +#define RADEON_PTE_COUNT (RADEON_DEFAULT_PTE_COUNT << RADEON_BLOCK_SIZE) >>> + >>> +/* Get last PDE number containing nth PTE */ >>> +#define RADEON_GET_LAST_PDE_FOR_PFN(_n) ((_n) / RADEON_PTE_COUNT) >>> + >>> +/* Get PTE number to containing nth pfn */ >>> +#define RADEON_GET_PTE_FOR_PFN(_n) ((_n) % RADEON_PTE_COUNT) >>> + >>> +/* Number of PDE tables to cover n PTEs */ >>> +#define RADEON_PDE_COUNT_FOR_N_PAGES(_n) \ >>> + (((_n) + RADEON_PTE_COUNT - 1) / RADEON_PTE_COUNT) >>> + >>> +/* Number of PDE tables to cover max_pfn (maximum number of PTEs) */ >>> +#define RADEON_TOTAL_PDE_COUNT(rdev) \ >>> + RADEON_PDE_COUNT_FOR_N_PAGES(rdev->vm_manager.max_pfn) >>> + >>> +#define PTE_SIZE 8 >>> +#define PDE_SIZE 8 >>> + >>> +/* offset for nth PDE starting from beginning of PDE table */ >>> +#define PDE_OFFSET(_rdev, _npde) ((_npde) * PDE_SIZE) >>> + >>> +#define PT_OFFSET(_rdev) \ >>> + (RADEON_GPU_PAGE_ALIGN(RADEON_TOTAL_PDE_COUNT(rdev) * PDE_SIZE)) >>> + >>> +/* offset for nth PTE of mth PDE starting from beginning of PDE table */ >>> +#define PTE_OFFSET(_rdev, _npde, _npte) \ >>> + (PT_OFFSET(_rdev) + \ >>> + (_npde) * RADEON_PTE_COUNT * PTE_SIZE + \ >>> + (_npte) * PTE_SIZE) >> All defines should have RADEON_ prefix, also I'm not 100% sure why we need >> that as a global define, only the vm code should deal with that. >> >> >>> + >>> +/* cpu address of gpuvm page table */ >>> +#define RADEON_BASE_CPU_ADDR(_vm) \ >>> + radeon_sa_bo_cpu_addr(vm->sa_bo) >>> + >>> +/* gpu address of gpuvm page table */ >>> +#define RADEON_BASE_GPU_ADDR(_vm) \ >>> + radeon_sa_bo_gpu_addr(vm->sa_bo) >>> + >>> +#define RADEON_PDE_FLAG_VALID 1 >>> + >>> /* >>> * file private structure >>> */ >>> @@ -1141,9 +1192,20 @@ struct radeon_asic { >>> void (*fini)(struct radeon_device *rdev); >>> u32 pt_ring_index; >>> - void (*set_page)(struct radeon_device *rdev, struct >>> radeon_vm *vm, >>> - unsigned pfn, struct ttm_mem_reg *mem, >>> - unsigned npages, uint32_t flags); >>> + void (*set_page)( >>> + struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + uint64_t pte_gpu_addr, uint64_t mem_pfn_offset, >>> + unsigned count, uint32_t flags >>> +); >>> + >>> + void (*set_pdes)( >>> + struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + struct radeon_vm *vm, >>> + uint64_t pde_index, >>> + unsigned pde_count, uint32_t flags); >>> + >>> } vm; >>> /* ring specific callbacks */ >>> struct { >>> @@ -1755,7 +1817,12 @@ void radeon_ring_write(struct radeon_ring *ring, >>> uint32_t v); >>> #define radeon_gart_set_page(rdev, i, p) >>> (rdev)->asic->gart.set_page((rdev), (i), (p)) >>> #define radeon_asic_vm_init(rdev) (rdev)->asic->vm.init((rdev)) >>> #define radeon_asic_vm_fini(rdev) (rdev)->asic->vm.fini((rdev)) >>> -#define radeon_asic_vm_set_page(rdev, v, pfn, mem, npages, flags) >>> (rdev)->asic->vm.set_page((rdev), (v), (pfn), (mem), (npages), (flags)) >>> +#define radeon_asic_vm_set_page(rdev, mem, start_pte, start_pfn, count, >>> flags) \ >>> +((rdev)->asic->vm.set_page((rdev), (mem), (start_pte), (start_pfn), \ >>> + (count), (flags))) >>> +#define radeon_asic_vm_set_pdes(rdev, mem, vm, pde_index, pde_count, >>> flags) \ >>> +((rdev)->asic->vm.set_pdes((rdev), (mem), (vm), (pde_index), \ >>> + (pde_count), (flags))) >>> #define radeon_ring_start(rdev, r, cp) >>> (rdev)->asic->ring[(r)].ring_start((rdev), (cp)) >>> #define radeon_ring_test(rdev, r, cp) >>> (rdev)->asic->ring[(r)].ring_test((rdev), (cp)) >>> #define radeon_ib_test(rdev, r, cp) >>> (rdev)->asic->ring[(r)].ib_test((rdev), (cp)) >>> diff --git a/drivers/gpu/drm/radeon/radeon_asic.c >>> b/drivers/gpu/drm/radeon/radeon_asic.c >>> index 2f7adea..0df6a55 100644 >>> --- a/drivers/gpu/drm/radeon/radeon_asic.c >>> +++ b/drivers/gpu/drm/radeon/radeon_asic.c >>> @@ -1377,6 +1377,7 @@ static struct radeon_asic cayman_asic = { >>> .fini = &cayman_vm_fini, >>> .pt_ring_index = RADEON_RING_TYPE_GFX_INDEX, >>> .set_page = &cayman_vm_set_page, >>> + .set_pdes = &cayman_vm_set_pdes, >>> }, >>> .ring = { >>> [RADEON_RING_TYPE_GFX_INDEX] = { >>> diff --git a/drivers/gpu/drm/radeon/radeon_asic.h >>> b/drivers/gpu/drm/radeon/radeon_asic.h >>> index 29b3d05..7a694c0 100644 >>> --- a/drivers/gpu/drm/radeon/radeon_asic.h >>> +++ b/drivers/gpu/drm/radeon/radeon_asic.h >>> @@ -442,9 +442,16 @@ int cayman_vm_init(struct radeon_device *rdev); >>> void cayman_vm_fini(struct radeon_device *rdev); >>> void cayman_vm_flush(struct radeon_device *rdev, struct radeon_ib *ib); >>> uint32_t cayman_vm_page_flags(struct radeon_device *rdev, uint32_t >>> flags); >>> -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, >>> - unsigned pfn, struct ttm_mem_reg *mem, >>> - unsigned npages, uint32_t flags); >>> +void cayman_vm_set_page(struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + uint64_t pte_gpu_addr, uint64_t mem_pfn_offset, >>> + unsigned count, uint32_t flags); >>> +void cayman_vm_set_pdes(struct radeon_device *rdev, >>> + struct ttm_mem_reg *mem, >>> + struct radeon_vm *vm, >>> + uint64_t pde_index, >>> + unsigned pde_count, uint32_t flags); >>> + >>> int evergreen_ib_parse(struct radeon_device *rdev, struct radeon_ib >>> *ib); >>> /* DCE6 - SI */ >>> diff --git a/drivers/gpu/drm/radeon/radeon_gart.c >>> b/drivers/gpu/drm/radeon/radeon_gart.c >>> index 2f28ff3..f9bda6e 100644 >>> --- a/drivers/gpu/drm/radeon/radeon_gart.c >>> +++ b/drivers/gpu/drm/radeon/radeon_gart.c >>> @@ -490,7 +490,6 @@ static void radeon_vm_free_pt(struct radeon_device >>> *rdev, >>> list_del_init(&vm->list); >>> radeon_sa_bo_free(rdev, &vm->sa_bo, vm->fence); >>> - vm->pt = NULL; >>> list_for_each_entry(bo_va, &vm->va, vm_list) { >>> bo_va->valid = false; >>> @@ -547,6 +546,10 @@ int radeon_vm_alloc_pt(struct radeon_device *rdev, >>> struct radeon_vm *vm) >>> struct radeon_vm *vm_evict; >>> int r; >>> + int gpuvm_tables_sz = >>> + RADEON_GPU_PAGE_ALIGN(RADEON_TOTAL_PDE_COUNT(rdev) * PDE_SIZE) >>> + >>> + vm->last_pfn * PTE_SIZE; >>> + >>> if (vm == NULL) { >>> return -EINVAL; >>> } >>> @@ -560,7 +563,7 @@ int radeon_vm_alloc_pt(struct radeon_device *rdev, >>> struct radeon_vm *vm) >>> retry: >>> r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, >>> &vm->sa_bo, >>> - RADEON_GPU_PAGE_ALIGN(vm->last_pfn * 8), >>> + gpuvm_tables_sz, >>> RADEON_GPU_PAGE_SIZE, false); >>> if (r == -ENOMEM) { >>> if (list_empty(&rdev->vm_manager.lru_vm)) { >>> @@ -576,9 +579,13 @@ retry: >>> return r; >>> } >>> - vm->pt = radeon_sa_bo_cpu_addr(vm->sa_bo); >>> - vm->pt_gpu_addr = radeon_sa_bo_gpu_addr(vm->sa_bo); >>> - memset(vm->pt, 0, RADEON_GPU_PAGE_ALIGN(vm->last_pfn * 8)); >>> + DRM_INFO("radeon: reserved %d kb pd & pt tables\n", >>> + gpuvm_tables_sz/1024); >>> + >>> + vm->pd_addr = radeon_sa_bo_cpu_addr(vm->sa_bo); >>> + vm->pd_gpu_addr = radeon_sa_bo_gpu_addr(vm->sa_bo); >>> + memset(vm->pd_addr, 0, gpuvm_tables_sz); >>> + >>> list_add_tail(&vm->list, &rdev->vm_manager.lru_vm); >>> return radeon_vm_bo_update_pte(rdev, vm, rdev->ring_tmp_bo.bo, >>> @@ -861,6 +868,63 @@ u64 radeon_vm_get_addr(struct radeon_device *rdev, >>> } >>> /** >>> + * radeon_vm_bo_set_pages - update PDE and PTE for pfn >>> + * >>> + * @rdev: radeon_device pointer >>> + * @first_pfn: first pfn in bo address space to start mapping from >>> + * @mem: ttm mem >>> + * @vm: requested vm >>> + * @bo: radeon buffer object >>> + * >>> + * Update page directory and table entries for pfn (cayman+). >>> + */ >>> + >>> +int radeon_vm_bo_set_pages(struct radeon_device *rdev, struct radeon_vm >>> *vm, >>> + unsigned first_pfn, struct ttm_mem_reg *mem, >>> + unsigned npages, uint32_t flags) >> Shouldn't that be a static function? >> >> >>> +{ >>> + u64 pde_num, pte_num, start_pde, pde_count = 0; >>> + >>> + unsigned pt_total, count, pfn_idx; >>> + unsigned last_pfn = first_pfn + npages, pfns_to_pt_edge, >>> pfns_to_end; >>> + unsigned mem_pfn_offset; >>> + >>> + pt_total = RADEON_PDE_COUNT_FOR_N_PAGES(npages); >>> + count = RADEON_PTE_COUNT; >>> + >>> + pfn_idx = first_pfn; >>> + >>> + for (mem_pfn_offset = 0; mem_pfn_offset < npages;) { >>> + >>> + pfns_to_end = last_pfn - pfn_idx; >>> + pfns_to_pt_edge = RADEON_PTE_COUNT - >>> + (pfn_idx % RADEON_PTE_COUNT); >>> + >>> + count = pfns_to_pt_edge < pfns_to_end ? >>> + pfns_to_pt_edge : pfns_to_end; >>> + >>> + pde_num = RADEON_GET_LAST_PDE_FOR_PFN(pfn_idx); >>> + pte_num = RADEON_GET_PTE_FOR_PFN(pfn_idx); >>> + >>> + radeon_asic_vm_set_page(rdev, mem, >>> + vm->pd_gpu_addr + >>> + PTE_OFFSET(rdev, pde_num, >>> pte_num), >>> + mem_pfn_offset, count, flags); >>> + pfn_idx += count; >>> + mem_pfn_offset += count; >>> + >>> + pde_count++; >>> + } >>> + >>> + start_pde = RADEON_GET_LAST_PDE_FOR_PFN(first_pfn); >>> + >>> + radeon_asic_vm_set_pdes(rdev, mem, vm, >>> + start_pde, pde_count, >>> RADEON_PDE_FLAG_VALID); >>> + >>> + return 0; >>> +} >>> + >>> +/** >>> * radeon_vm_bo_update_pte - map a bo into the vm page table >>> * >>> * @rdev: radeon_device pointer >>> @@ -882,7 +946,7 @@ int radeon_vm_bo_update_pte(struct radeon_device >>> *rdev, >>> struct radeon_ring *ring = &rdev->ring[ridx]; >>> struct radeon_semaphore *sem = NULL; >>> struct radeon_bo_va *bo_va; >>> - unsigned ngpu_pages, ndw; >>> + unsigned ngpu_pages, ndw, npdes; >>> uint64_t pfn; >>> int r; >>> @@ -935,10 +999,12 @@ int radeon_vm_bo_update_pte(struct radeon_device >>> *rdev, >>> } >>> } >>> - /* estimate number of dw needed */ >>> + npdes = (RADEON_PDE_COUNT_FOR_N_PAGES(pfn + ngpu_pages) - >>> + RADEON_GET_LAST_PDE_FOR_PFN(pfn)); >>> + >>> ndw = 32; >>> - ndw += (ngpu_pages >> 12) * 3; >>> - ndw += ngpu_pages * 2; >>> + ndw += ngpu_pages * 2 + 3 * npdes; >>> + ndw += npdes * 2 + 3; >>> r = radeon_ring_lock(rdev, ring, ndw); >>> if (r) { >>> @@ -950,7 +1016,7 @@ int radeon_vm_bo_update_pte(struct radeon_device >>> *rdev, >>> radeon_fence_note_sync(vm->fence, ridx); >>> } >>> - radeon_asic_vm_set_page(rdev, vm, pfn, mem, ngpu_pages, >>> bo_va->flags); >>> + radeon_vm_bo_set_pages(rdev, vm, pfn, mem, ngpu_pages, >>> bo_va->flags); >>> radeon_fence_unref(&vm->fence); >>> r = radeon_fence_emit(rdev, &vm->fence, ridx); >>> @@ -958,9 +1024,11 @@ int radeon_vm_bo_update_pte(struct radeon_device >>> *rdev, >>> radeon_ring_unlock_undo(rdev, ring); >>> return r; >>> } >>> + >>> radeon_ring_unlock_commit(rdev, ring); >>> radeon_semaphore_free(rdev, &sem, vm->fence); >>> radeon_fence_unref(&vm->last_flush); >>> + >>> return 0; >>> } >>> diff --git a/drivers/gpu/drm/radeon/si.c b/drivers/gpu/drm/radeon/si.c >>> index 2a5c337..156c994 100644 >>> --- a/drivers/gpu/drm/radeon/si.c >>> +++ b/drivers/gpu/drm/radeon/si.c >>> @@ -2426,7 +2426,7 @@ static int si_pcie_gart_enable(struct radeon_device >>> *rdev) >>> WREG32(VM_CONTEXT1_PROTECTION_FAULT_DEFAULT_ADDR, >>> (u32)(rdev->dummy_page.addr >> 12)); >>> WREG32(VM_CONTEXT1_CNTL2, 0); >>> - WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) | >>> + WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(1) | >>> RANGE_PROTECTION_FAULT_ENABLE_DEFAULT); >>> si_pcie_gart_tlb_flush(rdev); >>> @@ -2804,7 +2804,7 @@ void si_vm_flush(struct radeon_device *rdev, struct >>> radeon_ib *ib) >>> radeon_ring_write(ring, >>> PACKET0(VM_CONTEXT8_PAGE_TABLE_BASE_ADDR >>> + ((vm->id - 8) << 2), >>> 0)); >>> } >>> - radeon_ring_write(ring, vm->pt_gpu_addr >> 12); >>> + radeon_ring_write(ring, vm->pd_gpu_addr >> 12); >>> /* flush hdp cache */ >>> radeon_ring_write(ring, PACKET0(HDP_MEM_COHERENCY_FLUSH_CNTL, 0)); >> >> Cheers, >> Christian. >> > >
From 250216bfb15e024f4df0d9ae1de277233b75e92d Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Christian=20K=C3=B6nig?= <deathsimple@vodafone.de> Date: Thu, 13 Sep 2012 16:59:52 +0200 Subject: [PATCH] drm/radeon: refactor set_page chipset interface MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Cleanup the interface in preparation for hierarchical page tables. Signed-off-by: Christian König <deathsimple@vodafone.de> --- drivers/gpu/drm/radeon/ni.c | 39 +++++++++++++++----------- drivers/gpu/drm/radeon/radeon.h | 11 +++----- drivers/gpu/drm/radeon/radeon_asic.h | 5 ++-- drivers/gpu/drm/radeon/radeon_gart.c | 50 +++++++++++++--------------------- 4 files changed, 48 insertions(+), 57 deletions(-) diff --git a/drivers/gpu/drm/radeon/ni.c b/drivers/gpu/drm/radeon/ni.c index 2bcff7fe..df25996 100644 --- a/drivers/gpu/drm/radeon/ni.c +++ b/drivers/gpu/drm/radeon/ni.c @@ -1531,30 +1531,37 @@ uint32_t cayman_vm_page_flags(struct radeon_device *rdev, uint32_t flags) * cayman_vm_set_page - update the page tables using the CP * * @rdev: radeon_device pointer + * @pe: addr of the page entry + * @addr: dst addr to write into pe + * @count: number of page entries to update + * @flags: access flags * * Update the page tables using the CP (cayman-si). */ -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, - unsigned pfn, struct ttm_mem_reg *mem, - unsigned npages, uint32_t flags) +void cayman_vm_set_page(struct radeon_device *rdev, uint64_t pe, + uint64_t addr, unsigned count, uint32_t flags) { struct radeon_ring *ring = &rdev->ring[rdev->asic->vm.pt_ring_index]; - uint64_t addr, pt = vm->pt_gpu_addr + pfn * 8; + uint32_t r600_flags = cayman_vm_page_flags(rdev, flags); int i; - addr = flags = cayman_vm_page_flags(rdev, flags); - - radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + npages * 2)); - radeon_ring_write(ring, pt & 0xffffffff); - radeon_ring_write(ring, (pt >> 32) & 0xff); - for (i = 0; i < npages; ++i) { - if (mem) { - addr = radeon_vm_get_addr(rdev, mem, i); - addr = addr & 0xFFFFFFFFFFFFF000ULL; - addr |= flags; + radeon_ring_write(ring, PACKET3(PACKET3_ME_WRITE, 1 + count * 2)); + radeon_ring_write(ring, pe & 0xffffffff); + radeon_ring_write(ring, (pe >> 32) & 0xff); + for (i = 0; i < count; ++i) { + uint64_t value = 0; + if (flags & RADEON_VM_PAGE_SYSTEM) { + value = radeon_vm_map_gart(rdev, addr); + value &= 0xFFFFFFFFFFFFF000ULL; + addr += RADEON_GPU_PAGE_SIZE; + + } else if (flags & RADEON_VM_PAGE_VALID) { + value = addr; + addr += RADEON_GPU_PAGE_SIZE; } - radeon_ring_write(ring, addr & 0xffffffff); - radeon_ring_write(ring, (addr >> 32) & 0xffffffff); + value |= r600_flags; + radeon_ring_write(ring, value & 0xffffffff); + radeon_ring_write(ring, (value >> 32) & 0xffffffff); } } diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h index 74714e51..b7224a4 100644 --- a/drivers/gpu/drm/radeon/radeon.h +++ b/drivers/gpu/drm/radeon/radeon.h @@ -1141,9 +1141,8 @@ struct radeon_asic { void (*fini)(struct radeon_device *rdev); u32 pt_ring_index; - void (*set_page)(struct radeon_device *rdev, struct radeon_vm *vm, - unsigned pfn, struct ttm_mem_reg *mem, - unsigned npages, uint32_t flags); + void (*set_page)(struct radeon_device *rdev, uint64_t pe, + uint64_t addr, unsigned count, uint32_t flags); } vm; /* ring specific callbacks */ struct { @@ -1755,7 +1754,7 @@ void radeon_ring_write(struct radeon_ring *ring, uint32_t v); #define radeon_gart_set_page(rdev, i, p) (rdev)->asic->gart.set_page((rdev), (i), (p)) #define radeon_asic_vm_init(rdev) (rdev)->asic->vm.init((rdev)) #define radeon_asic_vm_fini(rdev) (rdev)->asic->vm.fini((rdev)) -#define radeon_asic_vm_set_page(rdev, v, pfn, mem, npages, flags) (rdev)->asic->vm.set_page((rdev), (v), (pfn), (mem), (npages), (flags)) +#define radeon_asic_vm_set_page(rdev, pe, addr, count, flags) (rdev)->asic->vm.set_page((rdev), (pe), (addr), (count), (flags)) #define radeon_ring_start(rdev, r, cp) (rdev)->asic->ring[(r)].ring_start((rdev), (cp)) #define radeon_ring_test(rdev, r, cp) (rdev)->asic->ring[(r)].ring_test((rdev), (cp)) #define radeon_ib_test(rdev, r, cp) (rdev)->asic->ring[(r)].ib_test((rdev), (cp)) @@ -1840,9 +1839,7 @@ struct radeon_fence *radeon_vm_grab_id(struct radeon_device *rdev, void radeon_vm_fence(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_fence *fence); -u64 radeon_vm_get_addr(struct radeon_device *rdev, - struct ttm_mem_reg *mem, - unsigned pfn); +uint64_t radeon_vm_map_gart(struct radeon_device *rdev, uint64_t addr); int radeon_vm_bo_update_pte(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_bo *bo, diff --git a/drivers/gpu/drm/radeon/radeon_asic.h b/drivers/gpu/drm/radeon/radeon_asic.h index 29b3d05..7be6778 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.h +++ b/drivers/gpu/drm/radeon/radeon_asic.h @@ -442,9 +442,8 @@ int cayman_vm_init(struct radeon_device *rdev); void cayman_vm_fini(struct radeon_device *rdev); void cayman_vm_flush(struct radeon_device *rdev, struct radeon_ib *ib); uint32_t cayman_vm_page_flags(struct radeon_device *rdev, uint32_t flags); -void cayman_vm_set_page(struct radeon_device *rdev, struct radeon_vm *vm, - unsigned pfn, struct ttm_mem_reg *mem, - unsigned npages, uint32_t flags); +void cayman_vm_set_page(struct radeon_device *rdev, uint64_t pe, + uint64_t addr, unsigned count, uint32_t flags); int evergreen_ib_parse(struct radeon_device *rdev, struct radeon_ib *ib); /* DCE6 - SI */ diff --git a/drivers/gpu/drm/radeon/radeon_gart.c b/drivers/gpu/drm/radeon/radeon_gart.c index 2f28ff3..e9a72dc 100644 --- a/drivers/gpu/drm/radeon/radeon_gart.c +++ b/drivers/gpu/drm/radeon/radeon_gart.c @@ -822,42 +822,26 @@ int radeon_vm_bo_set_addr(struct radeon_device *rdev, } /** - * radeon_vm_get_addr - get the physical address of the page + * radeon_vm_map_gart - get the physical address of a gart page * * @rdev: radeon_device pointer - * @mem: ttm mem - * @pfn: pfn + * @addr: the unmapped addr * * Look up the physical address of the page that the pte resolves * to (cayman+). * Returns the physical address of the page. */ -u64 radeon_vm_get_addr(struct radeon_device *rdev, - struct ttm_mem_reg *mem, - unsigned pfn) +uint64_t radeon_vm_map_gart(struct radeon_device *rdev, uint64_t addr) { - u64 addr = 0; - - switch (mem->mem_type) { - case TTM_PL_VRAM: - addr = (mem->start << PAGE_SHIFT); - addr += pfn * RADEON_GPU_PAGE_SIZE; - addr += rdev->vm_manager.vram_base_offset; - break; - case TTM_PL_TT: - /* offset inside page table */ - addr = mem->start << PAGE_SHIFT; - addr += pfn * RADEON_GPU_PAGE_SIZE; - addr = addr >> PAGE_SHIFT; - /* page table offset */ - addr = rdev->gart.pages_addr[addr]; - /* in case cpu page size != gpu page size*/ - addr += (pfn * RADEON_GPU_PAGE_SIZE) & (~PAGE_MASK); - break; - default: - break; - } - return addr; + uint64_t result; + + /* page table offset */ + result = rdev->gart.pages_addr[addr >> PAGE_SHIFT]; + + /* in case cpu page size != gpu page size*/ + result |= addr & (~PAGE_MASK); + + return result; } /** @@ -883,7 +867,7 @@ int radeon_vm_bo_update_pte(struct radeon_device *rdev, struct radeon_semaphore *sem = NULL; struct radeon_bo_va *bo_va; unsigned ngpu_pages, ndw; - uint64_t pfn; + uint64_t pfn, addr; int r; /* nothing to do if vm isn't bound */ @@ -908,21 +892,25 @@ int radeon_vm_bo_update_pte(struct radeon_device *rdev, ngpu_pages = radeon_bo_ngpu_pages(bo); bo_va->flags &= ~RADEON_VM_PAGE_VALID; bo_va->flags &= ~RADEON_VM_PAGE_SYSTEM; + pfn = bo_va->soffset / RADEON_GPU_PAGE_SIZE; if (mem) { + addr = mem->start << PAGE_SHIFT; if (mem->mem_type != TTM_PL_SYSTEM) { bo_va->flags |= RADEON_VM_PAGE_VALID; bo_va->valid = true; } if (mem->mem_type == TTM_PL_TT) { bo_va->flags |= RADEON_VM_PAGE_SYSTEM; + } else { + addr += rdev->vm_manager.vram_base_offset; } if (!bo_va->valid) { mem = NULL; } } else { + addr = 0; bo_va->valid = false; } - pfn = bo_va->soffset / RADEON_GPU_PAGE_SIZE; if (vm->fence && radeon_fence_signaled(vm->fence)) { radeon_fence_unref(&vm->fence); @@ -950,7 +938,7 @@ int radeon_vm_bo_update_pte(struct radeon_device *rdev, radeon_fence_note_sync(vm->fence, ridx); } - radeon_asic_vm_set_page(rdev, vm, pfn, mem, ngpu_pages, bo_va->flags); + radeon_asic_vm_set_page(rdev, vm->pt_gpu_addr + pfn * 8, addr, ngpu_pages, bo_va->flags); radeon_fence_unref(&vm->fence); r = radeon_fence_emit(rdev, &vm->fence, ridx); -- 1.7.9.5