diff options
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c')
| -rw-r--r-- | drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 53 | 
1 files changed, 32 insertions, 21 deletions
| diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c index 4e2391c83d7c..3abfa66d72a2 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c @@ -333,7 +333,7 @@ void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base,  	base->next = bo->vm_bo;  	bo->vm_bo = base; -	if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv) +	if (!amdgpu_vm_is_bo_always_valid(vm, bo))  		return;  	dma_resv_assert_held(vm->root.bo->tbo.base.resv); @@ -1055,7 +1055,7 @@ int amdgpu_vm_update_range(struct amdgpu_device *adev, struct amdgpu_vm *vm,  				params.pages_addr = NULL;  			} -		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) { +		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT_FLAG(adev))) {  			addr = vram_base + cursor.start;  		} else {  			addr = 0; @@ -1101,13 +1101,13 @@ static void amdgpu_vm_bo_get_memory(struct amdgpu_bo_va *bo_va,  	 * For now ignore BOs which are currently locked and potentially  	 * changing their location.  	 */ -	if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv && +	if (!amdgpu_vm_is_bo_always_valid(vm, bo) &&  	    !dma_resv_trylock(bo->tbo.base.resv))  		return;  	amdgpu_bo_get_memory(bo, stats); -	if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv) -	    dma_resv_unlock(bo->tbo.base.resv); +	if (!amdgpu_vm_is_bo_always_valid(vm, bo)) +		dma_resv_unlock(bo->tbo.base.resv);  }  void amdgpu_vm_get_memory(struct amdgpu_vm *vm, @@ -1203,8 +1203,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,  		uncached = false;  	} -	if (clear || (bo && bo->tbo.base.resv == -		      vm->root.bo->tbo.base.resv)) +	if (clear || amdgpu_vm_is_bo_always_valid(vm, bo))  		last_update = &vm->last_update;  	else  		last_update = &bo_va->last_pt_update; @@ -1246,7 +1245,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,  	 * the evicted list so that it gets validated again on the  	 * next command submission.  	 */ -	if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) { +	if (amdgpu_vm_is_bo_always_valid(vm, bo)) {  		uint32_t mem_type = bo->tbo.resource->mem_type;  		if (!(bo->preferred_domains & @@ -1369,7 +1368,7 @@ static void amdgpu_vm_free_mapping(struct amdgpu_device *adev,  				   struct amdgpu_bo_va_mapping *mapping,  				   struct dma_fence *fence)  { -	if (mapping->flags & AMDGPU_PTE_PRT) +	if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev))  		amdgpu_vm_add_prt_cb(adev, fence);  	kfree(mapping);  } @@ -1637,13 +1636,12 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,  	list_add(&mapping->list, &bo_va->invalids);  	amdgpu_vm_it_insert(mapping, &vm->va); -	if (mapping->flags & AMDGPU_PTE_PRT) +	if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev))  		amdgpu_vm_prt_get(adev); -	if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && -	    !bo_va->base.moved) { +	if (amdgpu_vm_is_bo_always_valid(vm, bo) && !bo_va->base.moved)  		amdgpu_vm_bo_moved(&bo_va->base); -	} +  	trace_amdgpu_vm_bo_map(bo_va, mapping);  } @@ -1939,10 +1937,10 @@ int amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev,  		struct amdgpu_bo *bo = before->bo_va->base.bo;  		amdgpu_vm_it_insert(before, &vm->va); -		if (before->flags & AMDGPU_PTE_PRT) +		if (before->flags & AMDGPU_PTE_PRT_FLAG(adev))  			amdgpu_vm_prt_get(adev); -		if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && +		if (amdgpu_vm_is_bo_always_valid(vm, bo) &&  		    !before->bo_va->base.moved)  			amdgpu_vm_bo_moved(&before->bo_va->base);  	} else { @@ -1954,10 +1952,10 @@ int amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev,  		struct amdgpu_bo *bo = after->bo_va->base.bo;  		amdgpu_vm_it_insert(after, &vm->va); -		if (after->flags & AMDGPU_PTE_PRT) +		if (after->flags & AMDGPU_PTE_PRT_FLAG(adev))  			amdgpu_vm_prt_get(adev); -		if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && +		if (amdgpu_vm_is_bo_always_valid(vm, bo) &&  		    !after->bo_va->base.moved)  			amdgpu_vm_bo_moved(&after->bo_va->base);  	} else { @@ -2037,7 +2035,7 @@ void amdgpu_vm_bo_del(struct amdgpu_device *adev,  	if (bo) {  		dma_resv_assert_held(bo->tbo.base.resv); -		if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv) +		if (amdgpu_vm_is_bo_always_valid(vm, bo))  			ttm_bo_set_bulk_move(&bo->tbo, NULL);  		for (base = &bo_va->base.bo->vm_bo; *base; @@ -2131,7 +2129,7 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,  	for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {  		struct amdgpu_vm *vm = bo_base->vm; -		if (evicted && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) { +		if (evicted && amdgpu_vm_is_bo_always_valid(vm, bo)) {  			amdgpu_vm_bo_evicted(bo_base);  			continue;  		} @@ -2142,7 +2140,7 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,  		if (bo->tbo.type == ttm_bo_type_kernel)  			amdgpu_vm_bo_relocated(bo_base); -		else if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv) +		else if (amdgpu_vm_is_bo_always_valid(vm, bo))  			amdgpu_vm_bo_moved(bo_base);  		else  			amdgpu_vm_bo_invalidated(bo_base); @@ -2605,7 +2603,7 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)  	dma_fence_put(vm->last_tlb_flush);  	list_for_each_entry_safe(mapping, tmp, &vm->freed, list) { -		if (mapping->flags & AMDGPU_PTE_PRT && prt_fini_needed) { +		if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev) && prt_fini_needed) {  			amdgpu_vm_prt_fini(adev, vm);  			prt_fini_needed = false;  		} @@ -3006,3 +3004,16 @@ void amdgpu_vm_update_fault_cache(struct amdgpu_device *adev,  	xa_unlock_irqrestore(&adev->vm_manager.pasids, flags);  } +/** + * amdgpu_vm_is_bo_always_valid - check if the BO is VM always valid + * + * @vm: VM to test against. + * @bo: BO to be tested. + * + * Returns true if the BO shares the dma_resv object with the root PD and is + * always guaranteed to be valid inside the VM. + */ +bool amdgpu_vm_is_bo_always_valid(struct amdgpu_vm *vm, struct amdgpu_bo *bo) +{ +	return bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv; +} |