struct kfd_process_device *pdd;
void *mem;
struct kfd_dev *dev;
- struct svm_range_list *svms = &p->svms;
int idr_handle;
long err;
uint64_t offset = args->mmap_offset;
/* Flush pending deferred work to avoid racing with deferred actions
* from previous memory map changes (e.g. munmap).
*/
- svm_range_list_lock_and_flush_work(svms, current->mm);
- mutex_lock(&svms->lock);
+ svm_range_list_lock_and_flush_work(&p->svms, current->mm);
+ mutex_lock(&p->svms.lock);
mmap_write_unlock(current->mm);
- if (interval_tree_iter_first(&svms->objects,
+ if (interval_tree_iter_first(&p->svms.objects,
args->va_addr >> PAGE_SHIFT,
(args->va_addr + args->size - 1) >> PAGE_SHIFT)) {
pr_err("Address: 0x%llx already allocated by SVM\n",
args->va_addr);
- mutex_unlock(&svms->lock);
+ mutex_unlock(&p->svms.lock);
return -EADDRINUSE;
}
- mutex_unlock(&svms->lock);
+ mutex_unlock(&p->svms.lock);
#endif
dev = kfd_device_by_id(args->gpu_id);
if (!dev)