int ret;
ret = drm_bo_lock_kmm(bo);
- if (ret) {
- if (ret == -EAGAIN)
- schedule();
+ if (ret)
return ret;
- }
drm_bo_unmap_virtual(bo);
if (old_is_pci)
drm_bo_finish_unmap(bo);
}
+
+
static int drm_bo_handle_move_mem(drm_buffer_object_t *bo,
drm_bo_mem_reg_t *mem,
int evict,
drm_mem_type_manager_t *new_man = &bm->man[mem->mem_type];
int ret = 0;
-
+
if (old_is_pci || new_is_pci)
ret = drm_bo_vm_pre_move(bo, old_is_pci);
if (ret)
return ret;
+ /*
+ * Create and bind a ttm if required.
+ */
+
if (!(new_man->flags & _DRM_FLAG_MEMTYPE_FIXED) &&
- (bo->ttm == NULL))
+ (bo->ttm == NULL)) {
ret = drm_bo_add_ttm(bo);
- if (ret)
- return ret;
+ if (ret)
+ goto out_err;
- if (!(old_man->flags & _DRM_FLAG_MEMTYPE_FIXED) &&
+ if (mem->mem_type != DRM_BO_MEM_LOCAL) {
+ ret = drm_bind_ttm(bo->ttm, new_man->flags &
+ DRM_BO_FLAG_CACHED,
+ mem->mm_node->start);
+ if (ret)
+ goto out_err;
+ }
+ }
+
+ if ((bo->mem.mem_type == DRM_BO_MEM_LOCAL) && bo->ttm == NULL) {
+
+ drm_bo_mem_reg_t *old_mem = &bo->mem;
+ uint32_t save_flags = old_mem->flags;
+ uint32_t save_mask = old_mem->mask;
+
+ *old_mem = *mem;
+ mem->mm_node = NULL;
+ old_mem->mask = save_mask;
+ DRM_FLAG_MASKED(save_flags, mem->flags,
+ DRM_BO_MASK_MEMTYPE);
+
+ } else if (!(old_man->flags & _DRM_FLAG_MEMTYPE_FIXED) &&
!(new_man->flags & _DRM_FLAG_MEMTYPE_FIXED)) {
+
ret = drm_bo_move_ttm(bo, evict, no_wait, mem);
+
} else if (dev->driver->bo_driver->move) {
ret = dev->driver->bo_driver->move(bo, evict, no_wait, mem);
+
} else {
+
ret = drm_bo_move_memcpy(bo, evict, no_wait, mem);
+
}
+ if (ret)
+ goto out_err;
+
if (old_is_pci || new_is_pci)
drm_bo_vm_post_move(bo);
- if (ret)
- return ret;
-
if (bo->priv_flags & _DRM_BO_FLAG_EVICTED) {
ret = dev->driver->bo_driver->invalidate_caches(dev, bo->mem.flags);
if (ret)
bo->offset = bo->mem.mm_node->start << PAGE_SHIFT;
return 0;
+
+out_err:
+ if (old_is_pci || new_is_pci)
+ drm_bo_vm_post_move(bo);
+
+ new_man = &bm->man[bo->mem.mem_type];
+ if ((new_man->flags & _DRM_FLAG_MEMTYPE_FIXED) && bo->ttm) {
+ drm_ttm_unbind(bo->ttm);
+ drm_destroy_ttm(bo->ttm);
+ bo->ttm = NULL;
+ }
+
+ return ret;
}
/*
bm->nice_mode = 0;
DRM_ERROR("Detected GPU lockup or "
"fence driver was taken down. "
- "Evicting waiting buffers.\n");
+ "Evicting buffer.\n");
}
if (bo->fence) {
drm_fence_usage_deref_unlocked(dev, bo->fence);
* Note that new_mem_flags are NOT transferred to the bo->mem.mask.
*/
-static int drm_bo_move_buffer(drm_buffer_object_t * bo, uint32_t new_mem_flags,
- int no_wait, int force_no_move, int move_unfenced)
+int drm_bo_move_buffer(drm_buffer_object_t * bo, uint32_t new_mem_flags,
+ int no_wait, int force_no_move, int move_unfenced)
{
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
out:
mutex_unlock(&bo->mutex);
+
drm_bo_usage_deref_unlocked(bo);
return ret;
}
#endif
bo->dev = dev;
bo->type = type;
+ bo->mem.mem_type = DRM_BO_MEM_LOCAL;
bo->mem.num_pages = num_pages;
bo->mem.mm_node = NULL;
bo->mem.page_alignment = page_alignment;
bo->buffer_start = buffer_start;
}
bo->priv_flags = 0;
- bo->mem.flags = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
- bo->mem.mask = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
+ bo->mem.flags = 0;
+ bo->mem.mask = 0;
atomic_inc(&bm->count);
ret = drm_bo_new_mask(bo, mask, hint);
out_err:
mutex_unlock(&bo->mutex);
+
drm_bo_usage_deref_unlocked(bo);
return ret;
}
return -ENOMEM;
memcpy_fromio(dst, src, PAGE_SIZE);
- kunmap(dst);
+ kunmap(d);
return 0;
}
return -ENOMEM;
memcpy_toio(dst, src, PAGE_SIZE);
- kunmap(src);
+ kunmap(s);
return 0;
}
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
- fbo = drm_ctl_calloc(1, sizeof(*bo), DRM_MEM_BUFOBJ);
+ fbo = drm_ctl_calloc(1, sizeof(*fbo), DRM_MEM_BUFOBJ);
if (!fbo)
return -ENOMEM;
mutex_lock(&fbo->mutex);
mutex_lock(&dev->struct_mutex);
+ DRM_INIT_WAITQUEUE(&bo->event_queue);
INIT_LIST_HEAD(&fbo->ddestroy);
INIT_LIST_HEAD(&fbo->lru);
+#ifdef DRM_ODD_MM_COMPAT
+ INIT_LIST_HEAD(&fbo->vma_list);
+ INIT_LIST_HEAD(&fbo->p_mm_list);
+#endif
atomic_inc(&bo->fence->usage);
-
fbo->mem.mm_node->private = (void *)fbo;
atomic_set(&fbo->usage, 1);
atomic_inc(&bm->count);
return 0;
}
+/*
+ * Since move is underway, we need to block signals in this function.
+ * We cannot restart until it has finished.
+ */
+
int drm_bo_move_accel_cleanup(drm_buffer_object_t *bo,
int evict,
int no_wait,
if (bo->fence)
drm_fence_usage_deref_unlocked(dev, bo->fence);
-
ret = drm_fence_object_create(dev, fence_type,
fence_flags | DRM_FENCE_FLAG_EMIT,
&bo->fence);
if (ret)
return ret;
- if (evict) {
+#ifdef DRM_ODD_MM_COMPAT
+ /*
+ * In this mode, we don't allow pipelining a copy blit,
+ * since the buffer will be accessible from user space
+ * the moment we return and rebuild the page tables.
+ *
+ * With normal vm operation, page tables are rebuilt
+ * on demand using fault(), which waits for buffer idle.
+ */
+ if (1)
+#else
+ if (evict)
+#endif
+ {
ret = drm_bo_wait(bo, 0, 1, 0);
- if (ret) {
- DRM_ERROR("Wait failure\n");
+ if (ret)
return ret;
- }
if (old_mem->mm_node) {
mutex_lock(&dev->struct_mutex);
drm_mm_put_block(old_mem->mm_node);
*/
ret = drm_buffer_object_transfer(bo, &old_obj);
+
if (ret)
return ret;
else
bo->ttm = NULL;
- atomic_inc(&old_obj->fence->usage);
mutex_lock(&dev->struct_mutex);
list_del(&old_obj->lru);
+ DRM_FLAG_MASKED(bo->priv_flags, 0, _DRM_BO_FLAG_UNFENCED);
drm_bo_add_to_lru(old_obj, &old_obj->dev->bm);
drm_bo_usage_deref_locked(old_obj);
mutex_unlock(&dev->struct_mutex);
mutex_lock(&bo->mutex);
+ err = drm_bo_wait(bo, 0, 0, 0);
+ if (err) {
+ data->type = (err == -EAGAIN) ?
+ VM_FAULT_MINOR : VM_FAULT_SIGBUS;
+ goto out_unlock;
+ }
+
+
/*
* If buffer happens to be in a non-mappable location,
* move it to a mappable.
err = drm_bo_move_buffer(bo, new_mask, 0, 0);
bo->mem.mask = mask_save;
- if (!err)
- err = drm_bo_wait(bo, 0, 0, 0);
-
if (err) {
data->type = (err == -EAGAIN) ?
VM_FAULT_MINOR : VM_FAULT_SIGBUS;
goto out_unlock;
}
}
-
+
if (address > vma->vm_end) {
data->type = VM_FAULT_SIGBUS;
goto out_unlock;
{
drm_bo_mem_reg_t *old_mem = &bo->mem;
- if (old_mem->mem_type == DRM_BO_MEM_LOCAL)
+ if (old_mem->mem_type == DRM_BO_MEM_LOCAL) {
return drm_bo_move_memcpy(bo, evict, no_wait, new_mem);
- if (new_mem->mem_type == DRM_BO_MEM_LOCAL) {
+ } else if (new_mem->mem_type == DRM_BO_MEM_LOCAL) {
if (i915_move_flip(bo, evict, no_wait, new_mem))
return drm_bo_move_memcpy(bo, evict, no_wait, new_mem);
} else {