diff options
author | Thomas Hellstrom <thomas-at-tungstengraphics-dot-com> | 2007-02-10 12:06:36 +0100 |
---|---|---|
committer | Thomas Hellstrom <thomas-at-tungstengraphics-dot-com> | 2007-02-10 12:06:36 +0100 |
commit | 85ee2a8d044cd4d8de4894a794151af9471648e3 (patch) | |
tree | a5d8feb968026c059aafebc356cab7ed55624eb8 | |
parent | 53aee3122a1821b8ca24ed2bc5c1940cb0f2ff8e (diff) |
Various bugfixes.
-rw-r--r-- | linux-core/drm_bo.c | 79 | ||||
-rw-r--r-- | linux-core/drm_bo_move.c | 40 | ||||
-rw-r--r-- | linux-core/drm_vm.c | 13 | ||||
-rw-r--r-- | linux-core/i915_buffer.c | 4 |
4 files changed, 103 insertions, 33 deletions
diff --git a/linux-core/drm_bo.c b/linux-core/drm_bo.c index 8d2b544e..41f4e002 100644 --- a/linux-core/drm_bo.c +++ b/linux-core/drm_bo.c @@ -85,11 +85,8 @@ static int drm_bo_vm_pre_move(drm_buffer_object_t *bo, int ret; ret = drm_bo_lock_kmm(bo); - if (ret) { - if (ret == -EAGAIN) - schedule(); + if (ret) return ret; - } drm_bo_unmap_virtual(bo); if (old_is_pci) drm_bo_finish_unmap(bo); @@ -142,6 +139,8 @@ static int drm_bo_add_ttm(drm_buffer_object_t * bo) } + + static int drm_bo_handle_move_mem(drm_buffer_object_t *bo, drm_bo_mem_reg_t *mem, int evict, @@ -155,33 +154,63 @@ static int drm_bo_handle_move_mem(drm_buffer_object_t *bo, drm_mem_type_manager_t *new_man = &bm->man[mem->mem_type]; int ret = 0; - + if (old_is_pci || new_is_pci) ret = drm_bo_vm_pre_move(bo, old_is_pci); if (ret) return ret; + /* + * Create and bind a ttm if required. + */ + if (!(new_man->flags & _DRM_FLAG_MEMTYPE_FIXED) && - (bo->ttm == NULL)) + (bo->ttm == NULL)) { ret = drm_bo_add_ttm(bo); - if (ret) - return ret; + if (ret) + goto out_err; - if (!(old_man->flags & _DRM_FLAG_MEMTYPE_FIXED) && + if (mem->mem_type != DRM_BO_MEM_LOCAL) { + ret = drm_bind_ttm(bo->ttm, new_man->flags & + DRM_BO_FLAG_CACHED, + mem->mm_node->start); + if (ret) + goto out_err; + } + } + + if ((bo->mem.mem_type == DRM_BO_MEM_LOCAL) && bo->ttm == NULL) { + + drm_bo_mem_reg_t *old_mem = &bo->mem; + uint32_t save_flags = old_mem->flags; + uint32_t save_mask = old_mem->mask; + + *old_mem = *mem; + mem->mm_node = NULL; + old_mem->mask = save_mask; + DRM_FLAG_MASKED(save_flags, mem->flags, + DRM_BO_MASK_MEMTYPE); + + } else if (!(old_man->flags & _DRM_FLAG_MEMTYPE_FIXED) && !(new_man->flags & _DRM_FLAG_MEMTYPE_FIXED)) { + ret = drm_bo_move_ttm(bo, evict, no_wait, mem); + } else if (dev->driver->bo_driver->move) { ret = dev->driver->bo_driver->move(bo, evict, no_wait, mem); + } else { + ret = drm_bo_move_memcpy(bo, evict, no_wait, mem); + } + if (ret) + goto out_err; + if (old_is_pci || new_is_pci) drm_bo_vm_post_move(bo); - if (ret) - return ret; - if (bo->priv_flags & _DRM_BO_FLAG_EVICTED) { ret = dev->driver->bo_driver->invalidate_caches(dev, bo->mem.flags); if (ret) @@ -196,6 +225,19 @@ static int drm_bo_handle_move_mem(drm_buffer_object_t *bo, bo->offset = bo->mem.mm_node->start << PAGE_SHIFT; return 0; + +out_err: + if (old_is_pci || new_is_pci) + drm_bo_vm_post_move(bo); + + new_man = &bm->man[bo->mem.mem_type]; + if ((new_man->flags & _DRM_FLAG_MEMTYPE_FIXED) && bo->ttm) { + drm_ttm_unbind(bo->ttm); + drm_destroy_ttm(bo->ttm); + bo->ttm = NULL; + } + + return ret; } /* @@ -269,7 +311,7 @@ static void drm_bo_cleanup_refs(drm_buffer_object_t *bo, int remove_all) bm->nice_mode = 0; DRM_ERROR("Detected GPU lockup or " "fence driver was taken down. " - "Evicting waiting buffers.\n"); + "Evicting buffer.\n"); } if (bo->fence) { drm_fence_usage_deref_unlocked(dev, bo->fence); @@ -1148,8 +1190,8 @@ static void drm_buffer_user_object_unmap(drm_file_t * priv, * Note that new_mem_flags are NOT transferred to the bo->mem.mask. */ -static int drm_bo_move_buffer(drm_buffer_object_t * bo, uint32_t new_mem_flags, - int no_wait, int force_no_move, int move_unfenced) +int drm_bo_move_buffer(drm_buffer_object_t * bo, uint32_t new_mem_flags, + int no_wait, int force_no_move, int move_unfenced) { drm_device_t *dev = bo->dev; drm_buffer_manager_t *bm = &dev->bm; @@ -1387,6 +1429,7 @@ static int drm_bo_handle_validate(drm_file_t * priv, uint32_t handle, out: mutex_unlock(&bo->mutex); + drm_bo_usage_deref_unlocked(bo); return ret; } @@ -1481,6 +1524,7 @@ int drm_buffer_object_create(drm_file_t * priv, #endif bo->dev = dev; bo->type = type; + bo->mem.mem_type = DRM_BO_MEM_LOCAL; bo->mem.num_pages = num_pages; bo->mem.mm_node = NULL; bo->mem.page_alignment = page_alignment; @@ -1491,8 +1535,8 @@ int drm_buffer_object_create(drm_file_t * priv, bo->buffer_start = buffer_start; } bo->priv_flags = 0; - bo->mem.flags = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED; - bo->mem.mask = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED; + bo->mem.flags = 0; + bo->mem.mask = 0; atomic_inc(&bm->count); ret = drm_bo_new_mask(bo, mask, hint); @@ -1517,6 +1561,7 @@ int drm_buffer_object_create(drm_file_t * priv, out_err: mutex_unlock(&bo->mutex); + drm_bo_usage_deref_unlocked(bo); return ret; } diff --git a/linux-core/drm_bo_move.c b/linux-core/drm_bo_move.c index d712a70f..3347f945 100644 --- a/linux-core/drm_bo_move.c +++ b/linux-core/drm_bo_move.c @@ -168,7 +168,7 @@ static int drm_copy_io_ttm_page(drm_ttm_t *ttm, void *src, unsigned long page) return -ENOMEM; memcpy_fromio(dst, src, PAGE_SIZE); - kunmap(dst); + kunmap(d); return 0; } @@ -186,7 +186,7 @@ static int drm_copy_ttm_io_page(drm_ttm_t *ttm, void *dst, unsigned long page) return -ENOMEM; memcpy_toio(dst, src, PAGE_SIZE); - kunmap(src); + kunmap(s); return 0; } @@ -283,7 +283,7 @@ int drm_buffer_object_transfer(drm_buffer_object_t *bo, drm_device_t *dev = bo->dev; drm_buffer_manager_t *bm = &dev->bm; - fbo = drm_ctl_calloc(1, sizeof(*bo), DRM_MEM_BUFOBJ); + fbo = drm_ctl_calloc(1, sizeof(*fbo), DRM_MEM_BUFOBJ); if (!fbo) return -ENOMEM; @@ -292,11 +292,15 @@ int drm_buffer_object_transfer(drm_buffer_object_t *bo, mutex_lock(&fbo->mutex); mutex_lock(&dev->struct_mutex); + DRM_INIT_WAITQUEUE(&bo->event_queue); INIT_LIST_HEAD(&fbo->ddestroy); INIT_LIST_HEAD(&fbo->lru); +#ifdef DRM_ODD_MM_COMPAT + INIT_LIST_HEAD(&fbo->vma_list); + INIT_LIST_HEAD(&fbo->p_mm_list); +#endif atomic_inc(&bo->fence->usage); - fbo->mem.mm_node->private = (void *)fbo; atomic_set(&fbo->usage, 1); atomic_inc(&bm->count); @@ -307,6 +311,11 @@ int drm_buffer_object_transfer(drm_buffer_object_t *bo, return 0; } +/* + * Since move is underway, we need to block signals in this function. + * We cannot restart until it has finished. + */ + int drm_bo_move_accel_cleanup(drm_buffer_object_t *bo, int evict, int no_wait, @@ -324,19 +333,29 @@ int drm_bo_move_accel_cleanup(drm_buffer_object_t *bo, if (bo->fence) drm_fence_usage_deref_unlocked(dev, bo->fence); - ret = drm_fence_object_create(dev, fence_type, fence_flags | DRM_FENCE_FLAG_EMIT, &bo->fence); if (ret) return ret; - if (evict) { +#ifdef DRM_ODD_MM_COMPAT + /* + * In this mode, we don't allow pipelining a copy blit, + * since the buffer will be accessible from user space + * the moment we return and rebuild the page tables. + * + * With normal vm operation, page tables are rebuilt + * on demand using fault(), which waits for buffer idle. + */ + if (1) +#else + if (evict) +#endif + { ret = drm_bo_wait(bo, 0, 1, 0); - if (ret) { - DRM_ERROR("Wait failure\n"); + if (ret) return ret; - } if (old_mem->mm_node) { mutex_lock(&dev->struct_mutex); drm_mm_put_block(old_mem->mm_node); @@ -359,6 +378,7 @@ int drm_bo_move_accel_cleanup(drm_buffer_object_t *bo, */ ret = drm_buffer_object_transfer(bo, &old_obj); + if (ret) return ret; @@ -367,9 +387,9 @@ int drm_bo_move_accel_cleanup(drm_buffer_object_t *bo, else bo->ttm = NULL; - atomic_inc(&old_obj->fence->usage); mutex_lock(&dev->struct_mutex); list_del(&old_obj->lru); + DRM_FLAG_MASKED(bo->priv_flags, 0, _DRM_BO_FLAG_UNFENCED); drm_bo_add_to_lru(old_obj, &old_obj->dev->bm); drm_bo_usage_deref_locked(old_obj); mutex_unlock(&dev->struct_mutex); diff --git a/linux-core/drm_vm.c b/linux-core/drm_vm.c index 4a41e761..17778c26 100644 --- a/linux-core/drm_vm.c +++ b/linux-core/drm_vm.c @@ -746,6 +746,14 @@ struct page *drm_bo_vm_fault(struct vm_area_struct *vma, mutex_lock(&bo->mutex); + err = drm_bo_wait(bo, 0, 0, 0); + if (err) { + data->type = (err == -EAGAIN) ? + VM_FAULT_MINOR : VM_FAULT_SIGBUS; + goto out_unlock; + } + + /* * If buffer happens to be in a non-mappable location, * move it to a mappable. @@ -760,16 +768,13 @@ struct page *drm_bo_vm_fault(struct vm_area_struct *vma, err = drm_bo_move_buffer(bo, new_mask, 0, 0); bo->mem.mask = mask_save; - if (!err) - err = drm_bo_wait(bo, 0, 0, 0); - if (err) { data->type = (err == -EAGAIN) ? VM_FAULT_MINOR : VM_FAULT_SIGBUS; goto out_unlock; } } - + if (address > vma->vm_end) { data->type = VM_FAULT_SIGBUS; goto out_unlock; diff --git a/linux-core/i915_buffer.c b/linux-core/i915_buffer.c index 70ba9a67..a357a53e 100644 --- a/linux-core/i915_buffer.c +++ b/linux-core/i915_buffer.c @@ -234,9 +234,9 @@ int i915_move(drm_buffer_object_t *bo, { drm_bo_mem_reg_t *old_mem = &bo->mem; - if (old_mem->mem_type == DRM_BO_MEM_LOCAL) + if (old_mem->mem_type == DRM_BO_MEM_LOCAL) { return drm_bo_move_memcpy(bo, evict, no_wait, new_mem); - if (new_mem->mem_type == DRM_BO_MEM_LOCAL) { + } else if (new_mem->mem_type == DRM_BO_MEM_LOCAL) { if (i915_move_flip(bo, evict, no_wait, new_mem)) return drm_bo_move_memcpy(bo, evict, no_wait, new_mem); } else { |