Commit 57de4ba9 authored by Jerome Glisse's avatar Jerome Glisse Committed by Dave Airlie

drm/ttm: simplify memory accounting for ttm user v2

Provide helper function to compute the kernel memory size needed
for each buffer object. Move all the accounting inside ttm, simplifying
driver and avoiding code duplication accross them.

v2 fix accounting of ghost object, one would have thought that i
   would have run into the issue since a longtime but it seems
   ghost object are rare when you have plenty of vram ;)
Signed-off-by: default avatarJerome Glisse <jglisse@redhat.com>
Reviewed-by: default avatarThomas Hellstrom <thellstrom@vmware.com>
parent 8e7e7052
...@@ -93,6 +93,7 @@ nouveau_bo_new(struct drm_device *dev, int size, int align, ...@@ -93,6 +93,7 @@ nouveau_bo_new(struct drm_device *dev, int size, int align,
{ {
struct drm_nouveau_private *dev_priv = dev->dev_private; struct drm_nouveau_private *dev_priv = dev->dev_private;
struct nouveau_bo *nvbo; struct nouveau_bo *nvbo;
size_t acc_size;
int ret; int ret;
nvbo = kzalloc(sizeof(struct nouveau_bo), GFP_KERNEL); nvbo = kzalloc(sizeof(struct nouveau_bo), GFP_KERNEL);
...@@ -115,9 +116,12 @@ nouveau_bo_new(struct drm_device *dev, int size, int align, ...@@ -115,9 +116,12 @@ nouveau_bo_new(struct drm_device *dev, int size, int align,
nvbo->bo.mem.num_pages = size >> PAGE_SHIFT; nvbo->bo.mem.num_pages = size >> PAGE_SHIFT;
nouveau_bo_placement_set(nvbo, flags, 0); nouveau_bo_placement_set(nvbo, flags, 0);
acc_size = ttm_bo_dma_acc_size(&dev_priv->ttm.bdev, size,
sizeof(struct nouveau_bo));
ret = ttm_bo_init(&dev_priv->ttm.bdev, &nvbo->bo, size, ret = ttm_bo_init(&dev_priv->ttm.bdev, &nvbo->bo, size,
ttm_bo_type_device, &nvbo->placement, ttm_bo_type_device, &nvbo->placement,
align >> PAGE_SHIFT, 0, false, NULL, size, align >> PAGE_SHIFT, 0, false, NULL, acc_size,
nouveau_bo_del_ttm); nouveau_bo_del_ttm);
if (ret) { if (ret) {
/* ttm will call nouveau_bo_del_ttm if it fails.. */ /* ttm will call nouveau_bo_del_ttm if it fails.. */
......
...@@ -95,6 +95,7 @@ int radeon_bo_create(struct radeon_device *rdev, ...@@ -95,6 +95,7 @@ int radeon_bo_create(struct radeon_device *rdev,
enum ttm_bo_type type; enum ttm_bo_type type;
unsigned long page_align = roundup(byte_align, PAGE_SIZE) >> PAGE_SHIFT; unsigned long page_align = roundup(byte_align, PAGE_SIZE) >> PAGE_SHIFT;
unsigned long max_size = 0; unsigned long max_size = 0;
size_t acc_size;
int r; int r;
size = ALIGN(size, PAGE_SIZE); size = ALIGN(size, PAGE_SIZE);
...@@ -117,6 +118,9 @@ int radeon_bo_create(struct radeon_device *rdev, ...@@ -117,6 +118,9 @@ int radeon_bo_create(struct radeon_device *rdev,
return -ENOMEM; return -ENOMEM;
} }
acc_size = ttm_bo_dma_acc_size(&rdev->mman.bdev, size,
sizeof(struct radeon_bo));
retry: retry:
bo = kzalloc(sizeof(struct radeon_bo), GFP_KERNEL); bo = kzalloc(sizeof(struct radeon_bo), GFP_KERNEL);
if (bo == NULL) if (bo == NULL)
...@@ -134,8 +138,8 @@ int radeon_bo_create(struct radeon_device *rdev, ...@@ -134,8 +138,8 @@ int radeon_bo_create(struct radeon_device *rdev,
/* Kernel allocation are uninterruptible */ /* Kernel allocation are uninterruptible */
mutex_lock(&rdev->vram_mutex); mutex_lock(&rdev->vram_mutex);
r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type, r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type,
&bo->placement, page_align, 0, !kernel, NULL, size, &bo->placement, page_align, 0, !kernel, NULL,
&radeon_ttm_bo_destroy); acc_size, &radeon_ttm_bo_destroy);
mutex_unlock(&rdev->vram_mutex); mutex_unlock(&rdev->vram_mutex);
if (unlikely(r != 0)) { if (unlikely(r != 0)) {
if (r != -ERESTARTSYS) { if (r != -ERESTARTSYS) {
......
...@@ -137,6 +137,7 @@ static void ttm_bo_release_list(struct kref *list_kref) ...@@ -137,6 +137,7 @@ static void ttm_bo_release_list(struct kref *list_kref)
struct ttm_buffer_object *bo = struct ttm_buffer_object *bo =
container_of(list_kref, struct ttm_buffer_object, list_kref); container_of(list_kref, struct ttm_buffer_object, list_kref);
struct ttm_bo_device *bdev = bo->bdev; struct ttm_bo_device *bdev = bo->bdev;
size_t acc_size = bo->acc_size;
BUG_ON(atomic_read(&bo->list_kref.refcount)); BUG_ON(atomic_read(&bo->list_kref.refcount));
BUG_ON(atomic_read(&bo->kref.refcount)); BUG_ON(atomic_read(&bo->kref.refcount));
...@@ -152,9 +153,9 @@ static void ttm_bo_release_list(struct kref *list_kref) ...@@ -152,9 +153,9 @@ static void ttm_bo_release_list(struct kref *list_kref)
if (bo->destroy) if (bo->destroy)
bo->destroy(bo); bo->destroy(bo);
else { else {
ttm_mem_global_free(bdev->glob->mem_glob, bo->acc_size);
kfree(bo); kfree(bo);
} }
ttm_mem_global_free(bdev->glob->mem_glob, acc_size);
} }
int ttm_bo_wait_unreserved(struct ttm_buffer_object *bo, bool interruptible) int ttm_bo_wait_unreserved(struct ttm_buffer_object *bo, bool interruptible)
...@@ -1157,6 +1158,17 @@ int ttm_bo_init(struct ttm_bo_device *bdev, ...@@ -1157,6 +1158,17 @@ int ttm_bo_init(struct ttm_bo_device *bdev,
{ {
int ret = 0; int ret = 0;
unsigned long num_pages; unsigned long num_pages;
struct ttm_mem_global *mem_glob = bdev->glob->mem_glob;
ret = ttm_mem_global_alloc(mem_glob, acc_size, false, false);
if (ret) {
printk(KERN_ERR TTM_PFX "Out of kernel memory.\n");
if (destroy)
(*destroy)(bo);
else
kfree(bo);
return -ENOMEM;
}
size += buffer_start & ~PAGE_MASK; size += buffer_start & ~PAGE_MASK;
num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
...@@ -1227,14 +1239,34 @@ int ttm_bo_init(struct ttm_bo_device *bdev, ...@@ -1227,14 +1239,34 @@ int ttm_bo_init(struct ttm_bo_device *bdev,
} }
EXPORT_SYMBOL(ttm_bo_init); EXPORT_SYMBOL(ttm_bo_init);
static inline size_t ttm_bo_size(struct ttm_bo_global *glob, size_t ttm_bo_acc_size(struct ttm_bo_device *bdev,
unsigned long num_pages) unsigned long bo_size,
unsigned struct_size)
{ {
size_t page_array_size = (num_pages * sizeof(void *) + PAGE_SIZE - 1) & unsigned npages = (PAGE_ALIGN(bo_size)) >> PAGE_SHIFT;
PAGE_MASK; size_t size = 0;
return glob->ttm_bo_size + 2 * page_array_size; size += ttm_round_pot(struct_size);
size += PAGE_ALIGN(npages * sizeof(void *));
size += ttm_round_pot(sizeof(struct ttm_tt));
return size;
} }
EXPORT_SYMBOL(ttm_bo_acc_size);
size_t ttm_bo_dma_acc_size(struct ttm_bo_device *bdev,
unsigned long bo_size,
unsigned struct_size)
{
unsigned npages = (PAGE_ALIGN(bo_size)) >> PAGE_SHIFT;
size_t size = 0;
size += ttm_round_pot(struct_size);
size += PAGE_ALIGN(npages * sizeof(void *));
size += PAGE_ALIGN(npages * sizeof(dma_addr_t));
size += ttm_round_pot(sizeof(struct ttm_dma_tt));
return size;
}
EXPORT_SYMBOL(ttm_bo_dma_acc_size);
int ttm_bo_create(struct ttm_bo_device *bdev, int ttm_bo_create(struct ttm_bo_device *bdev,
unsigned long size, unsigned long size,
...@@ -1248,10 +1280,10 @@ int ttm_bo_create(struct ttm_bo_device *bdev, ...@@ -1248,10 +1280,10 @@ int ttm_bo_create(struct ttm_bo_device *bdev,
{ {
struct ttm_buffer_object *bo; struct ttm_buffer_object *bo;
struct ttm_mem_global *mem_glob = bdev->glob->mem_glob; struct ttm_mem_global *mem_glob = bdev->glob->mem_glob;
size_t acc_size;
int ret; int ret;
size_t acc_size = acc_size = ttm_bo_acc_size(bdev, size, sizeof(struct ttm_buffer_object));
ttm_bo_size(bdev->glob, (size + PAGE_SIZE - 1) >> PAGE_SHIFT);
ret = ttm_mem_global_alloc(mem_glob, acc_size, false, false); ret = ttm_mem_global_alloc(mem_glob, acc_size, false, false);
if (unlikely(ret != 0)) if (unlikely(ret != 0))
return ret; return ret;
...@@ -1437,10 +1469,6 @@ int ttm_bo_global_init(struct drm_global_reference *ref) ...@@ -1437,10 +1469,6 @@ int ttm_bo_global_init(struct drm_global_reference *ref)
goto out_no_shrink; goto out_no_shrink;
} }
glob->ttm_bo_extra_size = ttm_round_pot(sizeof(struct ttm_tt));
glob->ttm_bo_size = glob->ttm_bo_extra_size +
ttm_round_pot(sizeof(struct ttm_buffer_object));
atomic_set(&glob->bo_count, 0); atomic_set(&glob->bo_count, 0);
ret = kobject_init_and_add( ret = kobject_init_and_add(
......
...@@ -445,6 +445,7 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo, ...@@ -445,6 +445,7 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo,
kref_init(&fbo->list_kref); kref_init(&fbo->list_kref);
kref_init(&fbo->kref); kref_init(&fbo->kref);
fbo->destroy = &ttm_transfered_destroy; fbo->destroy = &ttm_transfered_destroy;
fbo->acc_size = 0;
*new_obj = fbo; *new_obj = fbo;
return 0; return 0;
......
...@@ -1517,29 +1517,10 @@ int vmw_surface_check(struct vmw_private *dev_priv, ...@@ -1517,29 +1517,10 @@ int vmw_surface_check(struct vmw_private *dev_priv,
/** /**
* Buffer management. * Buffer management.
*/ */
static size_t vmw_dmabuf_acc_size(struct ttm_bo_global *glob,
unsigned long num_pages)
{
static size_t bo_user_size = ~0;
size_t page_array_size =
(num_pages * sizeof(void *) + PAGE_SIZE - 1) & PAGE_MASK;
if (unlikely(bo_user_size == ~0)) {
bo_user_size = glob->ttm_bo_extra_size +
ttm_round_pot(sizeof(struct vmw_dma_buffer));
}
return bo_user_size + page_array_size;
}
void vmw_dmabuf_bo_free(struct ttm_buffer_object *bo) void vmw_dmabuf_bo_free(struct ttm_buffer_object *bo)
{ {
struct vmw_dma_buffer *vmw_bo = vmw_dma_buffer(bo); struct vmw_dma_buffer *vmw_bo = vmw_dma_buffer(bo);
struct ttm_bo_global *glob = bo->glob;
ttm_mem_global_free(glob->mem_glob, bo->acc_size);
kfree(vmw_bo); kfree(vmw_bo);
} }
...@@ -1550,24 +1531,12 @@ int vmw_dmabuf_init(struct vmw_private *dev_priv, ...@@ -1550,24 +1531,12 @@ int vmw_dmabuf_init(struct vmw_private *dev_priv,
void (*bo_free) (struct ttm_buffer_object *bo)) void (*bo_free) (struct ttm_buffer_object *bo))
{ {
struct ttm_bo_device *bdev = &dev_priv->bdev; struct ttm_bo_device *bdev = &dev_priv->bdev;
struct ttm_mem_global *mem_glob = bdev->glob->mem_glob;
size_t acc_size; size_t acc_size;
int ret; int ret;
BUG_ON(!bo_free); BUG_ON(!bo_free);
acc_size = acc_size = ttm_bo_acc_size(bdev, size, sizeof(struct vmw_dma_buffer));
vmw_dmabuf_acc_size(bdev->glob,
(size + PAGE_SIZE - 1) >> PAGE_SHIFT);
ret = ttm_mem_global_alloc(mem_glob, acc_size, false, false);
if (unlikely(ret != 0)) {
/* we must free the bo here as
* ttm_buffer_object_init does so as well */
bo_free(&vmw_bo->base);
return ret;
}
memset(vmw_bo, 0, sizeof(*vmw_bo)); memset(vmw_bo, 0, sizeof(*vmw_bo));
INIT_LIST_HEAD(&vmw_bo->validate_list); INIT_LIST_HEAD(&vmw_bo->validate_list);
...@@ -1582,9 +1551,7 @@ int vmw_dmabuf_init(struct vmw_private *dev_priv, ...@@ -1582,9 +1551,7 @@ int vmw_dmabuf_init(struct vmw_private *dev_priv,
static void vmw_user_dmabuf_destroy(struct ttm_buffer_object *bo) static void vmw_user_dmabuf_destroy(struct ttm_buffer_object *bo)
{ {
struct vmw_user_dma_buffer *vmw_user_bo = vmw_user_dma_buffer(bo); struct vmw_user_dma_buffer *vmw_user_bo = vmw_user_dma_buffer(bo);
struct ttm_bo_global *glob = bo->glob;
ttm_mem_global_free(glob->mem_glob, bo->acc_size);
kfree(vmw_user_bo); kfree(vmw_user_bo);
} }
......
...@@ -429,9 +429,9 @@ extern void ttm_bo_unlock_delayed_workqueue(struct ttm_bo_device *bdev, ...@@ -429,9 +429,9 @@ extern void ttm_bo_unlock_delayed_workqueue(struct ttm_bo_device *bdev,
* -EBUSY if the buffer is busy and no_wait is true. * -EBUSY if the buffer is busy and no_wait is true.
* -ERESTARTSYS if interrupted by a signal. * -ERESTARTSYS if interrupted by a signal.
*/ */
extern int extern int
ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait); ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait);
/** /**
* ttm_bo_synccpu_write_release: * ttm_bo_synccpu_write_release:
* *
...@@ -441,6 +441,22 @@ ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait); ...@@ -441,6 +441,22 @@ ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait);
*/ */
extern void ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo); extern void ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo);
/**
* ttm_bo_acc_size
*
* @bdev: Pointer to a ttm_bo_device struct.
* @bo_size: size of the buffer object in byte.
* @struct_size: size of the structure holding buffer object datas
*
* Returns size to account for a buffer object
*/
size_t ttm_bo_acc_size(struct ttm_bo_device *bdev,
unsigned long bo_size,
unsigned struct_size);
size_t ttm_bo_dma_acc_size(struct ttm_bo_device *bdev,
unsigned long bo_size,
unsigned struct_size);
/** /**
* ttm_bo_init * ttm_bo_init
* *
...@@ -488,6 +504,7 @@ extern int ttm_bo_init(struct ttm_bo_device *bdev, ...@@ -488,6 +504,7 @@ extern int ttm_bo_init(struct ttm_bo_device *bdev,
struct file *persistent_swap_storage, struct file *persistent_swap_storage,
size_t acc_size, size_t acc_size,
void (*destroy) (struct ttm_buffer_object *)); void (*destroy) (struct ttm_buffer_object *));
/** /**
* ttm_bo_synccpu_object_init * ttm_bo_synccpu_object_init
* *
......
...@@ -469,9 +469,6 @@ struct ttm_bo_global_ref { ...@@ -469,9 +469,6 @@ struct ttm_bo_global_ref {
* @dummy_read_page: Pointer to a dummy page used for mapping requests * @dummy_read_page: Pointer to a dummy page used for mapping requests
* of unpopulated pages. * of unpopulated pages.
* @shrink: A shrink callback object used for buffer object swap. * @shrink: A shrink callback object used for buffer object swap.
* @ttm_bo_extra_size: Extra size (sizeof(struct ttm_buffer_object) excluded)
* used by a buffer object. This is excluding page arrays and backing pages.
* @ttm_bo_size: This is @ttm_bo_extra_size + sizeof(struct ttm_buffer_object).
* @device_list_mutex: Mutex protecting the device list. * @device_list_mutex: Mutex protecting the device list.
* This mutex is held while traversing the device list for pm options. * This mutex is held while traversing the device list for pm options.
* @lru_lock: Spinlock protecting the bo subsystem lru lists. * @lru_lock: Spinlock protecting the bo subsystem lru lists.
...@@ -489,8 +486,6 @@ struct ttm_bo_global { ...@@ -489,8 +486,6 @@ struct ttm_bo_global {
struct ttm_mem_global *mem_glob; struct ttm_mem_global *mem_glob;
struct page *dummy_read_page; struct page *dummy_read_page;
struct ttm_mem_shrink shrink; struct ttm_mem_shrink shrink;
size_t ttm_bo_extra_size;
size_t ttm_bo_size;
struct mutex device_list_mutex; struct mutex device_list_mutex;
spinlock_t lru_lock; spinlock_t lru_lock;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment