Commit b3553065 authored by Alan Cox's avatar Alan Cox Committed by Greg Kroah-Hartman

atomisp: remove most of the uses of atomisp_kernel_malloc

They can be replaced by kmalloc. There are a few that do need to pick kmalloc
or vmalloc. Those we leave for the moment.
Signed-off-by: default avatarAlan Cox <alan@linux.intel.com>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
parent 5eb32874
......@@ -1669,11 +1669,7 @@ int atomisp_alloc_dis_coef_buf(struct atomisp_sub_device *asd)
int atomisp_alloc_metadata_output_buf(struct atomisp_sub_device *asd)
{
#ifndef ISP2401
int i; /* Coverity CID 298003 - index var must be signed */
#else
int i;
#endif
/* We allocate the cpu-side buffer used for communication with user
* space */
......
......@@ -724,8 +724,8 @@ static int alloc_private_pages(struct hmm_buffer_object *bo,
pgnr = bo->pgnr;
bo->page_obj = atomisp_kernel_malloc(
sizeof(struct hmm_page_object) * pgnr);
bo->page_obj = kmalloc(sizeof(struct hmm_page_object) * pgnr,
GFP_KERNEL);
if (unlikely(!bo->page_obj)) {
dev_err(atomisp_dev, "out of memory for bo->page_obj\n");
return -ENOMEM;
......@@ -859,7 +859,7 @@ static int alloc_private_pages(struct hmm_buffer_object *bo,
alloc_pgnr = i;
free_private_bo_pages(bo, dypool, repool, alloc_pgnr);
atomisp_kernel_free(bo->page_obj);
kfree(bo->page_obj);
return -ENOMEM;
}
......@@ -870,7 +870,7 @@ static void free_private_pages(struct hmm_buffer_object *bo,
{
free_private_bo_pages(bo, dypool, repool, bo->pgnr);
atomisp_kernel_free(bo->page_obj);
kfree(bo->page_obj);
}
/*
......@@ -989,17 +989,17 @@ static int alloc_user_pages(struct hmm_buffer_object *bo,
struct vm_area_struct *vma;
struct page **pages;
pages = atomisp_kernel_malloc(sizeof(struct page *) * bo->pgnr);
pages = kmalloc(sizeof(struct page *) * bo->pgnr, GFP_KERNEL);
if (unlikely(!pages)) {
dev_err(atomisp_dev, "out of memory for pages...\n");
return -ENOMEM;
}
bo->page_obj = atomisp_kernel_malloc(
sizeof(struct hmm_page_object) * bo->pgnr);
bo->page_obj = kmalloc(sizeof(struct hmm_page_object) * bo->pgnr,
GFP_KERNEL);
if (unlikely(!bo->page_obj)) {
dev_err(atomisp_dev, "out of memory for bo->page_obj...\n");
atomisp_kernel_free(pages);
kfree(pages);
return -ENOMEM;
}
......@@ -1009,8 +1009,8 @@ static int alloc_user_pages(struct hmm_buffer_object *bo,
up_read(&current->mm->mmap_sem);
if (vma == NULL) {
dev_err(atomisp_dev, "find_vma failed\n");
atomisp_kernel_free(bo->page_obj);
atomisp_kernel_free(pages);
kfree(bo->page_obj);
kfree(pages);
mutex_lock(&bo->mutex);
return -EFAULT;
}
......@@ -1050,15 +1050,15 @@ static int alloc_user_pages(struct hmm_buffer_object *bo,
bo->page_obj[i].type = HMM_PAGE_TYPE_GENERAL;
}
hmm_mem_stat.usr_size += bo->pgnr;
atomisp_kernel_free(pages);
kfree(pages);
return 0;
out_of_mem:
for (i = 0; i < page_nr; i++)
put_page(pages[i]);
atomisp_kernel_free(pages);
atomisp_kernel_free(bo->page_obj);
kfree(pages);
kfree(bo->page_obj);
return -ENOMEM;
}
......@@ -1071,7 +1071,7 @@ static void free_user_pages(struct hmm_buffer_object *bo)
put_page(bo->page_obj[i].page);
hmm_mem_stat.usr_size -= bo->pgnr;
atomisp_kernel_free(bo->page_obj);
kfree(bo->page_obj);
}
/*
......@@ -1362,7 +1362,7 @@ void *hmm_bo_vmap(struct hmm_buffer_object *bo, bool cached)
bo->status &= ~(HMM_BO_VMAPED | HMM_BO_VMAPED_CACHED);
}
pages = atomisp_kernel_malloc(sizeof(*pages) * bo->pgnr);
pages = kmalloc(sizeof(*pages) * bo->pgnr, GFP_KERNEL);
if (unlikely(!pages)) {
mutex_unlock(&bo->mutex);
dev_err(atomisp_dev, "out of memory for pages...\n");
......@@ -1375,14 +1375,14 @@ void *hmm_bo_vmap(struct hmm_buffer_object *bo, bool cached)
bo->vmap_addr = vmap(pages, bo->pgnr, VM_MAP,
cached ? PAGE_KERNEL : PAGE_KERNEL_NOCACHE);
if (unlikely(!bo->vmap_addr)) {
atomisp_kernel_free(pages);
kfree(pages);
mutex_unlock(&bo->mutex);
dev_err(atomisp_dev, "vmap failed...\n");
return NULL;
}
bo->status |= (cached ? HMM_BO_VMAPED_CACHED : HMM_BO_VMAPED);
atomisp_kernel_free(pages);
kfree(pages);
mutex_unlock(&bo->mutex);
return bo->vmap_addr;
......
......@@ -149,8 +149,8 @@ static int hmm_dynamic_pool_init(void **pool, unsigned int pool_size)
if (pool_size == 0)
return 0;
dypool_info = atomisp_kernel_malloc(
sizeof(struct hmm_dynamic_pool_info));
dypool_info = kmalloc(sizeof(struct hmm_dynamic_pool_info),
GFP_KERNEL);
if (unlikely(!dypool_info)) {
dev_err(atomisp_dev, "out of memory for repool_info.\n");
return -ENOMEM;
......@@ -160,7 +160,7 @@ static int hmm_dynamic_pool_init(void **pool, unsigned int pool_size)
sizeof(struct hmm_page), 0,
SLAB_HWCACHE_ALIGN, NULL);
if (!dypool_info->pgptr_cache) {
atomisp_kernel_free(dypool_info);
kfree(dypool_info);
return -ENOMEM;
}
......@@ -217,7 +217,7 @@ static void hmm_dynamic_pool_exit(void **pool)
kmem_cache_destroy(dypool_info->pgptr_cache);
atomisp_kernel_free(dypool_info);
kfree(dypool_info);
*pool = NULL;
}
......
......@@ -89,18 +89,18 @@ static int hmm_reserved_pool_setup(struct hmm_reserved_pool_info **repool_info,
{
struct hmm_reserved_pool_info *pool_info;
pool_info = atomisp_kernel_malloc(
sizeof(struct hmm_reserved_pool_info));
pool_info = kmalloc(sizeof(struct hmm_reserved_pool_info),
GFP_KERNEL);
if (unlikely(!pool_info)) {
dev_err(atomisp_dev, "out of memory for repool_info.\n");
return -ENOMEM;
}
pool_info->pages = atomisp_kernel_malloc(
sizeof(struct page *) * pool_size);
pool_info->pages = kmalloc(sizeof(struct page *) * pool_size,
GFP_KERNEL);
if (unlikely(!pool_info->pages)) {
dev_err(atomisp_dev, "out of memory for repool_info->pages.\n");
atomisp_kernel_free(pool_info);
kfree(pool_info);
return -ENOMEM;
}
......@@ -233,8 +233,8 @@ static void hmm_reserved_pool_exit(void **pool)
__free_pages(repool_info->pages[i], 0);
}
atomisp_kernel_free(repool_info->pages);
atomisp_kernel_free(repool_info);
kfree(repool_info->pages);
kfree(repool_info);
*pool = NULL;
}
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment