mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Makes creation of shm pools more flexible by adding new more primitive functions to allocate a shm pool. This makes it easier to add driver specific shm pool management. Signed-off-by: Jens Wiklander <jens.wiklander@linaro.org> Signed-off-by: Volodymyr Babchuk <vlad.babchuk@gmail.com>
		
			
				
	
	
		
			195 lines
		
	
	
	
		
			4.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			195 lines
		
	
	
	
		
			4.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright (c) 2015, Linaro Limited
 | 
						|
 *
 | 
						|
 * This software is licensed under the terms of the GNU General Public
 | 
						|
 * License version 2, as published by the Free Software Foundation, and
 | 
						|
 * may be copied, distributed, and modified under those terms.
 | 
						|
 *
 | 
						|
 * This program is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
 * GNU General Public License for more details.
 | 
						|
 *
 | 
						|
 */
 | 
						|
#include <linux/device.h>
 | 
						|
#include <linux/dma-buf.h>
 | 
						|
#include <linux/genalloc.h>
 | 
						|
#include <linux/slab.h>
 | 
						|
#include <linux/tee_drv.h>
 | 
						|
#include "tee_private.h"
 | 
						|
 | 
						|
static int pool_op_gen_alloc(struct tee_shm_pool_mgr *poolm,
 | 
						|
			     struct tee_shm *shm, size_t size)
 | 
						|
{
 | 
						|
	unsigned long va;
 | 
						|
	struct gen_pool *genpool = poolm->private_data;
 | 
						|
	size_t s = roundup(size, 1 << genpool->min_alloc_order);
 | 
						|
 | 
						|
	va = gen_pool_alloc(genpool, s);
 | 
						|
	if (!va)
 | 
						|
		return -ENOMEM;
 | 
						|
 | 
						|
	memset((void *)va, 0, s);
 | 
						|
	shm->kaddr = (void *)va;
 | 
						|
	shm->paddr = gen_pool_virt_to_phys(genpool, va);
 | 
						|
	shm->size = s;
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void pool_op_gen_free(struct tee_shm_pool_mgr *poolm,
 | 
						|
			     struct tee_shm *shm)
 | 
						|
{
 | 
						|
	gen_pool_free(poolm->private_data, (unsigned long)shm->kaddr,
 | 
						|
		      shm->size);
 | 
						|
	shm->kaddr = NULL;
 | 
						|
}
 | 
						|
 | 
						|
static void pool_op_gen_destroy_poolmgr(struct tee_shm_pool_mgr *poolm)
 | 
						|
{
 | 
						|
	gen_pool_destroy(poolm->private_data);
 | 
						|
	kfree(poolm);
 | 
						|
}
 | 
						|
 | 
						|
static const struct tee_shm_pool_mgr_ops pool_ops_generic = {
 | 
						|
	.alloc = pool_op_gen_alloc,
 | 
						|
	.free = pool_op_gen_free,
 | 
						|
	.destroy_poolmgr = pool_op_gen_destroy_poolmgr,
 | 
						|
};
 | 
						|
 | 
						|
/**
 | 
						|
 * tee_shm_pool_alloc_res_mem() - Create a shared memory pool from reserved
 | 
						|
 * memory range
 | 
						|
 * @priv_info:	Information for driver private shared memory pool
 | 
						|
 * @dmabuf_info: Information for dma-buf shared memory pool
 | 
						|
 *
 | 
						|
 * Start and end of pools will must be page aligned.
 | 
						|
 *
 | 
						|
 * Allocation with the flag TEE_SHM_DMA_BUF set will use the range supplied
 | 
						|
 * in @dmabuf, others will use the range provided by @priv.
 | 
						|
 *
 | 
						|
 * @returns pointer to a 'struct tee_shm_pool' or an ERR_PTR on failure.
 | 
						|
 */
 | 
						|
struct tee_shm_pool *
 | 
						|
tee_shm_pool_alloc_res_mem(struct tee_shm_pool_mem_info *priv_info,
 | 
						|
			   struct tee_shm_pool_mem_info *dmabuf_info)
 | 
						|
{
 | 
						|
	struct tee_shm_pool_mgr *priv_mgr;
 | 
						|
	struct tee_shm_pool_mgr *dmabuf_mgr;
 | 
						|
	void *rc;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Create the pool for driver private shared memory
 | 
						|
	 */
 | 
						|
	rc = tee_shm_pool_mgr_alloc_res_mem(priv_info->vaddr, priv_info->paddr,
 | 
						|
					    priv_info->size,
 | 
						|
					    3 /* 8 byte aligned */);
 | 
						|
	if (IS_ERR(rc))
 | 
						|
		return rc;
 | 
						|
	priv_mgr = rc;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Create the pool for dma_buf shared memory
 | 
						|
	 */
 | 
						|
	rc = tee_shm_pool_mgr_alloc_res_mem(dmabuf_info->vaddr,
 | 
						|
					    dmabuf_info->paddr,
 | 
						|
					    dmabuf_info->size, PAGE_SHIFT);
 | 
						|
	if (IS_ERR(rc))
 | 
						|
		goto err_free_priv_mgr;
 | 
						|
	dmabuf_mgr = rc;
 | 
						|
 | 
						|
	rc = tee_shm_pool_alloc(priv_mgr, dmabuf_mgr);
 | 
						|
	if (IS_ERR(rc))
 | 
						|
		goto err_free_dmabuf_mgr;
 | 
						|
 | 
						|
	return rc;
 | 
						|
 | 
						|
err_free_dmabuf_mgr:
 | 
						|
	tee_shm_pool_mgr_destroy(dmabuf_mgr);
 | 
						|
err_free_priv_mgr:
 | 
						|
	tee_shm_pool_mgr_destroy(priv_mgr);
 | 
						|
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(tee_shm_pool_alloc_res_mem);
 | 
						|
 | 
						|
struct tee_shm_pool_mgr *tee_shm_pool_mgr_alloc_res_mem(unsigned long vaddr,
 | 
						|
							phys_addr_t paddr,
 | 
						|
							size_t size,
 | 
						|
							int min_alloc_order)
 | 
						|
{
 | 
						|
	const size_t page_mask = PAGE_SIZE - 1;
 | 
						|
	struct tee_shm_pool_mgr *mgr;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	/* Start and end must be page aligned */
 | 
						|
	if (vaddr & page_mask || paddr & page_mask || size & page_mask)
 | 
						|
		return ERR_PTR(-EINVAL);
 | 
						|
 | 
						|
	mgr = kzalloc(sizeof(*mgr), GFP_KERNEL);
 | 
						|
	if (!mgr)
 | 
						|
		return ERR_PTR(-ENOMEM);
 | 
						|
 | 
						|
	mgr->private_data = gen_pool_create(min_alloc_order, -1);
 | 
						|
	if (!mgr->private_data) {
 | 
						|
		rc = -ENOMEM;
 | 
						|
		goto err;
 | 
						|
	}
 | 
						|
 | 
						|
	gen_pool_set_algo(mgr->private_data, gen_pool_best_fit, NULL);
 | 
						|
	rc = gen_pool_add_virt(mgr->private_data, vaddr, paddr, size, -1);
 | 
						|
	if (rc) {
 | 
						|
		gen_pool_destroy(mgr->private_data);
 | 
						|
		goto err;
 | 
						|
	}
 | 
						|
 | 
						|
	mgr->ops = &pool_ops_generic;
 | 
						|
 | 
						|
	return mgr;
 | 
						|
err:
 | 
						|
	kfree(mgr);
 | 
						|
 | 
						|
	return ERR_PTR(rc);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(tee_shm_pool_mgr_alloc_res_mem);
 | 
						|
 | 
						|
static bool check_mgr_ops(struct tee_shm_pool_mgr *mgr)
 | 
						|
{
 | 
						|
	return mgr && mgr->ops && mgr->ops->alloc && mgr->ops->free &&
 | 
						|
		mgr->ops->destroy_poolmgr;
 | 
						|
}
 | 
						|
 | 
						|
struct tee_shm_pool *tee_shm_pool_alloc(struct tee_shm_pool_mgr *priv_mgr,
 | 
						|
					struct tee_shm_pool_mgr *dmabuf_mgr)
 | 
						|
{
 | 
						|
	struct tee_shm_pool *pool;
 | 
						|
 | 
						|
	if (!check_mgr_ops(priv_mgr) || !check_mgr_ops(dmabuf_mgr))
 | 
						|
		return ERR_PTR(-EINVAL);
 | 
						|
 | 
						|
	pool = kzalloc(sizeof(*pool), GFP_KERNEL);
 | 
						|
	if (!pool)
 | 
						|
		return ERR_PTR(-ENOMEM);
 | 
						|
 | 
						|
	pool->private_mgr = priv_mgr;
 | 
						|
	pool->dma_buf_mgr = dmabuf_mgr;
 | 
						|
 | 
						|
	return pool;
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(tee_shm_pool_alloc);
 | 
						|
 | 
						|
/**
 | 
						|
 * tee_shm_pool_free() - Free a shared memory pool
 | 
						|
 * @pool:	The shared memory pool to free
 | 
						|
 *
 | 
						|
 * There must be no remaining shared memory allocated from this pool when
 | 
						|
 * this function is called.
 | 
						|
 */
 | 
						|
void tee_shm_pool_free(struct tee_shm_pool *pool)
 | 
						|
{
 | 
						|
	if (pool->private_mgr)
 | 
						|
		tee_shm_pool_mgr_destroy(pool->private_mgr);
 | 
						|
	if (pool->dma_buf_mgr)
 | 
						|
		tee_shm_pool_mgr_destroy(pool->dma_buf_mgr);
 | 
						|
	kfree(pool);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL_GPL(tee_shm_pool_free);
 |