forked from mirrors/linux
		
	 9bd405c48b
			
		
	
	
		9bd405c48b
		
	
	
	
	
		
			
			Align the end size to cache boundary size in ax45mp_dma_cache_wback()
callback likewise done in ax45mp_dma_cache_inv() callback.
Additionally return early in case of start == end.
Fixes: d34599bcd2 ("cache: Add L2 cache management for Andes AX45MP RISC-V core")
Reported-by: Pavel Machek <pavel@denx.de>
Link: https://lore.kernel.org/cip-dev/ZYsdKDiw7G+kxQ3m@duo.ucw.cz/
Signed-off-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com>
Signed-off-by: Conor Dooley <conor.dooley@microchip.com>
		
	
			
		
			
				
	
	
		
			217 lines
		
	
	
	
		
			5.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			217 lines
		
	
	
	
		
			5.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * non-coherent cache functions for Andes AX45MP
 | |
|  *
 | |
|  * Copyright (C) 2023 Renesas Electronics Corp.
 | |
|  */
 | |
| 
 | |
| #include <linux/cacheflush.h>
 | |
| #include <linux/cacheinfo.h>
 | |
| #include <linux/dma-direction.h>
 | |
| #include <linux/of_address.h>
 | |
| #include <linux/of_platform.h>
 | |
| 
 | |
| #include <asm/dma-noncoherent.h>
 | |
| 
 | |
| /* L2 cache registers */
 | |
| #define AX45MP_L2C_REG_CTL_OFFSET		0x8
 | |
| 
 | |
| #define AX45MP_L2C_REG_C0_CMD_OFFSET		0x40
 | |
| #define AX45MP_L2C_REG_C0_ACC_OFFSET		0x48
 | |
| #define AX45MP_L2C_REG_STATUS_OFFSET		0x80
 | |
| 
 | |
| /* D-cache operation */
 | |
| #define AX45MP_CCTL_L1D_VA_INVAL		0 /* Invalidate an L1 cache entry */
 | |
| #define AX45MP_CCTL_L1D_VA_WB			1 /* Write-back an L1 cache entry */
 | |
| 
 | |
| /* L2 CCTL status */
 | |
| #define AX45MP_CCTL_L2_STATUS_IDLE		0
 | |
| 
 | |
| /* L2 CCTL status cores mask */
 | |
| #define AX45MP_CCTL_L2_STATUS_C0_MASK		0xf
 | |
| 
 | |
| /* L2 cache operation */
 | |
| #define AX45MP_CCTL_L2_PA_INVAL			0x8 /* Invalidate an L2 cache entry */
 | |
| #define AX45MP_CCTL_L2_PA_WB			0x9 /* Write-back an L2 cache entry */
 | |
| 
 | |
| #define AX45MP_L2C_REG_PER_CORE_OFFSET		0x10
 | |
| #define AX45MP_CCTL_L2_STATUS_PER_CORE_OFFSET	4
 | |
| 
 | |
| #define AX45MP_L2C_REG_CN_CMD_OFFSET(n)	\
 | |
| 	(AX45MP_L2C_REG_C0_CMD_OFFSET + ((n) * AX45MP_L2C_REG_PER_CORE_OFFSET))
 | |
| #define AX45MP_L2C_REG_CN_ACC_OFFSET(n)	\
 | |
| 	(AX45MP_L2C_REG_C0_ACC_OFFSET + ((n) * AX45MP_L2C_REG_PER_CORE_OFFSET))
 | |
| #define AX45MP_CCTL_L2_STATUS_CN_MASK(n)	\
 | |
| 	(AX45MP_CCTL_L2_STATUS_C0_MASK << ((n) * AX45MP_CCTL_L2_STATUS_PER_CORE_OFFSET))
 | |
| 
 | |
| #define AX45MP_CCTL_REG_UCCTLBEGINADDR_NUM	0x80b
 | |
| #define AX45MP_CCTL_REG_UCCTLCOMMAND_NUM	0x80c
 | |
| 
 | |
| #define AX45MP_CACHE_LINE_SIZE			64
 | |
| 
 | |
| struct ax45mp_priv {
 | |
| 	void __iomem *l2c_base;
 | |
| 	u32 ax45mp_cache_line_size;
 | |
| };
 | |
| 
 | |
| static struct ax45mp_priv ax45mp_priv;
 | |
| 
 | |
| /* L2 Cache operations */
 | |
| static inline uint32_t ax45mp_cpu_l2c_get_cctl_status(void)
 | |
| {
 | |
| 	return readl(ax45mp_priv.l2c_base + AX45MP_L2C_REG_STATUS_OFFSET);
 | |
| }
 | |
| 
 | |
| static void ax45mp_cpu_cache_operation(unsigned long start, unsigned long end,
 | |
| 				       unsigned int l1_op, unsigned int l2_op)
 | |
| {
 | |
| 	unsigned long line_size = ax45mp_priv.ax45mp_cache_line_size;
 | |
| 	void __iomem *base = ax45mp_priv.l2c_base;
 | |
| 	int mhartid = smp_processor_id();
 | |
| 	unsigned long pa;
 | |
| 
 | |
| 	while (end > start) {
 | |
| 		csr_write(AX45MP_CCTL_REG_UCCTLBEGINADDR_NUM, start);
 | |
| 		csr_write(AX45MP_CCTL_REG_UCCTLCOMMAND_NUM, l1_op);
 | |
| 
 | |
| 		pa = virt_to_phys((void *)start);
 | |
| 		writel(pa, base + AX45MP_L2C_REG_CN_ACC_OFFSET(mhartid));
 | |
| 		writel(l2_op, base + AX45MP_L2C_REG_CN_CMD_OFFSET(mhartid));
 | |
| 		while ((ax45mp_cpu_l2c_get_cctl_status() &
 | |
| 			AX45MP_CCTL_L2_STATUS_CN_MASK(mhartid)) !=
 | |
| 			AX45MP_CCTL_L2_STATUS_IDLE)
 | |
| 			;
 | |
| 
 | |
| 		start += line_size;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Write-back L1 and L2 cache entry */
 | |
| static inline void ax45mp_cpu_dcache_wb_range(unsigned long start, unsigned long end)
 | |
| {
 | |
| 	ax45mp_cpu_cache_operation(start, end, AX45MP_CCTL_L1D_VA_WB,
 | |
| 				   AX45MP_CCTL_L2_PA_WB);
 | |
| }
 | |
| 
 | |
| /* Invalidate the L1 and L2 cache entry */
 | |
| static inline void ax45mp_cpu_dcache_inval_range(unsigned long start, unsigned long end)
 | |
| {
 | |
| 	ax45mp_cpu_cache_operation(start, end, AX45MP_CCTL_L1D_VA_INVAL,
 | |
| 				   AX45MP_CCTL_L2_PA_INVAL);
 | |
| }
 | |
| 
 | |
| static void ax45mp_dma_cache_inv(phys_addr_t paddr, size_t size)
 | |
| {
 | |
| 	unsigned long start = (unsigned long)phys_to_virt(paddr);
 | |
| 	unsigned long end = start + size;
 | |
| 	unsigned long line_size;
 | |
| 	unsigned long flags;
 | |
| 
 | |
| 	if (unlikely(start == end))
 | |
| 		return;
 | |
| 
 | |
| 	line_size = ax45mp_priv.ax45mp_cache_line_size;
 | |
| 
 | |
| 	start = start & (~(line_size - 1));
 | |
| 	end = ((end + line_size - 1) & (~(line_size - 1)));
 | |
| 
 | |
| 	local_irq_save(flags);
 | |
| 
 | |
| 	ax45mp_cpu_dcache_inval_range(start, end);
 | |
| 
 | |
| 	local_irq_restore(flags);
 | |
| }
 | |
| 
 | |
| static void ax45mp_dma_cache_wback(phys_addr_t paddr, size_t size)
 | |
| {
 | |
| 	unsigned long start = (unsigned long)phys_to_virt(paddr);
 | |
| 	unsigned long end = start + size;
 | |
| 	unsigned long line_size;
 | |
| 	unsigned long flags;
 | |
| 
 | |
| 	if (unlikely(start == end))
 | |
| 		return;
 | |
| 
 | |
| 	line_size = ax45mp_priv.ax45mp_cache_line_size;
 | |
| 	start = start & (~(line_size - 1));
 | |
| 	end = ((end + line_size - 1) & (~(line_size - 1)));
 | |
| 	local_irq_save(flags);
 | |
| 	ax45mp_cpu_dcache_wb_range(start, end);
 | |
| 	local_irq_restore(flags);
 | |
| }
 | |
| 
 | |
| static void ax45mp_dma_cache_wback_inv(phys_addr_t paddr, size_t size)
 | |
| {
 | |
| 	ax45mp_dma_cache_wback(paddr, size);
 | |
| 	ax45mp_dma_cache_inv(paddr, size);
 | |
| }
 | |
| 
 | |
| static int ax45mp_get_l2_line_size(struct device_node *np)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = of_property_read_u32(np, "cache-line-size", &ax45mp_priv.ax45mp_cache_line_size);
 | |
| 	if (ret) {
 | |
| 		pr_err("Failed to get cache-line-size, defaulting to 64 bytes\n");
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	if (ax45mp_priv.ax45mp_cache_line_size != AX45MP_CACHE_LINE_SIZE) {
 | |
| 		pr_err("Expected cache-line-size to be 64 bytes (found:%u)\n",
 | |
| 		       ax45mp_priv.ax45mp_cache_line_size);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static const struct riscv_nonstd_cache_ops ax45mp_cmo_ops __initdata = {
 | |
| 	.wback = &ax45mp_dma_cache_wback,
 | |
| 	.inv = &ax45mp_dma_cache_inv,
 | |
| 	.wback_inv = &ax45mp_dma_cache_wback_inv,
 | |
| };
 | |
| 
 | |
| static const struct of_device_id ax45mp_cache_ids[] = {
 | |
| 	{ .compatible = "andestech,ax45mp-cache" },
 | |
| 	{ /* sentinel */ }
 | |
| };
 | |
| 
 | |
| static int __init ax45mp_cache_init(void)
 | |
| {
 | |
| 	struct device_node *np;
 | |
| 	struct resource res;
 | |
| 	int ret;
 | |
| 
 | |
| 	np = of_find_matching_node(NULL, ax45mp_cache_ids);
 | |
| 	if (!of_device_is_available(np))
 | |
| 		return -ENODEV;
 | |
| 
 | |
| 	ret = of_address_to_resource(np, 0, &res);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	/*
 | |
| 	 * If IOCP is present on the Andes AX45MP core riscv_cbom_block_size
 | |
| 	 * will be 0 for sure, so we can definitely rely on it. If
 | |
| 	 * riscv_cbom_block_size = 0 we don't need to handle CMO using SW any
 | |
| 	 * more so we just return success here and only if its being set we
 | |
| 	 * continue further in the probe path.
 | |
| 	 */
 | |
| 	if (!riscv_cbom_block_size)
 | |
| 		return 0;
 | |
| 
 | |
| 	ax45mp_priv.l2c_base = ioremap(res.start, resource_size(&res));
 | |
| 	if (!ax45mp_priv.l2c_base)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	ret = ax45mp_get_l2_line_size(np);
 | |
| 	if (ret) {
 | |
| 		iounmap(ax45mp_priv.l2c_base);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	riscv_noncoherent_register_cache_ops(&ax45mp_cmo_ops);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| early_initcall(ax45mp_cache_init);
 |