forked from mirrors/linux
		
	 a814a14ea4
			
		
	
	
		a814a14ea4
		
	
	
	
	
		
			
			Get rid of boilerplate code by using module_platform_driver macro. Signed-off-by: Axel Lin <axel.lin@ingics.com> Acked-by: Jan Glauber <jglauber@cavium.com> Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org>
		
			
				
	
	
		
			349 lines
		
	
	
	
		
			8.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			349 lines
		
	
	
	
		
			8.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Driver for MMC and SSD cards for Cavium OCTEON SOCs.
 | |
|  *
 | |
|  * This file is subject to the terms and conditions of the GNU General Public
 | |
|  * License.  See the file "COPYING" in the main directory of this archive
 | |
|  * for more details.
 | |
|  *
 | |
|  * Copyright (C) 2012-2017 Cavium Inc.
 | |
|  */
 | |
| #include <linux/dma-mapping.h>
 | |
| #include <linux/gpio/consumer.h>
 | |
| #include <linux/interrupt.h>
 | |
| #include <linux/mmc/mmc.h>
 | |
| #include <linux/mmc/slot-gpio.h>
 | |
| #include <linux/module.h>
 | |
| #include <linux/of_platform.h>
 | |
| #include <asm/octeon/octeon.h>
 | |
| #include "cavium.h"
 | |
| 
 | |
| #define CVMX_MIO_BOOT_CTL CVMX_ADD_IO_SEG(0x00011800000000D0ull)
 | |
| 
 | |
| /*
 | |
|  * The l2c* functions below are used for the EMMC-17978 workaround.
 | |
|  *
 | |
|  * Due to a bug in the design of the MMC bus hardware, the 2nd to last
 | |
|  * cache block of a DMA read must be locked into the L2 Cache.
 | |
|  * Otherwise, data corruption may occur.
 | |
|  */
 | |
| static inline void *phys_to_ptr(u64 address)
 | |
| {
 | |
| 	return (void *)(address | (1ull << 63)); /* XKPHYS */
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Lock a single line into L2. The line is zeroed before locking
 | |
|  * to make sure no dram accesses are made.
 | |
|  */
 | |
| static void l2c_lock_line(u64 addr)
 | |
| {
 | |
| 	char *addr_ptr = phys_to_ptr(addr);
 | |
| 
 | |
| 	asm volatile (
 | |
| 		"cache 31, %[line]"	/* Unlock the line */
 | |
| 		::[line] "m" (*addr_ptr));
 | |
| }
 | |
| 
 | |
| /* Unlock a single line in the L2 cache. */
 | |
| static void l2c_unlock_line(u64 addr)
 | |
| {
 | |
| 	char *addr_ptr = phys_to_ptr(addr);
 | |
| 
 | |
| 	asm volatile (
 | |
| 		"cache 23, %[line]"	/* Unlock the line */
 | |
| 		::[line] "m" (*addr_ptr));
 | |
| }
 | |
| 
 | |
| /* Locks a memory region in the L2 cache. */
 | |
| static void l2c_lock_mem_region(u64 start, u64 len)
 | |
| {
 | |
| 	u64 end;
 | |
| 
 | |
| 	/* Round start/end to cache line boundaries */
 | |
| 	end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE);
 | |
| 	start = ALIGN(start, CVMX_CACHE_LINE_SIZE);
 | |
| 
 | |
| 	while (start <= end) {
 | |
| 		l2c_lock_line(start);
 | |
| 		start += CVMX_CACHE_LINE_SIZE;
 | |
| 	}
 | |
| 	asm volatile("sync");
 | |
| }
 | |
| 
 | |
| /* Unlock a memory region in the L2 cache. */
 | |
| static void l2c_unlock_mem_region(u64 start, u64 len)
 | |
| {
 | |
| 	u64 end;
 | |
| 
 | |
| 	/* Round start/end to cache line boundaries */
 | |
| 	end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE);
 | |
| 	start = ALIGN(start, CVMX_CACHE_LINE_SIZE);
 | |
| 
 | |
| 	while (start <= end) {
 | |
| 		l2c_unlock_line(start);
 | |
| 		start += CVMX_CACHE_LINE_SIZE;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_acquire_bus(struct cvm_mmc_host *host)
 | |
| {
 | |
| 	if (!host->has_ciu3) {
 | |
| 		down(&octeon_bootbus_sem);
 | |
| 		/* For CN70XX, switch the MMC controller onto the bus. */
 | |
| 		if (OCTEON_IS_MODEL(OCTEON_CN70XX))
 | |
| 			writeq(0, (void __iomem *)CVMX_MIO_BOOT_CTL);
 | |
| 	} else {
 | |
| 		down(&host->mmc_serializer);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_release_bus(struct cvm_mmc_host *host)
 | |
| {
 | |
| 	if (!host->has_ciu3)
 | |
| 		up(&octeon_bootbus_sem);
 | |
| 	else
 | |
| 		up(&host->mmc_serializer);
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_int_enable(struct cvm_mmc_host *host, u64 val)
 | |
| {
 | |
| 	writeq(val, host->base + MIO_EMM_INT(host));
 | |
| 	if (!host->has_ciu3)
 | |
| 		writeq(val, host->base + MIO_EMM_INT_EN(host));
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_set_shared_power(struct cvm_mmc_host *host, int dir)
 | |
| {
 | |
| 	if (dir == 0)
 | |
| 		if (!atomic_dec_return(&host->shared_power_users))
 | |
| 			gpiod_set_value_cansleep(host->global_pwr_gpiod, 0);
 | |
| 	if (dir == 1)
 | |
| 		if (atomic_inc_return(&host->shared_power_users) == 1)
 | |
| 			gpiod_set_value_cansleep(host->global_pwr_gpiod, 1);
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_dmar_fixup(struct cvm_mmc_host *host,
 | |
| 				  struct mmc_command *cmd,
 | |
| 				  struct mmc_data *data,
 | |
| 				  u64 addr)
 | |
| {
 | |
| 	if (cmd->opcode != MMC_WRITE_MULTIPLE_BLOCK)
 | |
| 		return;
 | |
| 	if (data->blksz * data->blocks <= 1024)
 | |
| 		return;
 | |
| 
 | |
| 	host->n_minus_one = addr + (data->blksz * data->blocks) - 1024;
 | |
| 	l2c_lock_mem_region(host->n_minus_one, 512);
 | |
| }
 | |
| 
 | |
| static void octeon_mmc_dmar_fixup_done(struct cvm_mmc_host *host)
 | |
| {
 | |
| 	if (!host->n_minus_one)
 | |
| 		return;
 | |
| 	l2c_unlock_mem_region(host->n_minus_one, 512);
 | |
| 	host->n_minus_one = 0;
 | |
| }
 | |
| 
 | |
| static int octeon_mmc_probe(struct platform_device *pdev)
 | |
| {
 | |
| 	struct device_node *cn, *node = pdev->dev.of_node;
 | |
| 	struct cvm_mmc_host *host;
 | |
| 	struct resource	*res;
 | |
| 	void __iomem *base;
 | |
| 	int mmc_irq[9];
 | |
| 	int i, ret = 0;
 | |
| 	u64 val;
 | |
| 
 | |
| 	host = devm_kzalloc(&pdev->dev, sizeof(*host), GFP_KERNEL);
 | |
| 	if (!host)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	spin_lock_init(&host->irq_handler_lock);
 | |
| 	sema_init(&host->mmc_serializer, 1);
 | |
| 
 | |
| 	host->dev = &pdev->dev;
 | |
| 	host->acquire_bus = octeon_mmc_acquire_bus;
 | |
| 	host->release_bus = octeon_mmc_release_bus;
 | |
| 	host->int_enable = octeon_mmc_int_enable;
 | |
| 	host->set_shared_power = octeon_mmc_set_shared_power;
 | |
| 	if (OCTEON_IS_MODEL(OCTEON_CN6XXX) ||
 | |
| 	    OCTEON_IS_MODEL(OCTEON_CNF7XXX)) {
 | |
| 		host->dmar_fixup = octeon_mmc_dmar_fixup;
 | |
| 		host->dmar_fixup_done = octeon_mmc_dmar_fixup_done;
 | |
| 	}
 | |
| 
 | |
| 	host->sys_freq = octeon_get_io_clock_rate();
 | |
| 
 | |
| 	if (of_device_is_compatible(node, "cavium,octeon-7890-mmc")) {
 | |
| 		host->big_dma_addr = true;
 | |
| 		host->need_irq_handler_lock = true;
 | |
| 		host->has_ciu3 = true;
 | |
| 		host->use_sg = true;
 | |
| 		/*
 | |
| 		 * First seven are the EMM_INT bits 0..6, then two for
 | |
| 		 * the EMM_DMA_INT bits
 | |
| 		 */
 | |
| 		for (i = 0; i < 9; i++) {
 | |
| 			mmc_irq[i] = platform_get_irq(pdev, i);
 | |
| 			if (mmc_irq[i] < 0)
 | |
| 				return mmc_irq[i];
 | |
| 
 | |
| 			/* work around legacy u-boot device trees */
 | |
| 			irq_set_irq_type(mmc_irq[i], IRQ_TYPE_EDGE_RISING);
 | |
| 		}
 | |
| 	} else {
 | |
| 		host->big_dma_addr = false;
 | |
| 		host->need_irq_handler_lock = false;
 | |
| 		host->has_ciu3 = false;
 | |
| 		/* First one is EMM second DMA */
 | |
| 		for (i = 0; i < 2; i++) {
 | |
| 			mmc_irq[i] = platform_get_irq(pdev, i);
 | |
| 			if (mmc_irq[i] < 0)
 | |
| 				return mmc_irq[i];
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	host->last_slot = -1;
 | |
| 
 | |
| 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
 | |
| 	if (!res) {
 | |
| 		dev_err(&pdev->dev, "Platform resource[0] is missing\n");
 | |
| 		return -ENXIO;
 | |
| 	}
 | |
| 	base = devm_ioremap_resource(&pdev->dev, res);
 | |
| 	if (IS_ERR(base))
 | |
| 		return PTR_ERR(base);
 | |
| 	host->base = (void __iomem *)base;
 | |
| 	host->reg_off = 0;
 | |
| 
 | |
| 	res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
 | |
| 	if (!res) {
 | |
| 		dev_err(&pdev->dev, "Platform resource[1] is missing\n");
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	base = devm_ioremap_resource(&pdev->dev, res);
 | |
| 	if (IS_ERR(base))
 | |
| 		return PTR_ERR(base);
 | |
| 	host->dma_base = (void __iomem *)base;
 | |
| 	/*
 | |
| 	 * To keep the register addresses shared we intentionaly use
 | |
| 	 * a negative offset here, first register used on Octeon therefore
 | |
| 	 * starts at 0x20 (MIO_EMM_DMA_CFG).
 | |
| 	 */
 | |
| 	host->reg_off_dma = -0x20;
 | |
| 
 | |
| 	ret = dma_set_mask(&pdev->dev, DMA_BIT_MASK(64));
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	/*
 | |
| 	 * Clear out any pending interrupts that may be left over from
 | |
| 	 * bootloader.
 | |
| 	 */
 | |
| 	val = readq(host->base + MIO_EMM_INT(host));
 | |
| 	writeq(val, host->base + MIO_EMM_INT(host));
 | |
| 
 | |
| 	if (host->has_ciu3) {
 | |
| 		/* Only CMD_DONE, DMA_DONE, CMD_ERR, DMA_ERR */
 | |
| 		for (i = 1; i <= 4; i++) {
 | |
| 			ret = devm_request_irq(&pdev->dev, mmc_irq[i],
 | |
| 					       cvm_mmc_interrupt,
 | |
| 					       0, cvm_mmc_irq_names[i], host);
 | |
| 			if (ret < 0) {
 | |
| 				dev_err(&pdev->dev, "Error: devm_request_irq %d\n",
 | |
| 					mmc_irq[i]);
 | |
| 				return ret;
 | |
| 			}
 | |
| 		}
 | |
| 	} else {
 | |
| 		ret = devm_request_irq(&pdev->dev, mmc_irq[0],
 | |
| 				       cvm_mmc_interrupt, 0, KBUILD_MODNAME,
 | |
| 				       host);
 | |
| 		if (ret < 0) {
 | |
| 			dev_err(&pdev->dev, "Error: devm_request_irq %d\n",
 | |
| 				mmc_irq[0]);
 | |
| 			return ret;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	host->global_pwr_gpiod = devm_gpiod_get_optional(&pdev->dev,
 | |
| 							 "power",
 | |
| 							 GPIOD_OUT_HIGH);
 | |
| 	if (IS_ERR(host->global_pwr_gpiod)) {
 | |
| 		dev_err(&pdev->dev, "Invalid power GPIO\n");
 | |
| 		return PTR_ERR(host->global_pwr_gpiod);
 | |
| 	}
 | |
| 
 | |
| 	platform_set_drvdata(pdev, host);
 | |
| 
 | |
| 	i = 0;
 | |
| 	for_each_child_of_node(node, cn) {
 | |
| 		host->slot_pdev[i] =
 | |
| 			of_platform_device_create(cn, NULL, &pdev->dev);
 | |
| 		if (!host->slot_pdev[i]) {
 | |
| 			i++;
 | |
| 			continue;
 | |
| 		}
 | |
| 		ret = cvm_mmc_of_slot_probe(&host->slot_pdev[i]->dev, host);
 | |
| 		if (ret) {
 | |
| 			dev_err(&pdev->dev, "Error populating slots\n");
 | |
| 			octeon_mmc_set_shared_power(host, 0);
 | |
| 			goto error;
 | |
| 		}
 | |
| 		i++;
 | |
| 	}
 | |
| 	return 0;
 | |
| 
 | |
| error:
 | |
| 	for (i = 0; i < CAVIUM_MAX_MMC; i++) {
 | |
| 		if (host->slot[i])
 | |
| 			cvm_mmc_of_slot_remove(host->slot[i]);
 | |
| 		if (host->slot_pdev[i])
 | |
| 			of_platform_device_destroy(&host->slot_pdev[i]->dev, NULL);
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int octeon_mmc_remove(struct platform_device *pdev)
 | |
| {
 | |
| 	struct cvm_mmc_host *host = platform_get_drvdata(pdev);
 | |
| 	u64 dma_cfg;
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < CAVIUM_MAX_MMC; i++)
 | |
| 		if (host->slot[i])
 | |
| 			cvm_mmc_of_slot_remove(host->slot[i]);
 | |
| 
 | |
| 	dma_cfg = readq(host->dma_base + MIO_EMM_DMA_CFG(host));
 | |
| 	dma_cfg &= ~MIO_EMM_DMA_CFG_EN;
 | |
| 	writeq(dma_cfg, host->dma_base + MIO_EMM_DMA_CFG(host));
 | |
| 
 | |
| 	octeon_mmc_set_shared_power(host, 0);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static const struct of_device_id octeon_mmc_match[] = {
 | |
| 	{
 | |
| 		.compatible = "cavium,octeon-6130-mmc",
 | |
| 	},
 | |
| 	{
 | |
| 		.compatible = "cavium,octeon-7890-mmc",
 | |
| 	},
 | |
| 	{},
 | |
| };
 | |
| MODULE_DEVICE_TABLE(of, octeon_mmc_match);
 | |
| 
 | |
| static struct platform_driver octeon_mmc_driver = {
 | |
| 	.probe		= octeon_mmc_probe,
 | |
| 	.remove		= octeon_mmc_remove,
 | |
| 	.driver		= {
 | |
| 		.name	= KBUILD_MODNAME,
 | |
| 		.of_match_table = octeon_mmc_match,
 | |
| 	},
 | |
| };
 | |
| 
 | |
| module_platform_driver(octeon_mmc_driver);
 | |
| 
 | |
| MODULE_AUTHOR("Cavium Inc. <support@cavium.com>");
 | |
| MODULE_DESCRIPTION("Low-level driver for Cavium OCTEON MMC/SSD card");
 | |
| MODULE_LICENSE("GPL");
 |