mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 08:38:45 +02:00 
			
		
		
		
	 f6b8e86b7a
			
		
	
	
		f6b8e86b7a
		
	
	
	
	
		
			
			Here is the big set of tty/serial driver changes for 6.10-rc1.  Included
 in here are:
   - Usual good set of api cleanups and evolution by Jiri Slaby to make
     the serial interfaces move out of the 1990's by using kfifos instead
     of hand-rolling their own logic.
   - 8250_exar driver updates
   - max3100 driver updates
   - sc16is7xx driver updates
   - exar driver updates
   - sh-sci driver updates
   - tty ldisc api addition to help refuse bindings
   - other smaller serial driver updates
 
 All of these have been in linux-next for a while with no reported
 issues.
 
 Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
 -----BEGIN PGP SIGNATURE-----
 
 iG0EABECAC0WIQT0tgzFv3jCIUoxPcsxR9QN2y37KQUCZk4Cvg8cZ3JlZ0Brcm9h
 aC5jb20ACgkQMUfUDdst+ymqpwCgnHU1NeBBUsvoSDOLk5oApIQ4jVgAn102jWlw
 3dNDhA4i3Ay/mZdv8/Kj
 =TI+P
 -----END PGP SIGNATURE-----
Merge tag 'tty-6.10-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/gregkh/tty
Pull tty / serial updates from Greg KH:
 "Here is the big set of tty/serial driver changes for 6.10-rc1.
  Included in here are:
   - Usual good set of api cleanups and evolution by Jiri Slaby to make
     the serial interfaces move out of the 1990's by using kfifos
     instead of hand-rolling their own logic.
   - 8250_exar driver updates
   - max3100 driver updates
   - sc16is7xx driver updates
   - exar driver updates
   - sh-sci driver updates
   - tty ldisc api addition to help refuse bindings
   - other smaller serial driver updates
  All of these have been in linux-next for a while with no reported
  issues"
* tag 'tty-6.10-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/gregkh/tty: (113 commits)
  serial: Clear UPF_DEAD before calling tty_port_register_device_attr_serdev()
  serial: imx: Raise TX trigger level to 8
  serial: 8250_pnp: Simplify "line" related code
  serial: sh-sci: simplify locking when re-issuing RXDMA fails
  serial: sh-sci: let timeout timer only run when DMA is scheduled
  serial: sh-sci: describe locking requirements for invalidating RXDMA
  serial: sh-sci: protect invalidating RXDMA on shutdown
  tty: add the option to have a tty reject a new ldisc
  serial: core: Call device_set_awake_path() for console port
  dt-bindings: serial: brcm,bcm2835-aux-uart: convert to dtschema
  tty: serial: uartps: Add support for uartps controller reset
  arm64: zynqmp: Add resets property for UART nodes
  dt-bindings: serial: cdns,uart: Add optional reset property
  serial: 8250_pnp: Switch to DEFINE_SIMPLE_DEV_PM_OPS()
  serial: 8250_exar: Keep the includes sorted
  serial: 8250_exar: Make type of bit the same in exar_ee_*_bit()
  serial: 8250_exar: Use BIT() in exar_ee_read()
  serial: 8250_exar: Switch to use dev_err_probe()
  serial: 8250_exar: Return directly from switch-cases
  serial: 8250_exar: Decrease indentation level
  ...
		
	
			
		
			
				
	
	
		
			595 lines
		
	
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			595 lines
		
	
	
	
		
			12 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-or-later
 | |
| /*
 | |
|  * A generic kernel FIFO implementation
 | |
|  *
 | |
|  * Copyright (C) 2009/2010 Stefani Seibold <stefani@seibold.net>
 | |
|  */
 | |
| 
 | |
| #include <linux/dma-mapping.h>
 | |
| #include <linux/err.h>
 | |
| #include <linux/export.h>
 | |
| #include <linux/kfifo.h>
 | |
| #include <linux/log2.h>
 | |
| #include <linux/scatterlist.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/uaccess.h>
 | |
| 
 | |
| /*
 | |
|  * internal helper to calculate the unused elements in a fifo
 | |
|  */
 | |
| static inline unsigned int kfifo_unused(struct __kfifo *fifo)
 | |
| {
 | |
| 	return (fifo->mask + 1) - (fifo->in - fifo->out);
 | |
| }
 | |
| 
 | |
| int __kfifo_alloc(struct __kfifo *fifo, unsigned int size,
 | |
| 		size_t esize, gfp_t gfp_mask)
 | |
| {
 | |
| 	/*
 | |
| 	 * round up to the next power of 2, since our 'let the indices
 | |
| 	 * wrap' technique works only in this case.
 | |
| 	 */
 | |
| 	size = roundup_pow_of_two(size);
 | |
| 
 | |
| 	fifo->in = 0;
 | |
| 	fifo->out = 0;
 | |
| 	fifo->esize = esize;
 | |
| 
 | |
| 	if (size < 2) {
 | |
| 		fifo->data = NULL;
 | |
| 		fifo->mask = 0;
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	fifo->data = kmalloc_array(esize, size, gfp_mask);
 | |
| 
 | |
| 	if (!fifo->data) {
 | |
| 		fifo->mask = 0;
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 	fifo->mask = size - 1;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_alloc);
 | |
| 
 | |
| void __kfifo_free(struct __kfifo *fifo)
 | |
| {
 | |
| 	kfree(fifo->data);
 | |
| 	fifo->in = 0;
 | |
| 	fifo->out = 0;
 | |
| 	fifo->esize = 0;
 | |
| 	fifo->data = NULL;
 | |
| 	fifo->mask = 0;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_free);
 | |
| 
 | |
| int __kfifo_init(struct __kfifo *fifo, void *buffer,
 | |
| 		unsigned int size, size_t esize)
 | |
| {
 | |
| 	size /= esize;
 | |
| 
 | |
| 	if (!is_power_of_2(size))
 | |
| 		size = rounddown_pow_of_two(size);
 | |
| 
 | |
| 	fifo->in = 0;
 | |
| 	fifo->out = 0;
 | |
| 	fifo->esize = esize;
 | |
| 	fifo->data = buffer;
 | |
| 
 | |
| 	if (size < 2) {
 | |
| 		fifo->mask = 0;
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	fifo->mask = size - 1;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_init);
 | |
| 
 | |
| static void kfifo_copy_in(struct __kfifo *fifo, const void *src,
 | |
| 		unsigned int len, unsigned int off)
 | |
| {
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	off &= fifo->mask;
 | |
| 	if (esize != 1) {
 | |
| 		off *= esize;
 | |
| 		size *= esize;
 | |
| 		len *= esize;
 | |
| 	}
 | |
| 	l = min(len, size - off);
 | |
| 
 | |
| 	memcpy(fifo->data + off, src, l);
 | |
| 	memcpy(fifo->data, src + l, len - l);
 | |
| 	/*
 | |
| 	 * make sure that the data in the fifo is up to date before
 | |
| 	 * incrementing the fifo->in index counter
 | |
| 	 */
 | |
| 	smp_wmb();
 | |
| }
 | |
| 
 | |
| unsigned int __kfifo_in(struct __kfifo *fifo,
 | |
| 		const void *buf, unsigned int len)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	l = kfifo_unused(fifo);
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 
 | |
| 	kfifo_copy_in(fifo, buf, len, fifo->in);
 | |
| 	fifo->in += len;
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_in);
 | |
| 
 | |
| static void kfifo_copy_out(struct __kfifo *fifo, void *dst,
 | |
| 		unsigned int len, unsigned int off)
 | |
| {
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	off &= fifo->mask;
 | |
| 	if (esize != 1) {
 | |
| 		off *= esize;
 | |
| 		size *= esize;
 | |
| 		len *= esize;
 | |
| 	}
 | |
| 	l = min(len, size - off);
 | |
| 
 | |
| 	memcpy(dst, fifo->data + off, l);
 | |
| 	memcpy(dst + l, fifo->data, len - l);
 | |
| 	/*
 | |
| 	 * make sure that the data is copied before
 | |
| 	 * incrementing the fifo->out index counter
 | |
| 	 */
 | |
| 	smp_wmb();
 | |
| }
 | |
| 
 | |
| unsigned int __kfifo_out_peek(struct __kfifo *fifo,
 | |
| 		void *buf, unsigned int len)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	l = fifo->in - fifo->out;
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 
 | |
| 	kfifo_copy_out(fifo, buf, len, fifo->out);
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out_peek);
 | |
| 
 | |
| unsigned int __kfifo_out_linear(struct __kfifo *fifo,
 | |
| 		unsigned int *tail, unsigned int n)
 | |
| {
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int off = fifo->out & fifo->mask;
 | |
| 
 | |
| 	if (tail)
 | |
| 		*tail = off;
 | |
| 
 | |
| 	return min3(n, fifo->in - fifo->out, size - off);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out_linear);
 | |
| 
 | |
| unsigned int __kfifo_out(struct __kfifo *fifo,
 | |
| 		void *buf, unsigned int len)
 | |
| {
 | |
| 	len = __kfifo_out_peek(fifo, buf, len);
 | |
| 	fifo->out += len;
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out);
 | |
| 
 | |
| static unsigned long kfifo_copy_from_user(struct __kfifo *fifo,
 | |
| 	const void __user *from, unsigned int len, unsigned int off,
 | |
| 	unsigned int *copied)
 | |
| {
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	unsigned int l;
 | |
| 	unsigned long ret;
 | |
| 
 | |
| 	off &= fifo->mask;
 | |
| 	if (esize != 1) {
 | |
| 		off *= esize;
 | |
| 		size *= esize;
 | |
| 		len *= esize;
 | |
| 	}
 | |
| 	l = min(len, size - off);
 | |
| 
 | |
| 	ret = copy_from_user(fifo->data + off, from, l);
 | |
| 	if (unlikely(ret))
 | |
| 		ret = DIV_ROUND_UP(ret + len - l, esize);
 | |
| 	else {
 | |
| 		ret = copy_from_user(fifo->data, from + l, len - l);
 | |
| 		if (unlikely(ret))
 | |
| 			ret = DIV_ROUND_UP(ret, esize);
 | |
| 	}
 | |
| 	/*
 | |
| 	 * make sure that the data in the fifo is up to date before
 | |
| 	 * incrementing the fifo->in index counter
 | |
| 	 */
 | |
| 	smp_wmb();
 | |
| 	*copied = len - ret * esize;
 | |
| 	/* return the number of elements which are not copied */
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int __kfifo_from_user(struct __kfifo *fifo, const void __user *from,
 | |
| 		unsigned long len, unsigned int *copied)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 	unsigned long ret;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	int err;
 | |
| 
 | |
| 	if (esize != 1)
 | |
| 		len /= esize;
 | |
| 
 | |
| 	l = kfifo_unused(fifo);
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 
 | |
| 	ret = kfifo_copy_from_user(fifo, from, len, fifo->in, copied);
 | |
| 	if (unlikely(ret)) {
 | |
| 		len -= ret;
 | |
| 		err = -EFAULT;
 | |
| 	} else
 | |
| 		err = 0;
 | |
| 	fifo->in += len;
 | |
| 	return err;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_from_user);
 | |
| 
 | |
| static unsigned long kfifo_copy_to_user(struct __kfifo *fifo, void __user *to,
 | |
| 		unsigned int len, unsigned int off, unsigned int *copied)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 	unsigned long ret;
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 
 | |
| 	off &= fifo->mask;
 | |
| 	if (esize != 1) {
 | |
| 		off *= esize;
 | |
| 		size *= esize;
 | |
| 		len *= esize;
 | |
| 	}
 | |
| 	l = min(len, size - off);
 | |
| 
 | |
| 	ret = copy_to_user(to, fifo->data + off, l);
 | |
| 	if (unlikely(ret))
 | |
| 		ret = DIV_ROUND_UP(ret + len - l, esize);
 | |
| 	else {
 | |
| 		ret = copy_to_user(to + l, fifo->data, len - l);
 | |
| 		if (unlikely(ret))
 | |
| 			ret = DIV_ROUND_UP(ret, esize);
 | |
| 	}
 | |
| 	/*
 | |
| 	 * make sure that the data is copied before
 | |
| 	 * incrementing the fifo->out index counter
 | |
| 	 */
 | |
| 	smp_wmb();
 | |
| 	*copied = len - ret * esize;
 | |
| 	/* return the number of elements which are not copied */
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int __kfifo_to_user(struct __kfifo *fifo, void __user *to,
 | |
| 		unsigned long len, unsigned int *copied)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 	unsigned long ret;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	int err;
 | |
| 
 | |
| 	if (esize != 1)
 | |
| 		len /= esize;
 | |
| 
 | |
| 	l = fifo->in - fifo->out;
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 	ret = kfifo_copy_to_user(fifo, to, len, fifo->out, copied);
 | |
| 	if (unlikely(ret)) {
 | |
| 		len -= ret;
 | |
| 		err = -EFAULT;
 | |
| 	} else
 | |
| 		err = 0;
 | |
| 	fifo->out += len;
 | |
| 	return err;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_to_user);
 | |
| 
 | |
| static unsigned int setup_sgl_buf(struct __kfifo *fifo, struct scatterlist *sgl,
 | |
| 				  unsigned int data_offset, int nents,
 | |
| 				  unsigned int len, dma_addr_t dma)
 | |
| {
 | |
| 	const void *buf = fifo->data + data_offset;
 | |
| 
 | |
| 	if (!nents || !len)
 | |
| 		return 0;
 | |
| 
 | |
| 	sg_set_buf(sgl, buf, len);
 | |
| 
 | |
| 	if (dma != DMA_MAPPING_ERROR) {
 | |
| 		sg_dma_address(sgl) = dma + data_offset;
 | |
| 		sg_dma_len(sgl) = len;
 | |
| 	}
 | |
| 
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl,
 | |
| 		int nents, unsigned int len, unsigned int off, dma_addr_t dma)
 | |
| {
 | |
| 	unsigned int size = fifo->mask + 1;
 | |
| 	unsigned int esize = fifo->esize;
 | |
| 	unsigned int len_to_end;
 | |
| 	unsigned int n;
 | |
| 
 | |
| 	off &= fifo->mask;
 | |
| 	if (esize != 1) {
 | |
| 		off *= esize;
 | |
| 		size *= esize;
 | |
| 		len *= esize;
 | |
| 	}
 | |
| 	len_to_end = min(len, size - off);
 | |
| 
 | |
| 	n = setup_sgl_buf(fifo, sgl, off, nents, len_to_end, dma);
 | |
| 	n += setup_sgl_buf(fifo, sgl + n, 0, nents - n, len - len_to_end, dma);
 | |
| 
 | |
| 	return n;
 | |
| }
 | |
| 
 | |
| unsigned int __kfifo_dma_in_prepare(struct __kfifo *fifo,
 | |
| 		struct scatterlist *sgl, int nents, unsigned int len,
 | |
| 		dma_addr_t dma)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	l = kfifo_unused(fifo);
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 
 | |
| 	return setup_sgl(fifo, sgl, nents, len, fifo->in, dma);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_dma_in_prepare);
 | |
| 
 | |
| unsigned int __kfifo_dma_out_prepare(struct __kfifo *fifo,
 | |
| 		struct scatterlist *sgl, int nents, unsigned int len,
 | |
| 		dma_addr_t dma)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 
 | |
| 	l = fifo->in - fifo->out;
 | |
| 	if (len > l)
 | |
| 		len = l;
 | |
| 
 | |
| 	return setup_sgl(fifo, sgl, nents, len, fifo->out, dma);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_dma_out_prepare);
 | |
| 
 | |
| unsigned int __kfifo_max_r(unsigned int len, size_t recsize)
 | |
| {
 | |
| 	unsigned int max = (1 << (recsize << 3)) - 1;
 | |
| 
 | |
| 	if (len > max)
 | |
| 		return max;
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_max_r);
 | |
| 
 | |
| #define	__KFIFO_PEEK(data, out, mask) \
 | |
| 	((data)[(out) & (mask)])
 | |
| /*
 | |
|  * __kfifo_peek_n internal helper function for determinate the length of
 | |
|  * the next record in the fifo
 | |
|  */
 | |
| static unsigned int __kfifo_peek_n(struct __kfifo *fifo, size_t recsize)
 | |
| {
 | |
| 	unsigned int l;
 | |
| 	unsigned int mask = fifo->mask;
 | |
| 	unsigned char *data = fifo->data;
 | |
| 
 | |
| 	l = __KFIFO_PEEK(data, fifo->out, mask);
 | |
| 
 | |
| 	if (--recsize)
 | |
| 		l |= __KFIFO_PEEK(data, fifo->out + 1, mask) << 8;
 | |
| 
 | |
| 	return l;
 | |
| }
 | |
| 
 | |
| #define	__KFIFO_POKE(data, in, mask, val) \
 | |
| 	( \
 | |
| 	(data)[(in) & (mask)] = (unsigned char)(val) \
 | |
| 	)
 | |
| 
 | |
| /*
 | |
|  * __kfifo_poke_n internal helper function for storing the length of
 | |
|  * the record into the fifo
 | |
|  */
 | |
| static void __kfifo_poke_n(struct __kfifo *fifo, unsigned int n, size_t recsize)
 | |
| {
 | |
| 	unsigned int mask = fifo->mask;
 | |
| 	unsigned char *data = fifo->data;
 | |
| 
 | |
| 	__KFIFO_POKE(data, fifo->in, mask, n);
 | |
| 
 | |
| 	if (recsize > 1)
 | |
| 		__KFIFO_POKE(data, fifo->in + 1, mask, n >> 8);
 | |
| }
 | |
| 
 | |
| unsigned int __kfifo_len_r(struct __kfifo *fifo, size_t recsize)
 | |
| {
 | |
| 	return __kfifo_peek_n(fifo, recsize);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_len_r);
 | |
| 
 | |
| unsigned int __kfifo_in_r(struct __kfifo *fifo, const void *buf,
 | |
| 		unsigned int len, size_t recsize)
 | |
| {
 | |
| 	if (len + recsize > kfifo_unused(fifo))
 | |
| 		return 0;
 | |
| 
 | |
| 	__kfifo_poke_n(fifo, len, recsize);
 | |
| 
 | |
| 	kfifo_copy_in(fifo, buf, len, fifo->in + recsize);
 | |
| 	fifo->in += len + recsize;
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_in_r);
 | |
| 
 | |
| static unsigned int kfifo_out_copy_r(struct __kfifo *fifo,
 | |
| 	void *buf, unsigned int len, size_t recsize, unsigned int *n)
 | |
| {
 | |
| 	*n = __kfifo_peek_n(fifo, recsize);
 | |
| 
 | |
| 	if (len > *n)
 | |
| 		len = *n;
 | |
| 
 | |
| 	kfifo_copy_out(fifo, buf, len, fifo->out + recsize);
 | |
| 	return len;
 | |
| }
 | |
| 
 | |
| unsigned int __kfifo_out_peek_r(struct __kfifo *fifo, void *buf,
 | |
| 		unsigned int len, size_t recsize)
 | |
| {
 | |
| 	unsigned int n;
 | |
| 
 | |
| 	if (fifo->in == fifo->out)
 | |
| 		return 0;
 | |
| 
 | |
| 	return kfifo_out_copy_r(fifo, buf, len, recsize, &n);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out_peek_r);
 | |
| 
 | |
| unsigned int __kfifo_out_linear_r(struct __kfifo *fifo,
 | |
| 		unsigned int *tail, unsigned int n, size_t recsize)
 | |
| {
 | |
| 	if (fifo->in == fifo->out)
 | |
| 		return 0;
 | |
| 
 | |
| 	if (tail)
 | |
| 		*tail = fifo->out + recsize;
 | |
| 
 | |
| 	return min(n, __kfifo_peek_n(fifo, recsize));
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out_linear_r);
 | |
| 
 | |
| unsigned int __kfifo_out_r(struct __kfifo *fifo, void *buf,
 | |
| 		unsigned int len, size_t recsize)
 | |
| {
 | |
| 	unsigned int n;
 | |
| 
 | |
| 	if (fifo->in == fifo->out)
 | |
| 		return 0;
 | |
| 
 | |
| 	len = kfifo_out_copy_r(fifo, buf, len, recsize, &n);
 | |
| 	fifo->out += n + recsize;
 | |
| 	return len;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_out_r);
 | |
| 
 | |
| void __kfifo_skip_r(struct __kfifo *fifo, size_t recsize)
 | |
| {
 | |
| 	unsigned int n;
 | |
| 
 | |
| 	n = __kfifo_peek_n(fifo, recsize);
 | |
| 	fifo->out += n + recsize;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_skip_r);
 | |
| 
 | |
| int __kfifo_from_user_r(struct __kfifo *fifo, const void __user *from,
 | |
| 	unsigned long len, unsigned int *copied, size_t recsize)
 | |
| {
 | |
| 	unsigned long ret;
 | |
| 
 | |
| 	len = __kfifo_max_r(len, recsize);
 | |
| 
 | |
| 	if (len + recsize > kfifo_unused(fifo)) {
 | |
| 		*copied = 0;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	__kfifo_poke_n(fifo, len, recsize);
 | |
| 
 | |
| 	ret = kfifo_copy_from_user(fifo, from, len, fifo->in + recsize, copied);
 | |
| 	if (unlikely(ret)) {
 | |
| 		*copied = 0;
 | |
| 		return -EFAULT;
 | |
| 	}
 | |
| 	fifo->in += len + recsize;
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_from_user_r);
 | |
| 
 | |
| int __kfifo_to_user_r(struct __kfifo *fifo, void __user *to,
 | |
| 	unsigned long len, unsigned int *copied, size_t recsize)
 | |
| {
 | |
| 	unsigned long ret;
 | |
| 	unsigned int n;
 | |
| 
 | |
| 	if (fifo->in == fifo->out) {
 | |
| 		*copied = 0;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	n = __kfifo_peek_n(fifo, recsize);
 | |
| 	if (len > n)
 | |
| 		len = n;
 | |
| 
 | |
| 	ret = kfifo_copy_to_user(fifo, to, len, fifo->out + recsize, copied);
 | |
| 	if (unlikely(ret)) {
 | |
| 		*copied = 0;
 | |
| 		return -EFAULT;
 | |
| 	}
 | |
| 	fifo->out += n + recsize;
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_to_user_r);
 | |
| 
 | |
| unsigned int __kfifo_dma_in_prepare_r(struct __kfifo *fifo,
 | |
| 	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize,
 | |
| 	dma_addr_t dma)
 | |
| {
 | |
| 	BUG_ON(!nents);
 | |
| 
 | |
| 	len = __kfifo_max_r(len, recsize);
 | |
| 
 | |
| 	if (len + recsize > kfifo_unused(fifo))
 | |
| 		return 0;
 | |
| 
 | |
| 	return setup_sgl(fifo, sgl, nents, len, fifo->in + recsize, dma);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_dma_in_prepare_r);
 | |
| 
 | |
| void __kfifo_dma_in_finish_r(struct __kfifo *fifo,
 | |
| 	unsigned int len, size_t recsize)
 | |
| {
 | |
| 	len = __kfifo_max_r(len, recsize);
 | |
| 	__kfifo_poke_n(fifo, len, recsize);
 | |
| 	fifo->in += len + recsize;
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_dma_in_finish_r);
 | |
| 
 | |
| unsigned int __kfifo_dma_out_prepare_r(struct __kfifo *fifo,
 | |
| 	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize,
 | |
| 	dma_addr_t dma)
 | |
| {
 | |
| 	BUG_ON(!nents);
 | |
| 
 | |
| 	len = __kfifo_max_r(len, recsize);
 | |
| 
 | |
| 	if (len + recsize > fifo->in - fifo->out)
 | |
| 		return 0;
 | |
| 
 | |
| 	return setup_sgl(fifo, sgl, nents, len, fifo->out + recsize, dma);
 | |
| }
 | |
| EXPORT_SYMBOL(__kfifo_dma_out_prepare_r);
 | |
| 
 |