mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Remove cryptd_alloc_ablkcipher() and the ability of cryptd to create
algorithms with the deprecated "ablkcipher" type.
This has been unused since commit 0e145b477d ("crypto: ablk_helper -
remove ablk_helper").  Instead, cryptd_alloc_skcipher() is used.
Signed-off-by: Eric Biggers <ebiggers@google.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
		
	
			
		
			
				
	
	
		
			71 lines
		
	
	
	
		
			2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			71 lines
		
	
	
	
		
			2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* SPDX-License-Identifier: GPL-2.0 */
 | 
						|
/*
 | 
						|
 * Software async crypto daemon
 | 
						|
 *
 | 
						|
 * Added AEAD support to cryptd.
 | 
						|
 *    Authors: Tadeusz Struk (tadeusz.struk@intel.com)
 | 
						|
 *             Adrian Hoban <adrian.hoban@intel.com>
 | 
						|
 *             Gabriele Paoloni <gabriele.paoloni@intel.com>
 | 
						|
 *             Aidan O'Mahony (aidan.o.mahony@intel.com)
 | 
						|
 *    Copyright (c) 2010, Intel Corporation.
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef _CRYPTO_CRYPT_H
 | 
						|
#define _CRYPTO_CRYPT_H
 | 
						|
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <crypto/aead.h>
 | 
						|
#include <crypto/hash.h>
 | 
						|
#include <crypto/skcipher.h>
 | 
						|
 | 
						|
struct cryptd_skcipher {
 | 
						|
	struct crypto_skcipher base;
 | 
						|
};
 | 
						|
 | 
						|
/* alg_name should be algorithm to be cryptd-ed */
 | 
						|
struct cryptd_skcipher *cryptd_alloc_skcipher(const char *alg_name,
 | 
						|
					      u32 type, u32 mask);
 | 
						|
struct crypto_skcipher *cryptd_skcipher_child(struct cryptd_skcipher *tfm);
 | 
						|
/* Must be called without moving CPUs. */
 | 
						|
bool cryptd_skcipher_queued(struct cryptd_skcipher *tfm);
 | 
						|
void cryptd_free_skcipher(struct cryptd_skcipher *tfm);
 | 
						|
 | 
						|
struct cryptd_ahash {
 | 
						|
	struct crypto_ahash base;
 | 
						|
};
 | 
						|
 | 
						|
static inline struct cryptd_ahash *__cryptd_ahash_cast(
 | 
						|
	struct crypto_ahash *tfm)
 | 
						|
{
 | 
						|
	return (struct cryptd_ahash *)tfm;
 | 
						|
}
 | 
						|
 | 
						|
/* alg_name should be algorithm to be cryptd-ed */
 | 
						|
struct cryptd_ahash *cryptd_alloc_ahash(const char *alg_name,
 | 
						|
					u32 type, u32 mask);
 | 
						|
struct crypto_shash *cryptd_ahash_child(struct cryptd_ahash *tfm);
 | 
						|
struct shash_desc *cryptd_shash_desc(struct ahash_request *req);
 | 
						|
/* Must be called without moving CPUs. */
 | 
						|
bool cryptd_ahash_queued(struct cryptd_ahash *tfm);
 | 
						|
void cryptd_free_ahash(struct cryptd_ahash *tfm);
 | 
						|
 | 
						|
struct cryptd_aead {
 | 
						|
	struct crypto_aead base;
 | 
						|
};
 | 
						|
 | 
						|
static inline struct cryptd_aead *__cryptd_aead_cast(
 | 
						|
	struct crypto_aead *tfm)
 | 
						|
{
 | 
						|
	return (struct cryptd_aead *)tfm;
 | 
						|
}
 | 
						|
 | 
						|
struct cryptd_aead *cryptd_alloc_aead(const char *alg_name,
 | 
						|
					  u32 type, u32 mask);
 | 
						|
 | 
						|
struct crypto_aead *cryptd_aead_child(struct cryptd_aead *tfm);
 | 
						|
/* Must be called without moving CPUs. */
 | 
						|
bool cryptd_aead_queued(struct cryptd_aead *tfm);
 | 
						|
 | 
						|
void cryptd_free_aead(struct cryptd_aead *tfm);
 | 
						|
 | 
						|
#endif
 |