forked from mirrors/linux
		
	maple_tree: Add mtree_alloc_cyclic()
I need a cyclic allocator for the simple_offset implementation in fs/libfs.c. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Link: https://lore.kernel.org/r/170820144179.6328.12838600511394432325.stgit@91.116.238.104.host.secureserver.net Signed-off-by: Christian Brauner <brauner@kernel.org>
This commit is contained in:
		
							parent
							
								
									ecba88a3b3
								
							
						
					
					
						commit
						9b6713cc75
					
				
					 2 changed files with 100 additions and 0 deletions
				
			
		| 
						 | 
					@ -171,6 +171,7 @@ enum maple_type {
 | 
				
			||||||
#define MT_FLAGS_LOCK_IRQ	0x100
 | 
					#define MT_FLAGS_LOCK_IRQ	0x100
 | 
				
			||||||
#define MT_FLAGS_LOCK_BH	0x200
 | 
					#define MT_FLAGS_LOCK_BH	0x200
 | 
				
			||||||
#define MT_FLAGS_LOCK_EXTERN	0x300
 | 
					#define MT_FLAGS_LOCK_EXTERN	0x300
 | 
				
			||||||
 | 
					#define MT_FLAGS_ALLOC_WRAPPED	0x0800
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define MAPLE_HEIGHT_MAX	31
 | 
					#define MAPLE_HEIGHT_MAX	31
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -319,6 +320,9 @@ int mtree_insert_range(struct maple_tree *mt, unsigned long first,
 | 
				
			||||||
int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
 | 
					int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
		void *entry, unsigned long size, unsigned long min,
 | 
							void *entry, unsigned long size, unsigned long min,
 | 
				
			||||||
		unsigned long max, gfp_t gfp);
 | 
							unsigned long max, gfp_t gfp);
 | 
				
			||||||
 | 
					int mtree_alloc_cyclic(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
 | 
							void *entry, unsigned long range_lo, unsigned long range_hi,
 | 
				
			||||||
 | 
							unsigned long *next, gfp_t gfp);
 | 
				
			||||||
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
 | 
					int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
		void *entry, unsigned long size, unsigned long min,
 | 
							void *entry, unsigned long size, unsigned long min,
 | 
				
			||||||
		unsigned long max, gfp_t gfp);
 | 
							unsigned long max, gfp_t gfp);
 | 
				
			||||||
| 
						 | 
					@ -499,6 +503,9 @@ void *mas_find_range(struct ma_state *mas, unsigned long max);
 | 
				
			||||||
void *mas_find_rev(struct ma_state *mas, unsigned long min);
 | 
					void *mas_find_rev(struct ma_state *mas, unsigned long min);
 | 
				
			||||||
void *mas_find_range_rev(struct ma_state *mas, unsigned long max);
 | 
					void *mas_find_range_rev(struct ma_state *mas, unsigned long max);
 | 
				
			||||||
int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp);
 | 
					int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp);
 | 
				
			||||||
 | 
					int mas_alloc_cyclic(struct ma_state *mas, unsigned long *startp,
 | 
				
			||||||
 | 
							void *entry, unsigned long range_lo, unsigned long range_hi,
 | 
				
			||||||
 | 
							unsigned long *next, gfp_t gfp);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
bool mas_nomem(struct ma_state *mas, gfp_t gfp);
 | 
					bool mas_nomem(struct ma_state *mas, gfp_t gfp);
 | 
				
			||||||
void mas_pause(struct ma_state *mas);
 | 
					void mas_pause(struct ma_state *mas);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -4290,6 +4290,56 @@ static inline void *mas_insert(struct ma_state *mas, void *entry)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/**
 | 
				
			||||||
 | 
					 * mas_alloc_cyclic() - Internal call to find somewhere to store an entry
 | 
				
			||||||
 | 
					 * @mas: The maple state.
 | 
				
			||||||
 | 
					 * @startp: Pointer to ID.
 | 
				
			||||||
 | 
					 * @range_lo: Lower bound of range to search.
 | 
				
			||||||
 | 
					 * @range_hi: Upper bound of range to search.
 | 
				
			||||||
 | 
					 * @entry: The entry to store.
 | 
				
			||||||
 | 
					 * @next: Pointer to next ID to allocate.
 | 
				
			||||||
 | 
					 * @gfp: The GFP_FLAGS to use for allocations.
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 * Return: 0 if the allocation succeeded without wrapping, 1 if the
 | 
				
			||||||
 | 
					 * allocation succeeded after wrapping, or -EBUSY if there are no
 | 
				
			||||||
 | 
					 * free entries.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					int mas_alloc_cyclic(struct ma_state *mas, unsigned long *startp,
 | 
				
			||||||
 | 
							void *entry, unsigned long range_lo, unsigned long range_hi,
 | 
				
			||||||
 | 
							unsigned long *next, gfp_t gfp)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						unsigned long min = range_lo;
 | 
				
			||||||
 | 
						int ret = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						range_lo = max(min, *next);
 | 
				
			||||||
 | 
						ret = mas_empty_area(mas, range_lo, range_hi, 1);
 | 
				
			||||||
 | 
						if ((mas->tree->ma_flags & MT_FLAGS_ALLOC_WRAPPED) && ret == 0) {
 | 
				
			||||||
 | 
							mas->tree->ma_flags &= ~MT_FLAGS_ALLOC_WRAPPED;
 | 
				
			||||||
 | 
							ret = 1;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						if (ret < 0 && range_lo > min) {
 | 
				
			||||||
 | 
							ret = mas_empty_area(mas, min, range_hi, 1);
 | 
				
			||||||
 | 
							if (ret == 0)
 | 
				
			||||||
 | 
								ret = 1;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						if (ret < 0)
 | 
				
			||||||
 | 
							return ret;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						do {
 | 
				
			||||||
 | 
							mas_insert(mas, entry);
 | 
				
			||||||
 | 
						} while (mas_nomem(mas, gfp));
 | 
				
			||||||
 | 
						if (mas_is_err(mas))
 | 
				
			||||||
 | 
							return xa_err(mas->node);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						*startp = mas->index;
 | 
				
			||||||
 | 
						*next = *startp + 1;
 | 
				
			||||||
 | 
						if (*next == 0)
 | 
				
			||||||
 | 
							mas->tree->ma_flags |= MT_FLAGS_ALLOC_WRAPPED;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						return ret;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					EXPORT_SYMBOL(mas_alloc_cyclic);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static __always_inline void mas_rewalk(struct ma_state *mas, unsigned long index)
 | 
					static __always_inline void mas_rewalk(struct ma_state *mas, unsigned long index)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
retry:
 | 
					retry:
 | 
				
			||||||
| 
						 | 
					@ -6443,6 +6493,49 @@ int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(mtree_alloc_range);
 | 
					EXPORT_SYMBOL(mtree_alloc_range);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/**
 | 
				
			||||||
 | 
					 * mtree_alloc_cyclic() - Find somewhere to store this entry in the tree.
 | 
				
			||||||
 | 
					 * @mt: The maple tree.
 | 
				
			||||||
 | 
					 * @startp: Pointer to ID.
 | 
				
			||||||
 | 
					 * @range_lo: Lower bound of range to search.
 | 
				
			||||||
 | 
					 * @range_hi: Upper bound of range to search.
 | 
				
			||||||
 | 
					 * @entry: The entry to store.
 | 
				
			||||||
 | 
					 * @next: Pointer to next ID to allocate.
 | 
				
			||||||
 | 
					 * @gfp: The GFP_FLAGS to use for allocations.
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 * Finds an empty entry in @mt after @next, stores the new index into
 | 
				
			||||||
 | 
					 * the @id pointer, stores the entry at that index, then updates @next.
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 * @mt must be initialized with the MT_FLAGS_ALLOC_RANGE flag.
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 * Context: Any context.  Takes and releases the mt.lock.  May sleep if
 | 
				
			||||||
 | 
					 * the @gfp flags permit.
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 * Return: 0 if the allocation succeeded without wrapping, 1 if the
 | 
				
			||||||
 | 
					 * allocation succeeded after wrapping, -ENOMEM if memory could not be
 | 
				
			||||||
 | 
					 * allocated, -EINVAL if @mt cannot be used, or -EBUSY if there are no
 | 
				
			||||||
 | 
					 * free entries.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					int mtree_alloc_cyclic(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
 | 
							void *entry, unsigned long range_lo, unsigned long range_hi,
 | 
				
			||||||
 | 
							unsigned long *next, gfp_t gfp)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						int ret;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						MA_STATE(mas, mt, 0, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (!mt_is_alloc(mt))
 | 
				
			||||||
 | 
							return -EINVAL;
 | 
				
			||||||
 | 
						if (WARN_ON_ONCE(mt_is_reserved(entry)))
 | 
				
			||||||
 | 
							return -EINVAL;
 | 
				
			||||||
 | 
						mtree_lock(mt);
 | 
				
			||||||
 | 
						ret = mas_alloc_cyclic(&mas, startp, entry, range_lo, range_hi,
 | 
				
			||||||
 | 
								       next, gfp);
 | 
				
			||||||
 | 
						mtree_unlock(mt);
 | 
				
			||||||
 | 
						return ret;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					EXPORT_SYMBOL(mtree_alloc_cyclic);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
 | 
					int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
 | 
				
			||||||
		void *entry, unsigned long size, unsigned long min,
 | 
							void *entry, unsigned long size, unsigned long min,
 | 
				
			||||||
		unsigned long max, gfp_t gfp)
 | 
							unsigned long max, gfp_t gfp)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue