forked from mirrors/linux
		
	blkcg: change blkg reference counting to use percpu_ref
Now that every bio is associated with a blkg, this puts the use of blkg_get, blkg_try_get, and blkg_put on the hot path. This switches over the refcnt in blkg to use percpu_ref. Signed-off-by: Dennis Zhou <dennisszhou@gmail.com> Acked-by: Tejun Heo <tj@kernel.org> Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
		
							parent
							
								
									e2b0989954
								
							
						
					
					
						commit
						b3b9f24f5f
					
				
					 2 changed files with 44 additions and 35 deletions
				
			
		|  | @ -84,6 +84,37 @@ static void blkg_free(struct blkcg_gq *blkg) | ||||||
| 	kfree(blkg); | 	kfree(blkg); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | static void __blkg_release(struct rcu_head *rcu) | ||||||
|  | { | ||||||
|  | 	struct blkcg_gq *blkg = container_of(rcu, struct blkcg_gq, rcu_head); | ||||||
|  | 
 | ||||||
|  | 	percpu_ref_exit(&blkg->refcnt); | ||||||
|  | 
 | ||||||
|  | 	/* release the blkcg and parent blkg refs this blkg has been holding */ | ||||||
|  | 	css_put(&blkg->blkcg->css); | ||||||
|  | 	if (blkg->parent) | ||||||
|  | 		blkg_put(blkg->parent); | ||||||
|  | 
 | ||||||
|  | 	wb_congested_put(blkg->wb_congested); | ||||||
|  | 
 | ||||||
|  | 	blkg_free(blkg); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | /*
 | ||||||
|  |  * A group is RCU protected, but having an rcu lock does not mean that one | ||||||
|  |  * can access all the fields of blkg and assume these are valid.  For | ||||||
|  |  * example, don't try to follow throtl_data and request queue links. | ||||||
|  |  * | ||||||
|  |  * Having a reference to blkg under an rcu allows accesses to only values | ||||||
|  |  * local to groups like group stats and group rate limits. | ||||||
|  |  */ | ||||||
|  | static void blkg_release(struct percpu_ref *ref) | ||||||
|  | { | ||||||
|  | 	struct blkcg_gq *blkg = container_of(ref, struct blkcg_gq, refcnt); | ||||||
|  | 
 | ||||||
|  | 	call_rcu(&blkg->rcu_head, __blkg_release); | ||||||
|  | } | ||||||
|  | 
 | ||||||
| /**
 | /**
 | ||||||
|  * blkg_alloc - allocate a blkg |  * blkg_alloc - allocate a blkg | ||||||
|  * @blkcg: block cgroup the new blkg is associated with |  * @blkcg: block cgroup the new blkg is associated with | ||||||
|  | @ -110,7 +141,6 @@ static struct blkcg_gq *blkg_alloc(struct blkcg *blkcg, struct request_queue *q, | ||||||
| 	blkg->q = q; | 	blkg->q = q; | ||||||
| 	INIT_LIST_HEAD(&blkg->q_node); | 	INIT_LIST_HEAD(&blkg->q_node); | ||||||
| 	blkg->blkcg = blkcg; | 	blkg->blkcg = blkcg; | ||||||
| 	atomic_set(&blkg->refcnt, 1); |  | ||||||
| 
 | 
 | ||||||
| 	/* root blkg uses @q->root_rl, init rl only for !root blkgs */ | 	/* root blkg uses @q->root_rl, init rl only for !root blkgs */ | ||||||
| 	if (blkcg != &blkcg_root) { | 	if (blkcg != &blkcg_root) { | ||||||
|  | @ -217,6 +247,11 @@ static struct blkcg_gq *blkg_create(struct blkcg *blkcg, | ||||||
| 		blkg_get(blkg->parent); | 		blkg_get(blkg->parent); | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
|  | 	ret = percpu_ref_init(&blkg->refcnt, blkg_release, 0, | ||||||
|  | 			      GFP_NOWAIT | __GFP_NOWARN); | ||||||
|  | 	if (ret) | ||||||
|  | 		goto err_cancel_ref; | ||||||
|  | 
 | ||||||
| 	/* invoke per-policy init */ | 	/* invoke per-policy init */ | ||||||
| 	for (i = 0; i < BLKCG_MAX_POLS; i++) { | 	for (i = 0; i < BLKCG_MAX_POLS; i++) { | ||||||
| 		struct blkcg_policy *pol = blkcg_policy[i]; | 		struct blkcg_policy *pol = blkcg_policy[i]; | ||||||
|  | @ -249,6 +284,8 @@ static struct blkcg_gq *blkg_create(struct blkcg *blkcg, | ||||||
| 	blkg_put(blkg); | 	blkg_put(blkg); | ||||||
| 	return ERR_PTR(ret); | 	return ERR_PTR(ret); | ||||||
| 
 | 
 | ||||||
|  | err_cancel_ref: | ||||||
|  | 	percpu_ref_exit(&blkg->refcnt); | ||||||
| err_put_congested: | err_put_congested: | ||||||
| 	wb_congested_put(wb_congested); | 	wb_congested_put(wb_congested); | ||||||
| err_put_css: | err_put_css: | ||||||
|  | @ -387,7 +424,7 @@ static void blkg_destroy(struct blkcg_gq *blkg) | ||||||
| 	 * Put the reference taken at the time of creation so that when all | 	 * Put the reference taken at the time of creation so that when all | ||||||
| 	 * queues are gone, group can be destroyed. | 	 * queues are gone, group can be destroyed. | ||||||
| 	 */ | 	 */ | ||||||
| 	blkg_put(blkg); | 	percpu_ref_kill(&blkg->refcnt); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /**
 | /**
 | ||||||
|  | @ -414,29 +451,6 @@ static void blkg_destroy_all(struct request_queue *q) | ||||||
| 	q->root_rl.blkg = NULL; | 	q->root_rl.blkg = NULL; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /*
 |  | ||||||
|  * A group is RCU protected, but having an rcu lock does not mean that one |  | ||||||
|  * can access all the fields of blkg and assume these are valid.  For |  | ||||||
|  * example, don't try to follow throtl_data and request queue links. |  | ||||||
|  * |  | ||||||
|  * Having a reference to blkg under an rcu allows accesses to only values |  | ||||||
|  * local to groups like group stats and group rate limits. |  | ||||||
|  */ |  | ||||||
| void __blkg_release_rcu(struct rcu_head *rcu_head) |  | ||||||
| { |  | ||||||
| 	struct blkcg_gq *blkg = container_of(rcu_head, struct blkcg_gq, rcu_head); |  | ||||||
| 
 |  | ||||||
| 	/* release the blkcg and parent blkg refs this blkg has been holding */ |  | ||||||
| 	css_put(&blkg->blkcg->css); |  | ||||||
| 	if (blkg->parent) |  | ||||||
| 		blkg_put(blkg->parent); |  | ||||||
| 
 |  | ||||||
| 	wb_congested_put(blkg->wb_congested); |  | ||||||
| 
 |  | ||||||
| 	blkg_free(blkg); |  | ||||||
| } |  | ||||||
| EXPORT_SYMBOL_GPL(__blkg_release_rcu); |  | ||||||
| 
 |  | ||||||
| /*
 | /*
 | ||||||
|  * The next function used by blk_queue_for_each_rl().  It's a bit tricky |  * The next function used by blk_queue_for_each_rl().  It's a bit tricky | ||||||
|  * because the root blkg uses @q->root_rl instead of its own rl. |  * because the root blkg uses @q->root_rl instead of its own rl. | ||||||
|  |  | ||||||
|  | @ -126,7 +126,7 @@ struct blkcg_gq { | ||||||
| 	struct request_list		rl; | 	struct request_list		rl; | ||||||
| 
 | 
 | ||||||
| 	/* reference count */ | 	/* reference count */ | ||||||
| 	atomic_t			refcnt; | 	struct percpu_ref		refcnt; | ||||||
| 
 | 
 | ||||||
| 	/* is this blkg online? protected by both blkcg and q locks */ | 	/* is this blkg online? protected by both blkcg and q locks */ | ||||||
| 	bool				online; | 	bool				online; | ||||||
|  | @ -490,8 +490,7 @@ static inline int blkg_path(struct blkcg_gq *blkg, char *buf, int buflen) | ||||||
|  */ |  */ | ||||||
| static inline void blkg_get(struct blkcg_gq *blkg) | static inline void blkg_get(struct blkcg_gq *blkg) | ||||||
| { | { | ||||||
| 	WARN_ON_ONCE(atomic_read(&blkg->refcnt) <= 0); | 	percpu_ref_get(&blkg->refcnt); | ||||||
| 	atomic_inc(&blkg->refcnt); |  | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /**
 | /**
 | ||||||
|  | @ -503,7 +502,7 @@ static inline void blkg_get(struct blkcg_gq *blkg) | ||||||
|  */ |  */ | ||||||
| static inline struct blkcg_gq *blkg_try_get(struct blkcg_gq *blkg) | static inline struct blkcg_gq *blkg_try_get(struct blkcg_gq *blkg) | ||||||
| { | { | ||||||
| 	if (atomic_inc_not_zero(&blkg->refcnt)) | 	if (percpu_ref_tryget(&blkg->refcnt)) | ||||||
| 		return blkg; | 		return blkg; | ||||||
| 	return NULL; | 	return NULL; | ||||||
| } | } | ||||||
|  | @ -517,23 +516,19 @@ static inline struct blkcg_gq *blkg_try_get(struct blkcg_gq *blkg) | ||||||
|  */ |  */ | ||||||
| static inline struct blkcg_gq *blkg_try_get_closest(struct blkcg_gq *blkg) | static inline struct blkcg_gq *blkg_try_get_closest(struct blkcg_gq *blkg) | ||||||
| { | { | ||||||
| 	while (!atomic_inc_not_zero(&blkg->refcnt)) | 	while (!percpu_ref_tryget(&blkg->refcnt)) | ||||||
| 		blkg = blkg->parent; | 		blkg = blkg->parent; | ||||||
| 
 | 
 | ||||||
| 	return blkg; | 	return blkg; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| void __blkg_release_rcu(struct rcu_head *rcu); |  | ||||||
| 
 |  | ||||||
| /**
 | /**
 | ||||||
|  * blkg_put - put a blkg reference |  * blkg_put - put a blkg reference | ||||||
|  * @blkg: blkg to put |  * @blkg: blkg to put | ||||||
|  */ |  */ | ||||||
| static inline void blkg_put(struct blkcg_gq *blkg) | static inline void blkg_put(struct blkcg_gq *blkg) | ||||||
| { | { | ||||||
| 	WARN_ON_ONCE(atomic_read(&blkg->refcnt) <= 0); | 	percpu_ref_put(&blkg->refcnt); | ||||||
| 	if (atomic_dec_and_test(&blkg->refcnt)) |  | ||||||
| 		call_rcu(&blkg->rcu_head, __blkg_release_rcu); |  | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /**
 | /**
 | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Dennis Zhou (Facebook)
						Dennis Zhou (Facebook)