mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	sched/pelt: Add a new runnable average signal
Now that runnable_load_avg has been removed, we can replace it by a new signal that will highlight the runnable pressure on a cfs_rq. This signal track the waiting time of tasks on rq and can help to better define the state of rqs. At now, only util_avg is used to define the state of a rq: A rq with more that around 80% of utilization and more than 1 tasks is considered as overloaded. But the util_avg signal of a rq can become temporaly low after that a task migrated onto another rq which can bias the classification of the rq. When tasks compete for the same rq, their runnable average signal will be higher than util_avg as it will include the waiting time and we can use this signal to better classify cfs_rqs. The new runnable_avg will track the runnable time of a task which simply adds the waiting time to the running time. The runnable _avg of cfs_rq will be the /Sum of se's runnable_avg and the runnable_avg of group entity will follow the one of the rq similarly to util_avg. Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> Signed-off-by: Mel Gorman <mgorman@techsingularity.net> Signed-off-by: Ingo Molnar <mingo@kernel.org> Reviewed-by: "Dietmar Eggemann <dietmar.eggemann@arm.com>" Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Juri Lelli <juri.lelli@redhat.com> Cc: Valentin Schneider <valentin.schneider@arm.com> Cc: Phil Auld <pauld@redhat.com> Cc: Hillf Danton <hdanton@sina.com> Link: https://lore.kernel.org/r/20200224095223.13361-9-mgorman@techsingularity.net
This commit is contained in:
		
							parent
							
								
									0dacee1bfa
								
							
						
					
					
						commit
						9f68395333
					
				
					 5 changed files with 142 additions and 31 deletions
				
			
		| 
						 | 
					@ -356,28 +356,30 @@ struct util_est {
 | 
				
			||||||
} __attribute__((__aligned__(sizeof(u64))));
 | 
					} __attribute__((__aligned__(sizeof(u64))));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * The load_avg/util_avg accumulates an infinite geometric series
 | 
					 * The load/runnable/util_avg accumulates an infinite geometric series
 | 
				
			||||||
 * (see __update_load_avg_cfs_rq() in kernel/sched/pelt.c).
 | 
					 * (see __update_load_avg_cfs_rq() in kernel/sched/pelt.c).
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * [load_avg definition]
 | 
					 * [load_avg definition]
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   load_avg = runnable% * scale_load_down(load)
 | 
					 *   load_avg = runnable% * scale_load_down(load)
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * where runnable% is the time ratio that a sched_entity is runnable.
 | 
					 * [runnable_avg definition]
 | 
				
			||||||
 * For cfs_rq, it is the aggregated load_avg of all runnable and
 | 
					 *
 | 
				
			||||||
 * blocked sched_entities.
 | 
					 *   runnable_avg = runnable% * SCHED_CAPACITY_SCALE
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * [util_avg definition]
 | 
					 * [util_avg definition]
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   util_avg = running% * SCHED_CAPACITY_SCALE
 | 
					 *   util_avg = running% * SCHED_CAPACITY_SCALE
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * where running% is the time ratio that a sched_entity is running on
 | 
					 * where runnable% is the time ratio that a sched_entity is runnable and
 | 
				
			||||||
 * a CPU. For cfs_rq, it is the aggregated util_avg of all runnable
 | 
					 * running% the time ratio that a sched_entity is running.
 | 
				
			||||||
 * and blocked sched_entities.
 | 
					 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * load_avg and util_avg don't direcly factor frequency scaling and CPU
 | 
					 * For cfs_rq, they are the aggregated values of all runnable and blocked
 | 
				
			||||||
 * capacity scaling. The scaling is done through the rq_clock_pelt that
 | 
					 * sched_entities.
 | 
				
			||||||
 * is used for computing those signals (see update_rq_clock_pelt())
 | 
					 *
 | 
				
			||||||
 | 
					 * The load/runnable/util_avg doesn't direcly factor frequency scaling and CPU
 | 
				
			||||||
 | 
					 * capacity scaling. The scaling is done through the rq_clock_pelt that is used
 | 
				
			||||||
 | 
					 * for computing those signals (see update_rq_clock_pelt())
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * N.B., the above ratios (runnable% and running%) themselves are in the
 | 
					 * N.B., the above ratios (runnable% and running%) themselves are in the
 | 
				
			||||||
 * range of [0, 1]. To do fixed point arithmetics, we therefore scale them
 | 
					 * range of [0, 1]. To do fixed point arithmetics, we therefore scale them
 | 
				
			||||||
| 
						 | 
					@ -401,9 +403,11 @@ struct util_est {
 | 
				
			||||||
struct sched_avg {
 | 
					struct sched_avg {
 | 
				
			||||||
	u64				last_update_time;
 | 
						u64				last_update_time;
 | 
				
			||||||
	u64				load_sum;
 | 
						u64				load_sum;
 | 
				
			||||||
 | 
						u64				runnable_sum;
 | 
				
			||||||
	u32				util_sum;
 | 
						u32				util_sum;
 | 
				
			||||||
	u32				period_contrib;
 | 
						u32				period_contrib;
 | 
				
			||||||
	unsigned long			load_avg;
 | 
						unsigned long			load_avg;
 | 
				
			||||||
 | 
						unsigned long			runnable_avg;
 | 
				
			||||||
	unsigned long			util_avg;
 | 
						unsigned long			util_avg;
 | 
				
			||||||
	struct util_est			util_est;
 | 
						struct util_est			util_est;
 | 
				
			||||||
} ____cacheline_aligned;
 | 
					} ____cacheline_aligned;
 | 
				
			||||||
| 
						 | 
					@ -467,6 +471,8 @@ struct sched_entity {
 | 
				
			||||||
	struct cfs_rq			*cfs_rq;
 | 
						struct cfs_rq			*cfs_rq;
 | 
				
			||||||
	/* rq "owned" by this entity/group: */
 | 
						/* rq "owned" by this entity/group: */
 | 
				
			||||||
	struct cfs_rq			*my_q;
 | 
						struct cfs_rq			*my_q;
 | 
				
			||||||
 | 
						/* cached value of my_q->h_nr_running */
 | 
				
			||||||
 | 
						unsigned long			runnable_weight;
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_SMP
 | 
					#ifdef CONFIG_SMP
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -405,6 +405,7 @@ static void print_cfs_group_stats(struct seq_file *m, int cpu, struct task_group
 | 
				
			||||||
#ifdef CONFIG_SMP
 | 
					#ifdef CONFIG_SMP
 | 
				
			||||||
	P(se->avg.load_avg);
 | 
						P(se->avg.load_avg);
 | 
				
			||||||
	P(se->avg.util_avg);
 | 
						P(se->avg.util_avg);
 | 
				
			||||||
 | 
						P(se->avg.runnable_avg);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#undef PN_SCHEDSTAT
 | 
					#undef PN_SCHEDSTAT
 | 
				
			||||||
| 
						 | 
					@ -524,6 +525,8 @@ void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
 | 
				
			||||||
#ifdef CONFIG_SMP
 | 
					#ifdef CONFIG_SMP
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %lu\n", "load_avg",
 | 
						SEQ_printf(m, "  .%-30s: %lu\n", "load_avg",
 | 
				
			||||||
			cfs_rq->avg.load_avg);
 | 
								cfs_rq->avg.load_avg);
 | 
				
			||||||
 | 
						SEQ_printf(m, "  .%-30s: %lu\n", "runnable_avg",
 | 
				
			||||||
 | 
								cfs_rq->avg.runnable_avg);
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %lu\n", "util_avg",
 | 
						SEQ_printf(m, "  .%-30s: %lu\n", "util_avg",
 | 
				
			||||||
			cfs_rq->avg.util_avg);
 | 
								cfs_rq->avg.util_avg);
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %u\n", "util_est_enqueued",
 | 
						SEQ_printf(m, "  .%-30s: %u\n", "util_est_enqueued",
 | 
				
			||||||
| 
						 | 
					@ -532,8 +535,8 @@ void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
 | 
				
			||||||
			cfs_rq->removed.load_avg);
 | 
								cfs_rq->removed.load_avg);
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %ld\n", "removed.util_avg",
 | 
						SEQ_printf(m, "  .%-30s: %ld\n", "removed.util_avg",
 | 
				
			||||||
			cfs_rq->removed.util_avg);
 | 
								cfs_rq->removed.util_avg);
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %ld\n", "removed.runnable_sum",
 | 
						SEQ_printf(m, "  .%-30s: %ld\n", "removed.runnable_avg",
 | 
				
			||||||
			cfs_rq->removed.runnable_sum);
 | 
								cfs_rq->removed.runnable_avg);
 | 
				
			||||||
#ifdef CONFIG_FAIR_GROUP_SCHED
 | 
					#ifdef CONFIG_FAIR_GROUP_SCHED
 | 
				
			||||||
	SEQ_printf(m, "  .%-30s: %lu\n", "tg_load_avg_contrib",
 | 
						SEQ_printf(m, "  .%-30s: %lu\n", "tg_load_avg_contrib",
 | 
				
			||||||
			cfs_rq->tg_load_avg_contrib);
 | 
								cfs_rq->tg_load_avg_contrib);
 | 
				
			||||||
| 
						 | 
					@ -944,8 +947,10 @@ void proc_sched_show_task(struct task_struct *p, struct pid_namespace *ns,
 | 
				
			||||||
	P(se.load.weight);
 | 
						P(se.load.weight);
 | 
				
			||||||
#ifdef CONFIG_SMP
 | 
					#ifdef CONFIG_SMP
 | 
				
			||||||
	P(se.avg.load_sum);
 | 
						P(se.avg.load_sum);
 | 
				
			||||||
 | 
						P(se.avg.runnable_sum);
 | 
				
			||||||
	P(se.avg.util_sum);
 | 
						P(se.avg.util_sum);
 | 
				
			||||||
	P(se.avg.load_avg);
 | 
						P(se.avg.load_avg);
 | 
				
			||||||
 | 
						P(se.avg.runnable_avg);
 | 
				
			||||||
	P(se.avg.util_avg);
 | 
						P(se.avg.util_avg);
 | 
				
			||||||
	P(se.avg.last_update_time);
 | 
						P(se.avg.last_update_time);
 | 
				
			||||||
	P(se.avg.util_est.ewma);
 | 
						P(se.avg.util_est.ewma);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -794,6 +794,8 @@ void post_init_entity_util_avg(struct task_struct *p)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						sa->runnable_avg = cpu_scale;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (p->sched_class != &fair_sched_class) {
 | 
						if (p->sched_class != &fair_sched_class) {
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * For !fair tasks do:
 | 
							 * For !fair tasks do:
 | 
				
			||||||
| 
						 | 
					@ -3215,9 +3217,9 @@ void set_task_rq_fair(struct sched_entity *se,
 | 
				
			||||||
 * _IFF_ we look at the pure running and runnable sums. Because they
 | 
					 * _IFF_ we look at the pure running and runnable sums. Because they
 | 
				
			||||||
 * represent the very same entity, just at different points in the hierarchy.
 | 
					 * represent the very same entity, just at different points in the hierarchy.
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * Per the above update_tg_cfs_util() is trivial  * and simply copies the
 | 
					 * Per the above update_tg_cfs_util() and update_tg_cfs_runnable() are trivial
 | 
				
			||||||
 * running sum over (but still wrong, because the group entity and group rq do
 | 
					 * and simply copies the running/runnable sum over (but still wrong, because
 | 
				
			||||||
 * not have their PELT windows aligned).
 | 
					 * the group entity and group rq do not have their PELT windows aligned).
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * However, update_tg_cfs_load() is more complex. So we have:
 | 
					 * However, update_tg_cfs_load() is more complex. So we have:
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
| 
						 | 
					@ -3299,6 +3301,32 @@ update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq
 | 
				
			||||||
	cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * LOAD_AVG_MAX;
 | 
						cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * LOAD_AVG_MAX;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline void
 | 
				
			||||||
 | 
					update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						long delta = gcfs_rq->avg.runnable_avg - se->avg.runnable_avg;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* Nothing to update */
 | 
				
			||||||
 | 
						if (!delta)
 | 
				
			||||||
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/*
 | 
				
			||||||
 | 
						 * The relation between sum and avg is:
 | 
				
			||||||
 | 
						 *
 | 
				
			||||||
 | 
						 *   LOAD_AVG_MAX - 1024 + sa->period_contrib
 | 
				
			||||||
 | 
						 *
 | 
				
			||||||
 | 
						 * however, the PELT windows are not aligned between grq and gse.
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* Set new sched_entity's runnable */
 | 
				
			||||||
 | 
						se->avg.runnable_avg = gcfs_rq->avg.runnable_avg;
 | 
				
			||||||
 | 
						se->avg.runnable_sum = se->avg.runnable_avg * LOAD_AVG_MAX;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* Update parent cfs_rq runnable */
 | 
				
			||||||
 | 
						add_positive(&cfs_rq->avg.runnable_avg, delta);
 | 
				
			||||||
 | 
						cfs_rq->avg.runnable_sum = cfs_rq->avg.runnable_avg * LOAD_AVG_MAX;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static inline void
 | 
					static inline void
 | 
				
			||||||
update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
 | 
					update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
| 
						 | 
					@ -3379,6 +3407,7 @@ static inline int propagate_entity_load_avg(struct sched_entity *se)
 | 
				
			||||||
	add_tg_cfs_propagate(cfs_rq, gcfs_rq->prop_runnable_sum);
 | 
						add_tg_cfs_propagate(cfs_rq, gcfs_rq->prop_runnable_sum);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	update_tg_cfs_util(cfs_rq, se, gcfs_rq);
 | 
						update_tg_cfs_util(cfs_rq, se, gcfs_rq);
 | 
				
			||||||
 | 
						update_tg_cfs_runnable(cfs_rq, se, gcfs_rq);
 | 
				
			||||||
	update_tg_cfs_load(cfs_rq, se, gcfs_rq);
 | 
						update_tg_cfs_load(cfs_rq, se, gcfs_rq);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	trace_pelt_cfs_tp(cfs_rq);
 | 
						trace_pelt_cfs_tp(cfs_rq);
 | 
				
			||||||
| 
						 | 
					@ -3449,7 +3478,7 @@ static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum
 | 
				
			||||||
static inline int
 | 
					static inline int
 | 
				
			||||||
update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
 | 
					update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long removed_load = 0, removed_util = 0, removed_runnable_sum = 0;
 | 
						unsigned long removed_load = 0, removed_util = 0, removed_runnable = 0;
 | 
				
			||||||
	struct sched_avg *sa = &cfs_rq->avg;
 | 
						struct sched_avg *sa = &cfs_rq->avg;
 | 
				
			||||||
	int decayed = 0;
 | 
						int decayed = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3460,7 +3489,7 @@ update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
 | 
				
			||||||
		raw_spin_lock(&cfs_rq->removed.lock);
 | 
							raw_spin_lock(&cfs_rq->removed.lock);
 | 
				
			||||||
		swap(cfs_rq->removed.util_avg, removed_util);
 | 
							swap(cfs_rq->removed.util_avg, removed_util);
 | 
				
			||||||
		swap(cfs_rq->removed.load_avg, removed_load);
 | 
							swap(cfs_rq->removed.load_avg, removed_load);
 | 
				
			||||||
		swap(cfs_rq->removed.runnable_sum, removed_runnable_sum);
 | 
							swap(cfs_rq->removed.runnable_avg, removed_runnable);
 | 
				
			||||||
		cfs_rq->removed.nr = 0;
 | 
							cfs_rq->removed.nr = 0;
 | 
				
			||||||
		raw_spin_unlock(&cfs_rq->removed.lock);
 | 
							raw_spin_unlock(&cfs_rq->removed.lock);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3472,7 +3501,16 @@ update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
 | 
				
			||||||
		sub_positive(&sa->util_avg, r);
 | 
							sub_positive(&sa->util_avg, r);
 | 
				
			||||||
		sub_positive(&sa->util_sum, r * divider);
 | 
							sub_positive(&sa->util_sum, r * divider);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		add_tg_cfs_propagate(cfs_rq, -(long)removed_runnable_sum);
 | 
							r = removed_runnable;
 | 
				
			||||||
 | 
							sub_positive(&sa->runnable_avg, r);
 | 
				
			||||||
 | 
							sub_positive(&sa->runnable_sum, r * divider);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							/*
 | 
				
			||||||
 | 
							 * removed_runnable is the unweighted version of removed_load so we
 | 
				
			||||||
 | 
							 * can use it to estimate removed_load_sum.
 | 
				
			||||||
 | 
							 */
 | 
				
			||||||
 | 
							add_tg_cfs_propagate(cfs_rq,
 | 
				
			||||||
 | 
								-(long)(removed_runnable * divider) >> SCHED_CAPACITY_SHIFT);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		decayed = 1;
 | 
							decayed = 1;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					@ -3517,6 +3555,8 @@ static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *s
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	se->avg.util_sum = se->avg.util_avg * divider;
 | 
						se->avg.util_sum = se->avg.util_avg * divider;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						se->avg.runnable_sum = se->avg.runnable_avg * divider;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	se->avg.load_sum = divider;
 | 
						se->avg.load_sum = divider;
 | 
				
			||||||
	if (se_weight(se)) {
 | 
						if (se_weight(se)) {
 | 
				
			||||||
		se->avg.load_sum =
 | 
							se->avg.load_sum =
 | 
				
			||||||
| 
						 | 
					@ -3526,6 +3566,8 @@ static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *s
 | 
				
			||||||
	enqueue_load_avg(cfs_rq, se);
 | 
						enqueue_load_avg(cfs_rq, se);
 | 
				
			||||||
	cfs_rq->avg.util_avg += se->avg.util_avg;
 | 
						cfs_rq->avg.util_avg += se->avg.util_avg;
 | 
				
			||||||
	cfs_rq->avg.util_sum += se->avg.util_sum;
 | 
						cfs_rq->avg.util_sum += se->avg.util_sum;
 | 
				
			||||||
 | 
						cfs_rq->avg.runnable_avg += se->avg.runnable_avg;
 | 
				
			||||||
 | 
						cfs_rq->avg.runnable_sum += se->avg.runnable_sum;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	add_tg_cfs_propagate(cfs_rq, se->avg.load_sum);
 | 
						add_tg_cfs_propagate(cfs_rq, se->avg.load_sum);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3547,6 +3589,8 @@ static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *s
 | 
				
			||||||
	dequeue_load_avg(cfs_rq, se);
 | 
						dequeue_load_avg(cfs_rq, se);
 | 
				
			||||||
	sub_positive(&cfs_rq->avg.util_avg, se->avg.util_avg);
 | 
						sub_positive(&cfs_rq->avg.util_avg, se->avg.util_avg);
 | 
				
			||||||
	sub_positive(&cfs_rq->avg.util_sum, se->avg.util_sum);
 | 
						sub_positive(&cfs_rq->avg.util_sum, se->avg.util_sum);
 | 
				
			||||||
 | 
						sub_positive(&cfs_rq->avg.runnable_avg, se->avg.runnable_avg);
 | 
				
			||||||
 | 
						sub_positive(&cfs_rq->avg.runnable_sum, se->avg.runnable_sum);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	add_tg_cfs_propagate(cfs_rq, -se->avg.load_sum);
 | 
						add_tg_cfs_propagate(cfs_rq, -se->avg.load_sum);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3653,10 +3697,15 @@ static void remove_entity_load_avg(struct sched_entity *se)
 | 
				
			||||||
	++cfs_rq->removed.nr;
 | 
						++cfs_rq->removed.nr;
 | 
				
			||||||
	cfs_rq->removed.util_avg	+= se->avg.util_avg;
 | 
						cfs_rq->removed.util_avg	+= se->avg.util_avg;
 | 
				
			||||||
	cfs_rq->removed.load_avg	+= se->avg.load_avg;
 | 
						cfs_rq->removed.load_avg	+= se->avg.load_avg;
 | 
				
			||||||
	cfs_rq->removed.runnable_sum	+= se->avg.load_sum; /* == runnable_sum */
 | 
						cfs_rq->removed.runnable_avg	+= se->avg.runnable_avg;
 | 
				
			||||||
	raw_spin_unlock_irqrestore(&cfs_rq->removed.lock, flags);
 | 
						raw_spin_unlock_irqrestore(&cfs_rq->removed.lock, flags);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline unsigned long cfs_rq_runnable_avg(struct cfs_rq *cfs_rq)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return cfs_rq->avg.runnable_avg;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static inline unsigned long cfs_rq_load_avg(struct cfs_rq *cfs_rq)
 | 
					static inline unsigned long cfs_rq_load_avg(struct cfs_rq *cfs_rq)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	return cfs_rq->avg.load_avg;
 | 
						return cfs_rq->avg.load_avg;
 | 
				
			||||||
| 
						 | 
					@ -3983,11 +4032,13 @@ enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
	 * When enqueuing a sched_entity, we must:
 | 
						 * When enqueuing a sched_entity, we must:
 | 
				
			||||||
	 *   - Update loads to have both entity and cfs_rq synced with now.
 | 
						 *   - Update loads to have both entity and cfs_rq synced with now.
 | 
				
			||||||
 | 
						 *   - Add its load to cfs_rq->runnable_avg
 | 
				
			||||||
	 *   - For group_entity, update its weight to reflect the new share of
 | 
						 *   - For group_entity, update its weight to reflect the new share of
 | 
				
			||||||
	 *     its group cfs_rq
 | 
						 *     its group cfs_rq
 | 
				
			||||||
	 *   - Add its new weight to cfs_rq->load.weight
 | 
						 *   - Add its new weight to cfs_rq->load.weight
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	update_load_avg(cfs_rq, se, UPDATE_TG | DO_ATTACH);
 | 
						update_load_avg(cfs_rq, se, UPDATE_TG | DO_ATTACH);
 | 
				
			||||||
 | 
						se_update_runnable(se);
 | 
				
			||||||
	update_cfs_group(se);
 | 
						update_cfs_group(se);
 | 
				
			||||||
	account_entity_enqueue(cfs_rq, se);
 | 
						account_entity_enqueue(cfs_rq, se);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -4065,11 +4116,13 @@ dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
	 * When dequeuing a sched_entity, we must:
 | 
						 * When dequeuing a sched_entity, we must:
 | 
				
			||||||
	 *   - Update loads to have both entity and cfs_rq synced with now.
 | 
						 *   - Update loads to have both entity and cfs_rq synced with now.
 | 
				
			||||||
 | 
						 *   - Subtract its load from the cfs_rq->runnable_avg.
 | 
				
			||||||
	 *   - Subtract its previous weight from cfs_rq->load.weight.
 | 
						 *   - Subtract its previous weight from cfs_rq->load.weight.
 | 
				
			||||||
	 *   - For group entity, update its weight to reflect the new share
 | 
						 *   - For group entity, update its weight to reflect the new share
 | 
				
			||||||
	 *     of its group cfs_rq.
 | 
						 *     of its group cfs_rq.
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
						update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
				
			||||||
 | 
						se_update_runnable(se);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	update_stats_dequeue(cfs_rq, se, flags);
 | 
						update_stats_dequeue(cfs_rq, se, flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -5240,6 +5293,7 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
 | 
				
			||||||
			goto enqueue_throttle;
 | 
								goto enqueue_throttle;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
							update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
				
			||||||
 | 
							se_update_runnable(se);
 | 
				
			||||||
		update_cfs_group(se);
 | 
							update_cfs_group(se);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		cfs_rq->h_nr_running++;
 | 
							cfs_rq->h_nr_running++;
 | 
				
			||||||
| 
						 | 
					@ -5337,6 +5391,7 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
 | 
				
			||||||
			goto dequeue_throttle;
 | 
								goto dequeue_throttle;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
							update_load_avg(cfs_rq, se, UPDATE_TG);
 | 
				
			||||||
 | 
							se_update_runnable(se);
 | 
				
			||||||
		update_cfs_group(se);
 | 
							update_cfs_group(se);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		cfs_rq->h_nr_running--;
 | 
							cfs_rq->h_nr_running--;
 | 
				
			||||||
| 
						 | 
					@ -5409,6 +5464,11 @@ static unsigned long cpu_load_without(struct rq *rq, struct task_struct *p)
 | 
				
			||||||
	return load;
 | 
						return load;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static unsigned long cpu_runnable(struct rq *rq)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return cfs_rq_runnable_avg(&rq->cfs);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static unsigned long capacity_of(int cpu)
 | 
					static unsigned long capacity_of(int cpu)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	return cpu_rq(cpu)->cpu_capacity;
 | 
						return cpu_rq(cpu)->cpu_capacity;
 | 
				
			||||||
| 
						 | 
					@ -7554,6 +7614,9 @@ static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq)
 | 
				
			||||||
	if (cfs_rq->avg.util_sum)
 | 
						if (cfs_rq->avg.util_sum)
 | 
				
			||||||
		return false;
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (cfs_rq->avg.runnable_sum)
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return true;
 | 
						return true;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -108,7 +108,7 @@ static u32 __accumulate_pelt_segments(u64 periods, u32 d1, u32 d3)
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
static __always_inline u32
 | 
					static __always_inline u32
 | 
				
			||||||
accumulate_sum(u64 delta, struct sched_avg *sa,
 | 
					accumulate_sum(u64 delta, struct sched_avg *sa,
 | 
				
			||||||
	       unsigned long load, int running)
 | 
						       unsigned long load, unsigned long runnable, int running)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	u32 contrib = (u32)delta; /* p == 0 -> delta < 1024 */
 | 
						u32 contrib = (u32)delta; /* p == 0 -> delta < 1024 */
 | 
				
			||||||
	u64 periods;
 | 
						u64 periods;
 | 
				
			||||||
| 
						 | 
					@ -121,6 +121,8 @@ accumulate_sum(u64 delta, struct sched_avg *sa,
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (periods) {
 | 
						if (periods) {
 | 
				
			||||||
		sa->load_sum = decay_load(sa->load_sum, periods);
 | 
							sa->load_sum = decay_load(sa->load_sum, periods);
 | 
				
			||||||
 | 
							sa->runnable_sum =
 | 
				
			||||||
 | 
								decay_load(sa->runnable_sum, periods);
 | 
				
			||||||
		sa->util_sum = decay_load((u64)(sa->util_sum), periods);
 | 
							sa->util_sum = decay_load((u64)(sa->util_sum), periods);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
| 
						 | 
					@ -146,6 +148,8 @@ accumulate_sum(u64 delta, struct sched_avg *sa,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (load)
 | 
						if (load)
 | 
				
			||||||
		sa->load_sum += load * contrib;
 | 
							sa->load_sum += load * contrib;
 | 
				
			||||||
 | 
						if (runnable)
 | 
				
			||||||
 | 
							sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT;
 | 
				
			||||||
	if (running)
 | 
						if (running)
 | 
				
			||||||
		sa->util_sum += contrib << SCHED_CAPACITY_SHIFT;
 | 
							sa->util_sum += contrib << SCHED_CAPACITY_SHIFT;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -182,7 +186,7 @@ accumulate_sum(u64 delta, struct sched_avg *sa,
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
static __always_inline int
 | 
					static __always_inline int
 | 
				
			||||||
___update_load_sum(u64 now, struct sched_avg *sa,
 | 
					___update_load_sum(u64 now, struct sched_avg *sa,
 | 
				
			||||||
		  unsigned long load, int running)
 | 
							  unsigned long load, unsigned long runnable, int running)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	u64 delta;
 | 
						u64 delta;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -218,7 +222,7 @@ ___update_load_sum(u64 now, struct sched_avg *sa,
 | 
				
			||||||
	 * Also see the comment in accumulate_sum().
 | 
						 * Also see the comment in accumulate_sum().
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (!load)
 | 
						if (!load)
 | 
				
			||||||
		running = 0;
 | 
							runnable = running = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						/*
 | 
				
			||||||
	 * Now we know we crossed measurement unit boundaries. The *_avg
 | 
						 * Now we know we crossed measurement unit boundaries. The *_avg
 | 
				
			||||||
| 
						 | 
					@ -227,7 +231,7 @@ ___update_load_sum(u64 now, struct sched_avg *sa,
 | 
				
			||||||
	 * Step 1: accumulate *_sum since last_update_time. If we haven't
 | 
						 * Step 1: accumulate *_sum since last_update_time. If we haven't
 | 
				
			||||||
	 * crossed period boundaries, finish.
 | 
						 * crossed period boundaries, finish.
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (!accumulate_sum(delta, sa, load, running))
 | 
						if (!accumulate_sum(delta, sa, load, runnable, running))
 | 
				
			||||||
		return 0;
 | 
							return 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 1;
 | 
						return 1;
 | 
				
			||||||
| 
						 | 
					@ -242,6 +246,7 @@ ___update_load_avg(struct sched_avg *sa, unsigned long load)
 | 
				
			||||||
	 * Step 2: update *_avg.
 | 
						 * Step 2: update *_avg.
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	sa->load_avg = div_u64(load * sa->load_sum, divider);
 | 
						sa->load_avg = div_u64(load * sa->load_sum, divider);
 | 
				
			||||||
 | 
						sa->runnable_avg = div_u64(sa->runnable_sum, divider);
 | 
				
			||||||
	WRITE_ONCE(sa->util_avg, sa->util_sum / divider);
 | 
						WRITE_ONCE(sa->util_avg, sa->util_sum / divider);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -250,24 +255,30 @@ ___update_load_avg(struct sched_avg *sa, unsigned long load)
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   task:
 | 
					 *   task:
 | 
				
			||||||
 *     se_weight()   = se->load.weight
 | 
					 *     se_weight()   = se->load.weight
 | 
				
			||||||
 | 
					 *     se_runnable() = !!on_rq
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   group: [ see update_cfs_group() ]
 | 
					 *   group: [ see update_cfs_group() ]
 | 
				
			||||||
 *     se_weight()   = tg->weight * grq->load_avg / tg->load_avg
 | 
					 *     se_weight()   = tg->weight * grq->load_avg / tg->load_avg
 | 
				
			||||||
 | 
					 *     se_runnable() = grq->h_nr_running
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 | 
					 *   runnable_sum = se_runnable() * runnable = grq->runnable_sum
 | 
				
			||||||
 | 
					 *   runnable_avg = runnable_sum
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   load_sum := runnable
 | 
					 *   load_sum := runnable
 | 
				
			||||||
 *   load_avg = se_weight(se) * load_sum
 | 
					 *   load_avg = se_weight(se) * load_sum
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 * XXX collapse load_sum and runnable_load_sum
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * cfq_rq:
 | 
					 * cfq_rq:
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 | 
					 *   runnable_sum = \Sum se->avg.runnable_sum
 | 
				
			||||||
 | 
					 *   runnable_avg = \Sum se->avg.runnable_avg
 | 
				
			||||||
 | 
					 *
 | 
				
			||||||
 *   load_sum = \Sum se_weight(se) * se->avg.load_sum
 | 
					 *   load_sum = \Sum se_weight(se) * se->avg.load_sum
 | 
				
			||||||
 *   load_avg = \Sum se->avg.load_avg
 | 
					 *   load_avg = \Sum se->avg.load_avg
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __update_load_avg_blocked_se(u64 now, struct sched_entity *se)
 | 
					int __update_load_avg_blocked_se(u64 now, struct sched_entity *se)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (___update_load_sum(now, &se->avg, 0, 0)) {
 | 
						if (___update_load_sum(now, &se->avg, 0, 0, 0)) {
 | 
				
			||||||
		___update_load_avg(&se->avg, se_weight(se));
 | 
							___update_load_avg(&se->avg, se_weight(se));
 | 
				
			||||||
		trace_pelt_se_tp(se);
 | 
							trace_pelt_se_tp(se);
 | 
				
			||||||
		return 1;
 | 
							return 1;
 | 
				
			||||||
| 
						 | 
					@ -278,7 +289,8 @@ int __update_load_avg_blocked_se(u64 now, struct sched_entity *se)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se)
 | 
					int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (___update_load_sum(now, &se->avg, !!se->on_rq, cfs_rq->curr == se)) {
 | 
						if (___update_load_sum(now, &se->avg, !!se->on_rq, se_runnable(se),
 | 
				
			||||||
 | 
									cfs_rq->curr == se)) {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		___update_load_avg(&se->avg, se_weight(se));
 | 
							___update_load_avg(&se->avg, se_weight(se));
 | 
				
			||||||
		cfs_se_util_change(&se->avg);
 | 
							cfs_se_util_change(&se->avg);
 | 
				
			||||||
| 
						 | 
					@ -293,6 +305,7 @@ int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (___update_load_sum(now, &cfs_rq->avg,
 | 
						if (___update_load_sum(now, &cfs_rq->avg,
 | 
				
			||||||
				scale_load_down(cfs_rq->load.weight),
 | 
									scale_load_down(cfs_rq->load.weight),
 | 
				
			||||||
 | 
									cfs_rq->h_nr_running,
 | 
				
			||||||
				cfs_rq->curr != NULL)) {
 | 
									cfs_rq->curr != NULL)) {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		___update_load_avg(&cfs_rq->avg, 1);
 | 
							___update_load_avg(&cfs_rq->avg, 1);
 | 
				
			||||||
| 
						 | 
					@ -310,13 +323,14 @@ int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq)
 | 
				
			||||||
 *   util_sum = cpu_scale * load_sum
 | 
					 *   util_sum = cpu_scale * load_sum
 | 
				
			||||||
 *   runnable_sum = util_sum
 | 
					 *   runnable_sum = util_sum
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   load_avg is not supported and meaningless.
 | 
					 *   load_avg and runnable_avg are not supported and meaningless.
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int update_rt_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
					int update_rt_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (___update_load_sum(now, &rq->avg_rt,
 | 
						if (___update_load_sum(now, &rq->avg_rt,
 | 
				
			||||||
 | 
									running,
 | 
				
			||||||
				running,
 | 
									running,
 | 
				
			||||||
				running)) {
 | 
									running)) {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -335,13 +349,14 @@ int update_rt_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
				
			||||||
 *   util_sum = cpu_scale * load_sum
 | 
					 *   util_sum = cpu_scale * load_sum
 | 
				
			||||||
 *   runnable_sum = util_sum
 | 
					 *   runnable_sum = util_sum
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   load_avg is not supported and meaningless.
 | 
					 *   load_avg and runnable_avg are not supported and meaningless.
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
					int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (___update_load_sum(now, &rq->avg_dl,
 | 
						if (___update_load_sum(now, &rq->avg_dl,
 | 
				
			||||||
 | 
									running,
 | 
				
			||||||
				running,
 | 
									running,
 | 
				
			||||||
				running)) {
 | 
									running)) {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -361,7 +376,7 @@ int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
 | 
				
			||||||
 *   util_sum = cpu_scale * load_sum
 | 
					 *   util_sum = cpu_scale * load_sum
 | 
				
			||||||
 *   runnable_sum = util_sum
 | 
					 *   runnable_sum = util_sum
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 *   load_avg is not supported and meaningless.
 | 
					 *   load_avg and runnable_avg are not supported and meaningless.
 | 
				
			||||||
 *
 | 
					 *
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -389,9 +404,11 @@ int update_irq_load_avg(struct rq *rq, u64 running)
 | 
				
			||||||
	 * rq->clock += delta with delta >= running
 | 
						 * rq->clock += delta with delta >= running
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	ret = ___update_load_sum(rq->clock - running, &rq->avg_irq,
 | 
						ret = ___update_load_sum(rq->clock - running, &rq->avg_irq,
 | 
				
			||||||
 | 
									0,
 | 
				
			||||||
				0,
 | 
									0,
 | 
				
			||||||
				0);
 | 
									0);
 | 
				
			||||||
	ret += ___update_load_sum(rq->clock, &rq->avg_irq,
 | 
						ret += ___update_load_sum(rq->clock, &rq->avg_irq,
 | 
				
			||||||
 | 
									1,
 | 
				
			||||||
				1,
 | 
									1,
 | 
				
			||||||
				1);
 | 
									1);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -527,7 +527,7 @@ struct cfs_rq {
 | 
				
			||||||
		int		nr;
 | 
							int		nr;
 | 
				
			||||||
		unsigned long	load_avg;
 | 
							unsigned long	load_avg;
 | 
				
			||||||
		unsigned long	util_avg;
 | 
							unsigned long	util_avg;
 | 
				
			||||||
		unsigned long	runnable_sum;
 | 
							unsigned long	runnable_avg;
 | 
				
			||||||
	} removed;
 | 
						} removed;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_FAIR_GROUP_SCHED
 | 
					#ifdef CONFIG_FAIR_GROUP_SCHED
 | 
				
			||||||
| 
						 | 
					@ -688,9 +688,29 @@ struct dl_rq {
 | 
				
			||||||
/* An entity is a task if it doesn't "own" a runqueue */
 | 
					/* An entity is a task if it doesn't "own" a runqueue */
 | 
				
			||||||
#define entity_is_task(se)	(!se->my_q)
 | 
					#define entity_is_task(se)	(!se->my_q)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline void se_update_runnable(struct sched_entity *se)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						if (!entity_is_task(se))
 | 
				
			||||||
 | 
							se->runnable_weight = se->my_q->h_nr_running;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline long se_runnable(struct sched_entity *se)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						if (entity_is_task(se))
 | 
				
			||||||
 | 
							return !!se->on_rq;
 | 
				
			||||||
 | 
						else
 | 
				
			||||||
 | 
							return se->runnable_weight;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
#define entity_is_task(se)	1
 | 
					#define entity_is_task(se)	1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline void se_update_runnable(struct sched_entity *se) {}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline long se_runnable(struct sched_entity *se)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return !!se->on_rq;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_SMP
 | 
					#ifdef CONFIG_SMP
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue