mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	bcachefs: Convert clock code to u64s
Eliminate possible integer truncation bugs on 32 bit Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
This commit is contained in:
		
							parent
							
								
									ec8bf491a9
								
							
						
					
					
						commit
						27d033df35
					
				
					 3 changed files with 35 additions and 42 deletions
				
			
		| 
						 | 
					@ -15,18 +15,15 @@ static inline long io_timer_cmp(io_timer_heap *h,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_timer_add(struct io_clock *clock, struct io_timer *timer)
 | 
					void bch2_io_timer_add(struct io_clock *clock, struct io_timer *timer)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	size_t i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	spin_lock(&clock->timer_lock);
 | 
						spin_lock(&clock->timer_lock);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (time_after_eq((unsigned long) atomic64_read(&clock->now),
 | 
						if (time_after_eq64((u64) atomic64_read(&clock->now), timer->expire)) {
 | 
				
			||||||
			  timer->expire)) {
 | 
					 | 
				
			||||||
		spin_unlock(&clock->timer_lock);
 | 
							spin_unlock(&clock->timer_lock);
 | 
				
			||||||
		timer->fn(timer);
 | 
							timer->fn(timer);
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < clock->timers.used; i++)
 | 
						for (size_t i = 0; i < clock->timers.used; i++)
 | 
				
			||||||
		if (clock->timers.data[i] == timer)
 | 
							if (clock->timers.data[i] == timer)
 | 
				
			||||||
			goto out;
 | 
								goto out;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -37,11 +34,9 @@ void bch2_io_timer_add(struct io_clock *clock, struct io_timer *timer)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_timer_del(struct io_clock *clock, struct io_timer *timer)
 | 
					void bch2_io_timer_del(struct io_clock *clock, struct io_timer *timer)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	size_t i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	spin_lock(&clock->timer_lock);
 | 
						spin_lock(&clock->timer_lock);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < clock->timers.used; i++)
 | 
						for (size_t i = 0; i < clock->timers.used; i++)
 | 
				
			||||||
		if (clock->timers.data[i] == timer) {
 | 
							if (clock->timers.data[i] == timer) {
 | 
				
			||||||
			heap_del(&clock->timers, i, io_timer_cmp, NULL);
 | 
								heap_del(&clock->timers, i, io_timer_cmp, NULL);
 | 
				
			||||||
			break;
 | 
								break;
 | 
				
			||||||
| 
						 | 
					@ -75,33 +70,31 @@ static void io_clock_cpu_timeout(struct timer_list *timer)
 | 
				
			||||||
	wake_up_process(wait->task);
 | 
						wake_up_process(wait->task);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_clock_schedule_timeout(struct io_clock *clock, unsigned long until)
 | 
					void bch2_io_clock_schedule_timeout(struct io_clock *clock, u64 until)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct io_clock_wait wait;
 | 
						struct io_clock_wait wait = {
 | 
				
			||||||
 | 
							.io_timer.expire	= until,
 | 
				
			||||||
 | 
							.io_timer.fn		= io_clock_wait_fn,
 | 
				
			||||||
 | 
							.io_timer.fn2		= (void *) _RET_IP_,
 | 
				
			||||||
 | 
							.task			= current,
 | 
				
			||||||
 | 
						};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* XXX: calculate sleep time rigorously */
 | 
					 | 
				
			||||||
	wait.io_timer.expire	= until;
 | 
					 | 
				
			||||||
	wait.io_timer.fn	= io_clock_wait_fn;
 | 
					 | 
				
			||||||
	wait.task		= current;
 | 
					 | 
				
			||||||
	wait.expired		= 0;
 | 
					 | 
				
			||||||
	bch2_io_timer_add(clock, &wait.io_timer);
 | 
						bch2_io_timer_add(clock, &wait.io_timer);
 | 
				
			||||||
 | 
					 | 
				
			||||||
	schedule();
 | 
						schedule();
 | 
				
			||||||
 | 
					 | 
				
			||||||
	bch2_io_timer_del(clock, &wait.io_timer);
 | 
						bch2_io_timer_del(clock, &wait.io_timer);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_kthread_io_clock_wait(struct io_clock *clock,
 | 
					void bch2_kthread_io_clock_wait(struct io_clock *clock,
 | 
				
			||||||
				unsigned long io_until,
 | 
									u64 io_until, unsigned long cpu_timeout)
 | 
				
			||||||
				unsigned long cpu_timeout)
 | 
					 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	bool kthread = (current->flags & PF_KTHREAD) != 0;
 | 
						bool kthread = (current->flags & PF_KTHREAD) != 0;
 | 
				
			||||||
	struct io_clock_wait wait;
 | 
						struct io_clock_wait wait = {
 | 
				
			||||||
 | 
							.io_timer.expire	= io_until,
 | 
				
			||||||
 | 
							.io_timer.fn		= io_clock_wait_fn,
 | 
				
			||||||
 | 
							.io_timer.fn2		= (void *) _RET_IP_,
 | 
				
			||||||
 | 
							.task			= current,
 | 
				
			||||||
 | 
						};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	wait.io_timer.expire	= io_until;
 | 
					 | 
				
			||||||
	wait.io_timer.fn	= io_clock_wait_fn;
 | 
					 | 
				
			||||||
	wait.task		= current;
 | 
					 | 
				
			||||||
	wait.expired		= 0;
 | 
					 | 
				
			||||||
	bch2_io_timer_add(clock, &wait.io_timer);
 | 
						bch2_io_timer_add(clock, &wait.io_timer);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	timer_setup_on_stack(&wait.cpu_timer, io_clock_cpu_timeout, 0);
 | 
						timer_setup_on_stack(&wait.cpu_timer, io_clock_cpu_timeout, 0);
 | 
				
			||||||
| 
						 | 
					@ -127,21 +120,20 @@ void bch2_kthread_io_clock_wait(struct io_clock *clock,
 | 
				
			||||||
	bch2_io_timer_del(clock, &wait.io_timer);
 | 
						bch2_io_timer_del(clock, &wait.io_timer);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static struct io_timer *get_expired_timer(struct io_clock *clock,
 | 
					static struct io_timer *get_expired_timer(struct io_clock *clock, u64 now)
 | 
				
			||||||
					  unsigned long now)
 | 
					 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct io_timer *ret = NULL;
 | 
						struct io_timer *ret = NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (clock->timers.used &&
 | 
						if (clock->timers.used &&
 | 
				
			||||||
	    time_after_eq(now, clock->timers.data[0]->expire))
 | 
						    time_after_eq64(now, clock->timers.data[0]->expire))
 | 
				
			||||||
		heap_pop(&clock->timers, ret, io_timer_cmp, NULL);
 | 
							heap_pop(&clock->timers, ret, io_timer_cmp, NULL);
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void __bch2_increment_clock(struct io_clock *clock, unsigned sectors)
 | 
					void __bch2_increment_clock(struct io_clock *clock, u64 sectors)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct io_timer *timer;
 | 
						struct io_timer *timer;
 | 
				
			||||||
	unsigned long now = atomic64_add_return(sectors, &clock->now);
 | 
						u64 now = atomic64_add_return(sectors, &clock->now);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	spin_lock(&clock->timer_lock);
 | 
						spin_lock(&clock->timer_lock);
 | 
				
			||||||
	while ((timer = get_expired_timer(clock, now)))
 | 
						while ((timer = get_expired_timer(clock, now)))
 | 
				
			||||||
| 
						 | 
					@ -151,17 +143,18 @@ void __bch2_increment_clock(struct io_clock *clock, unsigned sectors)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_timers_to_text(struct printbuf *out, struct io_clock *clock)
 | 
					void bch2_io_timers_to_text(struct printbuf *out, struct io_clock *clock)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long now;
 | 
					 | 
				
			||||||
	unsigned i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	out->atomic++;
 | 
						out->atomic++;
 | 
				
			||||||
	spin_lock(&clock->timer_lock);
 | 
						spin_lock(&clock->timer_lock);
 | 
				
			||||||
	now = atomic64_read(&clock->now);
 | 
						u64 now = atomic64_read(&clock->now);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < clock->timers.used; i++)
 | 
						printbuf_tabstop_push(out, 40);
 | 
				
			||||||
		prt_printf(out, "%ps:\t%li\n",
 | 
						prt_printf(out, "current time:\t%llu\n", now);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						for (unsigned i = 0; i < clock->timers.used; i++)
 | 
				
			||||||
 | 
							prt_printf(out, "%ps %ps:\t%llu\n",
 | 
				
			||||||
		       clock->timers.data[i]->fn,
 | 
							       clock->timers.data[i]->fn,
 | 
				
			||||||
		       clock->timers.data[i]->expire - now);
 | 
							       clock->timers.data[i]->fn2,
 | 
				
			||||||
 | 
							       clock->timers.data[i]->expire);
 | 
				
			||||||
	spin_unlock(&clock->timer_lock);
 | 
						spin_unlock(&clock->timer_lock);
 | 
				
			||||||
	--out->atomic;
 | 
						--out->atomic;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -4,12 +4,11 @@
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_timer_add(struct io_clock *, struct io_timer *);
 | 
					void bch2_io_timer_add(struct io_clock *, struct io_timer *);
 | 
				
			||||||
void bch2_io_timer_del(struct io_clock *, struct io_timer *);
 | 
					void bch2_io_timer_del(struct io_clock *, struct io_timer *);
 | 
				
			||||||
void bch2_kthread_io_clock_wait(struct io_clock *, unsigned long,
 | 
					void bch2_kthread_io_clock_wait(struct io_clock *, u64, unsigned long);
 | 
				
			||||||
				unsigned long);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
void __bch2_increment_clock(struct io_clock *, unsigned);
 | 
					void __bch2_increment_clock(struct io_clock *, u64);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static inline void bch2_increment_clock(struct bch_fs *c, unsigned sectors,
 | 
					static inline void bch2_increment_clock(struct bch_fs *c, u64 sectors,
 | 
				
			||||||
					int rw)
 | 
										int rw)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct io_clock *clock = &c->io_clock[rw];
 | 
						struct io_clock *clock = &c->io_clock[rw];
 | 
				
			||||||
| 
						 | 
					@ -19,7 +18,7 @@ static inline void bch2_increment_clock(struct bch_fs *c, unsigned sectors,
 | 
				
			||||||
		__bch2_increment_clock(clock, this_cpu_xchg(*clock->pcpu_buf, 0));
 | 
							__bch2_increment_clock(clock, this_cpu_xchg(*clock->pcpu_buf, 0));
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void bch2_io_clock_schedule_timeout(struct io_clock *, unsigned long);
 | 
					void bch2_io_clock_schedule_timeout(struct io_clock *, u64);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define bch2_kthread_wait_event_ioclock_timeout(condition, clock, timeout)\
 | 
					#define bch2_kthread_wait_event_ioclock_timeout(condition, clock, timeout)\
 | 
				
			||||||
({									\
 | 
					({									\
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -17,7 +17,8 @@ typedef void (*io_timer_fn)(struct io_timer *);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
struct io_timer {
 | 
					struct io_timer {
 | 
				
			||||||
	io_timer_fn		fn;
 | 
						io_timer_fn		fn;
 | 
				
			||||||
	unsigned long		expire;
 | 
						void			*fn2;
 | 
				
			||||||
 | 
						u64			expire;
 | 
				
			||||||
};
 | 
					};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* Amount to buffer up on a percpu counter */
 | 
					/* Amount to buffer up on a percpu counter */
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue