mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 08:38:45 +02:00 
			
		
		
		
	 8fa7292fee
			
		
	
	
		8fa7292fee
		
	
	
	
	
		
			
			timer_delete[_sync]() replaces del_timer[_sync](). Convert the whole tree over and remove the historical wrapper inlines. Conversion was done with coccinelle plus manual fixups where necessary. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Ingo Molnar <mingo@kernel.org>
		
			
				
	
	
		
			539 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			539 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-or-later
 | |
| /* AFS fileserver probing
 | |
|  *
 | |
|  * Copyright (C) 2018, 2020 Red Hat, Inc. All Rights Reserved.
 | |
|  * Written by David Howells (dhowells@redhat.com)
 | |
|  */
 | |
| 
 | |
| #include <linux/sched.h>
 | |
| #include <linux/slab.h>
 | |
| #include "afs_fs.h"
 | |
| #include "internal.h"
 | |
| #include "protocol_afs.h"
 | |
| #include "protocol_yfs.h"
 | |
| 
 | |
| static unsigned int afs_fs_probe_fast_poll_interval = 30 * HZ;
 | |
| static unsigned int afs_fs_probe_slow_poll_interval = 5 * 60 * HZ;
 | |
| 
 | |
| struct afs_endpoint_state *afs_get_endpoint_state(struct afs_endpoint_state *estate,
 | |
| 						  enum afs_estate_trace where)
 | |
| {
 | |
| 	if (estate) {
 | |
| 		int r;
 | |
| 
 | |
| 		__refcount_inc(&estate->ref, &r);
 | |
| 		trace_afs_estate(estate->server_id, estate->probe_seq, r, where);
 | |
| 	}
 | |
| 	return estate;
 | |
| }
 | |
| 
 | |
| static void afs_endpoint_state_rcu(struct rcu_head *rcu)
 | |
| {
 | |
| 	struct afs_endpoint_state *estate = container_of(rcu, struct afs_endpoint_state, rcu);
 | |
| 
 | |
| 	trace_afs_estate(estate->server_id, estate->probe_seq, refcount_read(&estate->ref),
 | |
| 			 afs_estate_trace_free);
 | |
| 	afs_put_addrlist(estate->addresses, afs_alist_trace_put_estate);
 | |
| 	kfree(estate);
 | |
| }
 | |
| 
 | |
| void afs_put_endpoint_state(struct afs_endpoint_state *estate, enum afs_estate_trace where)
 | |
| {
 | |
| 	if (estate) {
 | |
| 		unsigned int server_id = estate->server_id, probe_seq = estate->probe_seq;
 | |
| 		bool dead;
 | |
| 		int r;
 | |
| 
 | |
| 		dead = __refcount_dec_and_test(&estate->ref, &r);
 | |
| 		trace_afs_estate(server_id, probe_seq, r, where);
 | |
| 		if (dead)
 | |
| 			call_rcu(&estate->rcu, afs_endpoint_state_rcu);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Start the probe polling timer.  We have to supply it with an inc on the
 | |
|  * outstanding server count.
 | |
|  */
 | |
| static void afs_schedule_fs_probe(struct afs_net *net,
 | |
| 				  struct afs_server *server, bool fast)
 | |
| {
 | |
| 	unsigned long atj;
 | |
| 
 | |
| 	if (!net->live)
 | |
| 		return;
 | |
| 
 | |
| 	atj = server->probed_at;
 | |
| 	atj += fast ? afs_fs_probe_fast_poll_interval : afs_fs_probe_slow_poll_interval;
 | |
| 
 | |
| 	afs_inc_servers_outstanding(net);
 | |
| 	if (timer_reduce(&net->fs_probe_timer, atj))
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Handle the completion of a set of probes.
 | |
|  */
 | |
| static void afs_finished_fs_probe(struct afs_net *net, struct afs_server *server,
 | |
| 				  struct afs_endpoint_state *estate)
 | |
| {
 | |
| 	bool responded = test_bit(AFS_ESTATE_RESPONDED, &estate->flags);
 | |
| 
 | |
| 	write_seqlock(&net->fs_lock);
 | |
| 	if (responded) {
 | |
| 		list_add_tail(&server->probe_link, &net->fs_probe_slow);
 | |
| 	} else {
 | |
| 		server->rtt = UINT_MAX;
 | |
| 		clear_bit(AFS_SERVER_FL_RESPONDING, &server->flags);
 | |
| 		list_add_tail(&server->probe_link, &net->fs_probe_fast);
 | |
| 	}
 | |
| 
 | |
| 	write_sequnlock(&net->fs_lock);
 | |
| 
 | |
| 	afs_schedule_fs_probe(net, server, !responded);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Handle the completion of a probe.
 | |
|  */
 | |
| static void afs_done_one_fs_probe(struct afs_net *net, struct afs_server *server,
 | |
| 				  struct afs_endpoint_state *estate)
 | |
| {
 | |
| 	_enter("");
 | |
| 
 | |
| 	if (atomic_dec_and_test(&estate->nr_probing))
 | |
| 		afs_finished_fs_probe(net, server, estate);
 | |
| 
 | |
| 	wake_up_all(&server->probe_wq);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Handle inability to send a probe due to ENOMEM when trying to allocate a
 | |
|  * call struct.
 | |
|  */
 | |
| static void afs_fs_probe_not_done(struct afs_net *net,
 | |
| 				  struct afs_server *server,
 | |
| 				  struct afs_endpoint_state *estate,
 | |
| 				  int index)
 | |
| {
 | |
| 	_enter("");
 | |
| 
 | |
| 	trace_afs_io_error(0, -ENOMEM, afs_io_error_fs_probe_fail);
 | |
| 	spin_lock(&server->probe_lock);
 | |
| 
 | |
| 	set_bit(AFS_ESTATE_LOCAL_FAILURE, &estate->flags);
 | |
| 	if (estate->error == 0)
 | |
| 		estate->error = -ENOMEM;
 | |
| 
 | |
| 	set_bit(index, &estate->failed_set);
 | |
| 
 | |
| 	spin_unlock(&server->probe_lock);
 | |
| 	return afs_done_one_fs_probe(net, server, estate);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Process the result of probing a fileserver.  This is called after successful
 | |
|  * or failed delivery of an FS.GetCapabilities operation.
 | |
|  */
 | |
| void afs_fileserver_probe_result(struct afs_call *call)
 | |
| {
 | |
| 	struct afs_endpoint_state *estate = call->probe;
 | |
| 	struct afs_addr_list *alist = estate->addresses;
 | |
| 	struct afs_address *addr = &alist->addrs[call->probe_index];
 | |
| 	struct afs_server *server = call->server;
 | |
| 	unsigned int index = call->probe_index;
 | |
| 	unsigned int rtt_us = -1, cap0;
 | |
| 	int ret = call->error;
 | |
| 
 | |
| 	_enter("%pU,%u", &server->uuid, index);
 | |
| 
 | |
| 	WRITE_ONCE(addr->last_error, ret);
 | |
| 
 | |
| 	spin_lock(&server->probe_lock);
 | |
| 
 | |
| 	switch (ret) {
 | |
| 	case 0:
 | |
| 		estate->error = 0;
 | |
| 		goto responded;
 | |
| 	case -ECONNABORTED:
 | |
| 		if (!test_bit(AFS_ESTATE_RESPONDED, &estate->flags)) {
 | |
| 			estate->abort_code = call->abort_code;
 | |
| 			estate->error = ret;
 | |
| 		}
 | |
| 		goto responded;
 | |
| 	case -ENOMEM:
 | |
| 	case -ENONET:
 | |
| 		clear_bit(index, &estate->responsive_set);
 | |
| 		set_bit(AFS_ESTATE_LOCAL_FAILURE, &estate->flags);
 | |
| 		trace_afs_io_error(call->debug_id, ret, afs_io_error_fs_probe_fail);
 | |
| 		goto out;
 | |
| 	case -ECONNRESET: /* Responded, but call expired. */
 | |
| 	case -ERFKILL:
 | |
| 	case -EADDRNOTAVAIL:
 | |
| 	case -ENETUNREACH:
 | |
| 	case -EHOSTUNREACH:
 | |
| 	case -EHOSTDOWN:
 | |
| 	case -ECONNREFUSED:
 | |
| 	case -ETIMEDOUT:
 | |
| 	case -ETIME:
 | |
| 	default:
 | |
| 		clear_bit(index, &estate->responsive_set);
 | |
| 		set_bit(index, &estate->failed_set);
 | |
| 		if (!test_bit(AFS_ESTATE_RESPONDED, &estate->flags) &&
 | |
| 		    (estate->error == 0 ||
 | |
| 		     estate->error == -ETIMEDOUT ||
 | |
| 		     estate->error == -ETIME))
 | |
| 			estate->error = ret;
 | |
| 		trace_afs_io_error(call->debug_id, ret, afs_io_error_fs_probe_fail);
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| responded:
 | |
| 	clear_bit(index, &estate->failed_set);
 | |
| 
 | |
| 	if (call->service_id == YFS_FS_SERVICE) {
 | |
| 		set_bit(AFS_ESTATE_IS_YFS, &estate->flags);
 | |
| 		set_bit(AFS_SERVER_FL_IS_YFS, &server->flags);
 | |
| 		server->service_id = call->service_id;
 | |
| 	} else {
 | |
| 		set_bit(AFS_ESTATE_NOT_YFS, &estate->flags);
 | |
| 		if (!test_bit(AFS_ESTATE_IS_YFS, &estate->flags)) {
 | |
| 			clear_bit(AFS_SERVER_FL_IS_YFS, &server->flags);
 | |
| 			server->service_id = call->service_id;
 | |
| 		}
 | |
| 		cap0 = ntohl(call->tmp);
 | |
| 		if (cap0 & AFS3_VICED_CAPABILITY_64BITFILES)
 | |
| 			set_bit(AFS_SERVER_FL_HAS_FS64, &server->flags);
 | |
| 		else
 | |
| 			clear_bit(AFS_SERVER_FL_HAS_FS64, &server->flags);
 | |
| 	}
 | |
| 
 | |
| 	rtt_us = rxrpc_kernel_get_srtt(addr->peer);
 | |
| 	if (rtt_us < estate->rtt) {
 | |
| 		estate->rtt = rtt_us;
 | |
| 		server->rtt = rtt_us;
 | |
| 		alist->preferred = index;
 | |
| 	}
 | |
| 
 | |
| 	smp_wmb(); /* Set rtt before responded. */
 | |
| 	set_bit(AFS_ESTATE_RESPONDED, &estate->flags);
 | |
| 	set_bit(index, &estate->responsive_set);
 | |
| 	set_bit(AFS_SERVER_FL_RESPONDING, &server->flags);
 | |
| out:
 | |
| 	spin_unlock(&server->probe_lock);
 | |
| 
 | |
| 	trace_afs_fs_probe(server, false, estate, index, call->error, call->abort_code, rtt_us);
 | |
| 	_debug("probe[%x] %pU [%u] %pISpc rtt=%d ret=%d",
 | |
| 	       estate->probe_seq, &server->uuid, index,
 | |
| 	       rxrpc_kernel_remote_addr(alist->addrs[index].peer),
 | |
| 	       rtt_us, ret);
 | |
| 
 | |
| 	return afs_done_one_fs_probe(call->net, server, estate);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Probe all of a fileserver's addresses to find out the best route and to
 | |
|  * query its capabilities.
 | |
|  */
 | |
| int afs_fs_probe_fileserver(struct afs_net *net, struct afs_server *server,
 | |
| 			    struct afs_addr_list *new_alist, struct key *key)
 | |
| {
 | |
| 	struct afs_endpoint_state *estate, *old;
 | |
| 	struct afs_addr_list *old_alist = NULL, *alist;
 | |
| 	unsigned long unprobed;
 | |
| 
 | |
| 	_enter("%pU", &server->uuid);
 | |
| 
 | |
| 	estate = kzalloc(sizeof(*estate), GFP_KERNEL);
 | |
| 	if (!estate)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	refcount_set(&estate->ref, 2);
 | |
| 	estate->server_id = server->debug_id;
 | |
| 	estate->rtt = UINT_MAX;
 | |
| 
 | |
| 	write_lock(&server->fs_lock);
 | |
| 
 | |
| 	old = rcu_dereference_protected(server->endpoint_state,
 | |
| 					lockdep_is_held(&server->fs_lock));
 | |
| 	if (old) {
 | |
| 		estate->responsive_set = old->responsive_set;
 | |
| 		if (!new_alist)
 | |
| 			new_alist = old->addresses;
 | |
| 	}
 | |
| 
 | |
| 	if (old_alist != new_alist)
 | |
| 		afs_set_peer_appdata(server, old_alist, new_alist);
 | |
| 
 | |
| 	estate->addresses = afs_get_addrlist(new_alist, afs_alist_trace_get_estate);
 | |
| 	alist = estate->addresses;
 | |
| 	estate->probe_seq = ++server->probe_counter;
 | |
| 	atomic_set(&estate->nr_probing, alist->nr_addrs);
 | |
| 
 | |
| 	if (new_alist)
 | |
| 		server->addr_version = new_alist->version;
 | |
| 	rcu_assign_pointer(server->endpoint_state, estate);
 | |
| 	write_unlock(&server->fs_lock);
 | |
| 	if (old)
 | |
| 		set_bit(AFS_ESTATE_SUPERSEDED, &old->flags);
 | |
| 
 | |
| 	trace_afs_estate(estate->server_id, estate->probe_seq, refcount_read(&estate->ref),
 | |
| 			 afs_estate_trace_alloc_probe);
 | |
| 
 | |
| 	afs_get_address_preferences(net, new_alist);
 | |
| 
 | |
| 	server->probed_at = jiffies;
 | |
| 	unprobed = (1UL << alist->nr_addrs) - 1;
 | |
| 	while (unprobed) {
 | |
| 		unsigned int index = 0, i;
 | |
| 		int best_prio = -1;
 | |
| 
 | |
| 		for (i = 0; i < alist->nr_addrs; i++) {
 | |
| 			if (test_bit(i, &unprobed) &&
 | |
| 			    alist->addrs[i].prio > best_prio) {
 | |
| 				index = i;
 | |
| 				best_prio = alist->addrs[i].prio;
 | |
| 			}
 | |
| 		}
 | |
| 		__clear_bit(index, &unprobed);
 | |
| 
 | |
| 		trace_afs_fs_probe(server, true, estate, index, 0, 0, 0);
 | |
| 		if (!afs_fs_get_capabilities(net, server, estate, index, key))
 | |
| 			afs_fs_probe_not_done(net, server, estate, index);
 | |
| 	}
 | |
| 
 | |
| 	afs_put_endpoint_state(old, afs_estate_trace_put_probe);
 | |
| 	afs_put_endpoint_state(estate, afs_estate_trace_put_probe);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Wait for the first as-yet untried fileserver to respond, for the probe state
 | |
|  * to be superseded or for all probes to finish.
 | |
|  */
 | |
| int afs_wait_for_fs_probes(struct afs_operation *op, struct afs_server_state *states, bool intr)
 | |
| {
 | |
| 	struct afs_endpoint_state *estate;
 | |
| 	struct afs_server_list *slist = op->server_list;
 | |
| 	bool still_probing = true;
 | |
| 	int ret = 0, i;
 | |
| 
 | |
| 	_enter("%u", slist->nr_servers);
 | |
| 
 | |
| 	for (i = 0; i < slist->nr_servers; i++) {
 | |
| 		estate = states[i].endpoint_state;
 | |
| 		if (test_bit(AFS_ESTATE_SUPERSEDED, &estate->flags))
 | |
| 			return 2;
 | |
| 		if (atomic_read(&estate->nr_probing))
 | |
| 			still_probing = true;
 | |
| 		if (estate->responsive_set & states[i].untried_addrs)
 | |
| 			return 1;
 | |
| 	}
 | |
| 	if (!still_probing)
 | |
| 		return 0;
 | |
| 
 | |
| 	for (i = 0; i < slist->nr_servers; i++)
 | |
| 		add_wait_queue(&slist->servers[i].server->probe_wq, &states[i].probe_waiter);
 | |
| 
 | |
| 	for (;;) {
 | |
| 		still_probing = false;
 | |
| 
 | |
| 		set_current_state(intr ? TASK_INTERRUPTIBLE : TASK_UNINTERRUPTIBLE);
 | |
| 		for (i = 0; i < slist->nr_servers; i++) {
 | |
| 			estate = states[i].endpoint_state;
 | |
| 			if (test_bit(AFS_ESTATE_SUPERSEDED, &estate->flags)) {
 | |
| 				ret = 2;
 | |
| 				goto stop;
 | |
| 			}
 | |
| 			if (atomic_read(&estate->nr_probing))
 | |
| 				still_probing = true;
 | |
| 			if (estate->responsive_set & states[i].untried_addrs) {
 | |
| 				ret = 1;
 | |
| 				goto stop;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if (!still_probing || signal_pending(current))
 | |
| 			goto stop;
 | |
| 		schedule();
 | |
| 	}
 | |
| 
 | |
| stop:
 | |
| 	set_current_state(TASK_RUNNING);
 | |
| 
 | |
| 	for (i = 0; i < slist->nr_servers; i++)
 | |
| 		remove_wait_queue(&slist->servers[i].server->probe_wq, &states[i].probe_waiter);
 | |
| 
 | |
| 	if (!ret && signal_pending(current))
 | |
| 		ret = -ERESTARTSYS;
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Probe timer.  We have an increment on fs_outstanding that we need to pass
 | |
|  * along to the work item.
 | |
|  */
 | |
| void afs_fs_probe_timer(struct timer_list *timer)
 | |
| {
 | |
| 	struct afs_net *net = container_of(timer, struct afs_net, fs_probe_timer);
 | |
| 
 | |
| 	if (!net->live || !queue_work(afs_wq, &net->fs_prober))
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispatch a probe to a server.
 | |
|  */
 | |
| static void afs_dispatch_fs_probe(struct afs_net *net, struct afs_server *server)
 | |
| 	__releases(&net->fs_lock)
 | |
| {
 | |
| 	struct key *key = NULL;
 | |
| 
 | |
| 	/* We remove it from the queues here - it will be added back to
 | |
| 	 * one of the queues on the completion of the probe.
 | |
| 	 */
 | |
| 	list_del_init(&server->probe_link);
 | |
| 
 | |
| 	afs_get_server(server, afs_server_trace_get_probe);
 | |
| 	write_sequnlock(&net->fs_lock);
 | |
| 
 | |
| 	afs_fs_probe_fileserver(net, server, NULL, key);
 | |
| 	afs_put_server(net, server, afs_server_trace_put_probe);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Probe a server immediately without waiting for its due time to come
 | |
|  * round.  This is used when all of the addresses have been tried.
 | |
|  */
 | |
| void afs_probe_fileserver(struct afs_net *net, struct afs_server *server)
 | |
| {
 | |
| 	write_seqlock(&net->fs_lock);
 | |
| 	if (!list_empty(&server->probe_link))
 | |
| 		return afs_dispatch_fs_probe(net, server);
 | |
| 	write_sequnlock(&net->fs_lock);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Probe dispatcher to regularly dispatch probes to keep NAT alive.
 | |
|  */
 | |
| void afs_fs_probe_dispatcher(struct work_struct *work)
 | |
| {
 | |
| 	struct afs_net *net = container_of(work, struct afs_net, fs_prober);
 | |
| 	struct afs_server *fast, *slow, *server;
 | |
| 	unsigned long nowj, timer_at, poll_at;
 | |
| 	bool first_pass = true, set_timer = false;
 | |
| 
 | |
| 	if (!net->live) {
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	_enter("");
 | |
| 
 | |
| 	if (list_empty(&net->fs_probe_fast) && list_empty(&net->fs_probe_slow)) {
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| 		_leave(" [none]");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| again:
 | |
| 	write_seqlock(&net->fs_lock);
 | |
| 
 | |
| 	fast = slow = server = NULL;
 | |
| 	nowj = jiffies;
 | |
| 	timer_at = nowj + MAX_JIFFY_OFFSET;
 | |
| 
 | |
| 	if (!list_empty(&net->fs_probe_fast)) {
 | |
| 		fast = list_first_entry(&net->fs_probe_fast, struct afs_server, probe_link);
 | |
| 		poll_at = fast->probed_at + afs_fs_probe_fast_poll_interval;
 | |
| 		if (time_before(nowj, poll_at)) {
 | |
| 			timer_at = poll_at;
 | |
| 			set_timer = true;
 | |
| 			fast = NULL;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (!list_empty(&net->fs_probe_slow)) {
 | |
| 		slow = list_first_entry(&net->fs_probe_slow, struct afs_server, probe_link);
 | |
| 		poll_at = slow->probed_at + afs_fs_probe_slow_poll_interval;
 | |
| 		if (time_before(nowj, poll_at)) {
 | |
| 			if (time_before(poll_at, timer_at))
 | |
| 			    timer_at = poll_at;
 | |
| 			set_timer = true;
 | |
| 			slow = NULL;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	server = fast ?: slow;
 | |
| 	if (server)
 | |
| 		_debug("probe %pU", &server->uuid);
 | |
| 
 | |
| 	if (server && (first_pass || !need_resched())) {
 | |
| 		afs_dispatch_fs_probe(net, server);
 | |
| 		first_pass = false;
 | |
| 		goto again;
 | |
| 	}
 | |
| 
 | |
| 	write_sequnlock(&net->fs_lock);
 | |
| 
 | |
| 	if (server) {
 | |
| 		if (!queue_work(afs_wq, &net->fs_prober))
 | |
| 			afs_dec_servers_outstanding(net);
 | |
| 		_leave(" [requeue]");
 | |
| 	} else if (set_timer) {
 | |
| 		if (timer_reduce(&net->fs_probe_timer, timer_at))
 | |
| 			afs_dec_servers_outstanding(net);
 | |
| 		_leave(" [timer]");
 | |
| 	} else {
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| 		_leave(" [quiesce]");
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Wait for a probe on a particular fileserver to complete for 2s.
 | |
|  */
 | |
| int afs_wait_for_one_fs_probe(struct afs_server *server, struct afs_endpoint_state *estate,
 | |
| 			      unsigned long exclude, bool is_intr)
 | |
| {
 | |
| 	struct wait_queue_entry wait;
 | |
| 	unsigned long timo = 2 * HZ;
 | |
| 
 | |
| 	if (atomic_read(&estate->nr_probing) == 0)
 | |
| 		goto dont_wait;
 | |
| 
 | |
| 	init_wait_entry(&wait, 0);
 | |
| 	for (;;) {
 | |
| 		prepare_to_wait_event(&server->probe_wq, &wait,
 | |
| 				      is_intr ? TASK_INTERRUPTIBLE : TASK_UNINTERRUPTIBLE);
 | |
| 		if (timo == 0 ||
 | |
| 		    test_bit(AFS_ESTATE_SUPERSEDED, &estate->flags) ||
 | |
| 		    (estate->responsive_set & ~exclude) ||
 | |
| 		    atomic_read(&estate->nr_probing) == 0 ||
 | |
| 		    (is_intr && signal_pending(current)))
 | |
| 			break;
 | |
| 		timo = schedule_timeout(timo);
 | |
| 	}
 | |
| 
 | |
| 	finish_wait(&server->probe_wq, &wait);
 | |
| 
 | |
| dont_wait:
 | |
| 	if (test_bit(AFS_ESTATE_SUPERSEDED, &estate->flags))
 | |
| 		return 0;
 | |
| 	if (estate->responsive_set & ~exclude)
 | |
| 		return 1;
 | |
| 	if (is_intr && signal_pending(current))
 | |
| 		return -ERESTARTSYS;
 | |
| 	if (timo == 0)
 | |
| 		return -ETIME;
 | |
| 	return -EDESTADDRREQ;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Clean up the probing when the namespace is killed off.
 | |
|  */
 | |
| void afs_fs_probe_cleanup(struct afs_net *net)
 | |
| {
 | |
| 	if (timer_delete_sync(&net->fs_probe_timer))
 | |
| 		afs_dec_servers_outstanding(net);
 | |
| }
 |