mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 16:48:26 +02:00 
			
		
		
		
	fs: get rid of the res2 iocb->ki_complete argument
The second argument was only used by the USB gadget code, yet everyone pays the overhead of passing a zero to be passed into aio, where it ends up being part of the aio res2 value. Now that everybody is passing in zero, kill off the extra argument. Reviewed-by: Darrick J. Wong <djwong@kernel.org> Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
		
							parent
							
								
									4c3d8accdc
								
							
						
					
					
						commit
						6b19b766e8
					
				
					 18 changed files with 33 additions and 34 deletions
				
			
		|  | @ -164,7 +164,7 @@ static void blkdev_bio_end_io(struct bio *bio) | |||
| 				ret = blk_status_to_errno(dio->bio.bi_status); | ||||
| 			} | ||||
| 
 | ||||
| 			dio->iocb->ki_complete(iocb, ret, 0); | ||||
| 			dio->iocb->ki_complete(iocb, ret); | ||||
| 			if (dio->flags & DIO_MULTI_BIO) | ||||
| 				bio_put(&dio->bio); | ||||
| 		} else { | ||||
|  |  | |||
|  | @ -1076,7 +1076,7 @@ void af_alg_async_cb(struct crypto_async_request *_req, int err) | |||
| 	af_alg_free_resources(areq); | ||||
| 	sock_put(sk); | ||||
| 
 | ||||
| 	iocb->ki_complete(iocb, err ? err : (int)resultlen, 0); | ||||
| 	iocb->ki_complete(iocb, err ? err : (int)resultlen); | ||||
| } | ||||
| EXPORT_SYMBOL_GPL(af_alg_async_cb); | ||||
| 
 | ||||
|  |  | |||
|  | @ -554,7 +554,7 @@ static void lo_rw_aio_do_completion(struct loop_cmd *cmd) | |||
| 		blk_mq_complete_request(rq); | ||||
| } | ||||
| 
 | ||||
| static void lo_rw_aio_complete(struct kiocb *iocb, long ret, long ret2) | ||||
| static void lo_rw_aio_complete(struct kiocb *iocb, long ret) | ||||
| { | ||||
| 	struct loop_cmd *cmd = container_of(iocb, struct loop_cmd, iocb); | ||||
| 
 | ||||
|  | @ -627,7 +627,7 @@ static int lo_rw_aio(struct loop_device *lo, struct loop_cmd *cmd, | |||
| 	lo_rw_aio_do_completion(cmd); | ||||
| 
 | ||||
| 	if (ret != -EIOCBQUEUED) | ||||
| 		cmd->iocb.ki_complete(&cmd->iocb, ret, 0); | ||||
| 		lo_rw_aio_complete(&cmd->iocb, ret); | ||||
| 	return 0; | ||||
| } | ||||
| 
 | ||||
|  |  | |||
|  | @ -125,7 +125,7 @@ static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, | |||
| 	return call_iter(iocb, &iter); | ||||
| } | ||||
| 
 | ||||
| static void nvmet_file_io_done(struct kiocb *iocb, long ret, long ret2) | ||||
| static void nvmet_file_io_done(struct kiocb *iocb, long ret) | ||||
| { | ||||
| 	struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); | ||||
| 	u16 status = NVME_SC_SUCCESS; | ||||
|  | @ -222,7 +222,7 @@ static bool nvmet_file_execute_io(struct nvmet_req *req, int ki_flags) | |||
| 	} | ||||
| 
 | ||||
| complete: | ||||
| 	nvmet_file_io_done(&req->f.iocb, ret, 0); | ||||
| 	nvmet_file_io_done(&req->f.iocb, ret); | ||||
| 	return true; | ||||
| } | ||||
| 
 | ||||
|  |  | |||
|  | @ -245,7 +245,7 @@ struct target_core_file_cmd { | |||
| 	struct bio_vec	bvecs[]; | ||||
| }; | ||||
| 
 | ||||
| static void cmd_rw_aio_complete(struct kiocb *iocb, long ret, long ret2) | ||||
| static void cmd_rw_aio_complete(struct kiocb *iocb, long ret) | ||||
| { | ||||
| 	struct target_core_file_cmd *cmd; | ||||
| 
 | ||||
|  | @ -303,7 +303,7 @@ fd_execute_rw_aio(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, | |||
| 		ret = call_read_iter(file, &aio_cmd->iocb, &iter); | ||||
| 
 | ||||
| 	if (ret != -EIOCBQUEUED) | ||||
| 		cmd_rw_aio_complete(&aio_cmd->iocb, ret, 0); | ||||
| 		cmd_rw_aio_complete(&aio_cmd->iocb, ret); | ||||
| 
 | ||||
| 	return 0; | ||||
| } | ||||
|  |  | |||
|  | @ -831,7 +831,7 @@ static void ffs_user_copy_worker(struct work_struct *work) | |||
| 		kthread_unuse_mm(io_data->mm); | ||||
| 	} | ||||
| 
 | ||||
| 	io_data->kiocb->ki_complete(io_data->kiocb, ret, 0); | ||||
| 	io_data->kiocb->ki_complete(io_data->kiocb, ret); | ||||
| 
 | ||||
| 	if (io_data->ffs->ffs_eventfd && !kiocb_has_eventfd) | ||||
| 		eventfd_signal(io_data->ffs->ffs_eventfd, 1); | ||||
|  |  | |||
|  | @ -469,7 +469,7 @@ static void ep_user_copy_worker(struct work_struct *work) | |||
| 		ret = -EFAULT; | ||||
| 
 | ||||
| 	/* completing the iocb can drop the ctx and mm, don't touch mm after */ | ||||
| 	iocb->ki_complete(iocb, ret, 0); | ||||
| 	iocb->ki_complete(iocb, ret); | ||||
| 
 | ||||
| 	kfree(priv->buf); | ||||
| 	kfree(priv->to_free); | ||||
|  | @ -497,8 +497,7 @@ static void ep_aio_complete(struct usb_ep *ep, struct usb_request *req) | |||
| 		kfree(priv); | ||||
| 		iocb->private = NULL; | ||||
| 		iocb->ki_complete(iocb, | ||||
| 				req->actual ? req->actual : (long)req->status, | ||||
| 				0); | ||||
| 				req->actual ? req->actual : (long)req->status); | ||||
| 	} else { | ||||
| 		/* ep_copy_to_user() won't report both; we hide some faults */ | ||||
| 		if (unlikely(0 != req->status)) | ||||
|  |  | |||
							
								
								
									
										6
									
								
								fs/aio.c
									
									
									
									
									
								
							
							
						
						
									
										6
									
								
								fs/aio.c
									
									
									
									
									
								
							|  | @ -1417,7 +1417,7 @@ static void aio_remove_iocb(struct aio_kiocb *iocb) | |||
| 	spin_unlock_irqrestore(&ctx->ctx_lock, flags); | ||||
| } | ||||
| 
 | ||||
| static void aio_complete_rw(struct kiocb *kiocb, long res, long res2) | ||||
| static void aio_complete_rw(struct kiocb *kiocb, long res) | ||||
| { | ||||
| 	struct aio_kiocb *iocb = container_of(kiocb, struct aio_kiocb, rw); | ||||
| 
 | ||||
|  | @ -1437,7 +1437,7 @@ static void aio_complete_rw(struct kiocb *kiocb, long res, long res2) | |||
| 	} | ||||
| 
 | ||||
| 	iocb->ki_res.res = res; | ||||
| 	iocb->ki_res.res2 = res2; | ||||
| 	iocb->ki_res.res2 = 0; | ||||
| 	iocb_put(iocb); | ||||
| } | ||||
| 
 | ||||
|  | @ -1508,7 +1508,7 @@ static inline void aio_rw_done(struct kiocb *req, ssize_t ret) | |||
| 		ret = -EINTR; | ||||
| 		fallthrough; | ||||
| 	default: | ||||
| 		req->ki_complete(req, ret, 0); | ||||
| 		req->ki_complete(req, ret); | ||||
| 	} | ||||
| } | ||||
| 
 | ||||
|  |  | |||
|  | @ -37,11 +37,11 @@ static inline void cachefiles_put_kiocb(struct cachefiles_kiocb *ki) | |||
| /*
 | ||||
|  * Handle completion of a read from the cache. | ||||
|  */ | ||||
| static void cachefiles_read_complete(struct kiocb *iocb, long ret, long ret2) | ||||
| static void cachefiles_read_complete(struct kiocb *iocb, long ret) | ||||
| { | ||||
| 	struct cachefiles_kiocb *ki = container_of(iocb, struct cachefiles_kiocb, iocb); | ||||
| 
 | ||||
| 	_enter("%ld,%ld", ret, ret2); | ||||
| 	_enter("%ld", ret); | ||||
| 
 | ||||
| 	if (ki->term_func) { | ||||
| 		if (ret >= 0) | ||||
|  | @ -139,7 +139,7 @@ static int cachefiles_read(struct netfs_cache_resources *cres, | |||
| 		fallthrough; | ||||
| 	default: | ||||
| 		ki->was_async = false; | ||||
| 		cachefiles_read_complete(&ki->iocb, ret, 0); | ||||
| 		cachefiles_read_complete(&ki->iocb, ret); | ||||
| 		if (ret > 0) | ||||
| 			ret = 0; | ||||
| 		break; | ||||
|  | @ -159,12 +159,12 @@ static int cachefiles_read(struct netfs_cache_resources *cres, | |||
| /*
 | ||||
|  * Handle completion of a write to the cache. | ||||
|  */ | ||||
| static void cachefiles_write_complete(struct kiocb *iocb, long ret, long ret2) | ||||
| static void cachefiles_write_complete(struct kiocb *iocb, long ret) | ||||
| { | ||||
| 	struct cachefiles_kiocb *ki = container_of(iocb, struct cachefiles_kiocb, iocb); | ||||
| 	struct inode *inode = file_inode(ki->iocb.ki_filp); | ||||
| 
 | ||||
| 	_enter("%ld,%ld", ret, ret2); | ||||
| 	_enter("%ld", ret); | ||||
| 
 | ||||
| 	/* Tell lockdep we inherited freeze protection from submission thread */ | ||||
| 	__sb_writers_acquired(inode->i_sb, SB_FREEZE_WRITE); | ||||
|  | @ -244,7 +244,7 @@ static int cachefiles_write(struct netfs_cache_resources *cres, | |||
| 		fallthrough; | ||||
| 	default: | ||||
| 		ki->was_async = false; | ||||
| 		cachefiles_write_complete(&ki->iocb, ret, 0); | ||||
| 		cachefiles_write_complete(&ki->iocb, ret); | ||||
| 		if (ret > 0) | ||||
| 			ret = 0; | ||||
| 		break; | ||||
|  |  | |||
|  | @ -1023,7 +1023,7 @@ static void ceph_aio_complete(struct inode *inode, | |||
| 	ceph_put_cap_refs(ci, (aio_req->write ? CEPH_CAP_FILE_WR : | ||||
| 						CEPH_CAP_FILE_RD)); | ||||
| 
 | ||||
| 	aio_req->iocb->ki_complete(aio_req->iocb, ret, 0); | ||||
| 	aio_req->iocb->ki_complete(aio_req->iocb, ret); | ||||
| 
 | ||||
| 	ceph_free_cap_flush(aio_req->prealloc_cf); | ||||
| 	kfree(aio_req); | ||||
|  |  | |||
|  | @ -3184,7 +3184,7 @@ static void collect_uncached_write_data(struct cifs_aio_ctx *ctx) | |||
| 	mutex_unlock(&ctx->aio_mutex); | ||||
| 
 | ||||
| 	if (ctx->iocb && ctx->iocb->ki_complete) | ||||
| 		ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0); | ||||
| 		ctx->iocb->ki_complete(ctx->iocb, ctx->rc); | ||||
| 	else | ||||
| 		complete(&ctx->done); | ||||
| } | ||||
|  | @ -3917,7 +3917,7 @@ collect_uncached_read_data(struct cifs_aio_ctx *ctx) | |||
| 	mutex_unlock(&ctx->aio_mutex); | ||||
| 
 | ||||
| 	if (ctx->iocb && ctx->iocb->ki_complete) | ||||
| 		ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0); | ||||
| 		ctx->iocb->ki_complete(ctx->iocb, ctx->rc); | ||||
| 	else | ||||
| 		complete(&ctx->done); | ||||
| } | ||||
|  |  | |||
|  | @ -307,7 +307,7 @@ static ssize_t dio_complete(struct dio *dio, ssize_t ret, unsigned int flags) | |||
| 
 | ||||
| 		if (ret > 0 && dio->op == REQ_OP_WRITE) | ||||
| 			ret = generic_write_sync(dio->iocb, ret); | ||||
| 		dio->iocb->ki_complete(dio->iocb, ret, 0); | ||||
| 		dio->iocb->ki_complete(dio->iocb, ret); | ||||
| 	} | ||||
| 
 | ||||
| 	kmem_cache_free(dio_cache, dio); | ||||
|  |  | |||
|  | @ -687,7 +687,7 @@ static void fuse_aio_complete(struct fuse_io_priv *io, int err, ssize_t pos) | |||
| 			spin_unlock(&fi->lock); | ||||
| 		} | ||||
| 
 | ||||
| 		io->iocb->ki_complete(io->iocb, res, 0); | ||||
| 		io->iocb->ki_complete(io->iocb, res); | ||||
| 	} | ||||
| 
 | ||||
| 	kref_put(&io->refcnt, fuse_io_release); | ||||
|  |  | |||
|  | @ -2689,7 +2689,7 @@ static void __io_complete_rw(struct io_kiocb *req, long res, long res2, | |||
| 	__io_req_complete(req, issue_flags, req->result, io_put_rw_kbuf(req)); | ||||
| } | ||||
| 
 | ||||
| static void io_complete_rw(struct kiocb *kiocb, long res, long res2) | ||||
| static void io_complete_rw(struct kiocb *kiocb, long res) | ||||
| { | ||||
| 	struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw.kiocb); | ||||
| 
 | ||||
|  | @ -2700,7 +2700,7 @@ static void io_complete_rw(struct kiocb *kiocb, long res, long res2) | |||
| 	io_req_task_work_add(req); | ||||
| } | ||||
| 
 | ||||
| static void io_complete_rw_iopoll(struct kiocb *kiocb, long res, long res2) | ||||
| static void io_complete_rw_iopoll(struct kiocb *kiocb, long res) | ||||
| { | ||||
| 	struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw.kiocb); | ||||
| 
 | ||||
|  | @ -2913,7 +2913,7 @@ static inline void io_rw_done(struct kiocb *kiocb, ssize_t ret) | |||
| 		ret = -EINTR; | ||||
| 		fallthrough; | ||||
| 	default: | ||||
| 		kiocb->ki_complete(kiocb, ret, 0); | ||||
| 		kiocb->ki_complete(kiocb, ret); | ||||
| 	} | ||||
| } | ||||
| 
 | ||||
|  |  | |||
|  | @ -125,7 +125,7 @@ static void iomap_dio_complete_work(struct work_struct *work) | |||
| 	struct iomap_dio *dio = container_of(work, struct iomap_dio, aio.work); | ||||
| 	struct kiocb *iocb = dio->iocb; | ||||
| 
 | ||||
| 	iocb->ki_complete(iocb, iomap_dio_complete(dio), 0); | ||||
| 	iocb->ki_complete(iocb, iomap_dio_complete(dio)); | ||||
| } | ||||
| 
 | ||||
| /*
 | ||||
|  |  | |||
|  | @ -275,7 +275,7 @@ static void nfs_direct_complete(struct nfs_direct_req *dreq) | |||
| 			res = (long) dreq->count; | ||||
| 			WARN_ON_ONCE(dreq->count < 0); | ||||
| 		} | ||||
| 		dreq->iocb->ki_complete(dreq->iocb, res, 0); | ||||
| 		dreq->iocb->ki_complete(dreq->iocb, res); | ||||
| 	} | ||||
| 
 | ||||
| 	complete(&dreq->completion); | ||||
|  |  | |||
|  | @ -272,14 +272,14 @@ static void ovl_aio_cleanup_handler(struct ovl_aio_req *aio_req) | |||
| 	kmem_cache_free(ovl_aio_request_cachep, aio_req); | ||||
| } | ||||
| 
 | ||||
| static void ovl_aio_rw_complete(struct kiocb *iocb, long res, long res2) | ||||
| static void ovl_aio_rw_complete(struct kiocb *iocb, long res) | ||||
| { | ||||
| 	struct ovl_aio_req *aio_req = container_of(iocb, | ||||
| 						   struct ovl_aio_req, iocb); | ||||
| 	struct kiocb *orig_iocb = aio_req->orig_iocb; | ||||
| 
 | ||||
| 	ovl_aio_cleanup_handler(aio_req); | ||||
| 	orig_iocb->ki_complete(orig_iocb, res, res2); | ||||
| 	orig_iocb->ki_complete(orig_iocb, res); | ||||
| } | ||||
| 
 | ||||
| static ssize_t ovl_read_iter(struct kiocb *iocb, struct iov_iter *iter) | ||||
|  |  | |||
|  | @ -330,7 +330,7 @@ struct kiocb { | |||
| 	randomized_struct_fields_start | ||||
| 
 | ||||
| 	loff_t			ki_pos; | ||||
| 	void (*ki_complete)(struct kiocb *iocb, long ret, long ret2); | ||||
| 	void (*ki_complete)(struct kiocb *iocb, long ret); | ||||
| 	void			*private; | ||||
| 	int			ki_flags; | ||||
| 	u16			ki_hint; | ||||
|  |  | |||
		Loading…
	
		Reference in a new issue
	
	 Jens Axboe
						Jens Axboe