Home
last modified time | relevance | path

Searched refs:req (Results 1 – 25 of 100) sorted by relevance

1234

/fs/nfs/
Dpagelist.c49 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init()
51 hdr->cred = nfs_req_openctx(hdr->req)->cred; in nfs_pgheader_init()
52 hdr->io_start = req_offset(hdr->req); in nfs_pgheader_init()
145 nfs_page_group_lock(struct nfs_page *req) in nfs_page_group_lock() argument
147 struct nfs_page *head = req->wb_head; in nfs_page_group_lock()
165 nfs_page_group_unlock(struct nfs_page *req) in nfs_page_group_unlock() argument
167 struct nfs_page *head = req->wb_head; in nfs_page_group_unlock()
185 nfs_page_group_sync_on_bit_locked(struct nfs_page *req, unsigned int bit) in nfs_page_group_sync_on_bit_locked() argument
187 struct nfs_page *head = req->wb_head; in nfs_page_group_sync_on_bit_locked()
191 WARN_ON_ONCE(test_and_set_bit(bit, &req->wb_flags)); in nfs_page_group_sync_on_bit_locked()
[all …]
Dwrite.c55 static void nfs_redirty_request(struct nfs_page *req);
60 static void nfs_inode_remove_request(struct nfs_page *req);
61 static void nfs_clear_request_commit(struct nfs_page *req);
171 struct nfs_page *req; in nfs_page_find_private_request() local
176 req = nfs_page_private_request(page); in nfs_page_find_private_request()
177 if (req) { in nfs_page_find_private_request()
178 WARN_ON_ONCE(req->wb_head != req); in nfs_page_find_private_request()
179 kref_get(&req->wb_kref); in nfs_page_find_private_request()
182 return req; in nfs_page_find_private_request()
190 struct nfs_page *req = NULL; in nfs_page_find_swap_request() local
[all …]
Ddirect.c401 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() local
402 struct page *page = req->wb_page; in nfs_direct_read_completion()
407 bytes += req->wb_bytes; in nfs_direct_read_completion()
408 nfs_list_remove_request(req); in nfs_direct_read_completion()
409 nfs_release_request(req); in nfs_direct_read_completion()
419 struct nfs_page *req; in nfs_read_sync_pgio_error() local
422 req = nfs_list_entry(head->next); in nfs_read_sync_pgio_error()
423 nfs_list_remove_request(req); in nfs_read_sync_pgio_error()
424 nfs_release_request(req); in nfs_read_sync_pgio_error()
478 struct nfs_page *req; in nfs_direct_read_schedule_iovec() local
[all …]
/fs/ubifs/
Dbudget.c363 const struct ubifs_budget_req *req) in calc_idx_growth() argument
367 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth()
368 req->new_dent; in calc_idx_growth()
379 const struct ubifs_budget_req *req) in calc_data_growth() argument
383 data_growth = req->new_ino ? c->bi.inode_budget : 0; in calc_data_growth()
384 if (req->new_page) in calc_data_growth()
386 if (req->new_dent) in calc_data_growth()
388 data_growth += req->new_ino_d; in calc_data_growth()
399 const struct ubifs_budget_req *req) in calc_dd_growth() argument
403 dd_growth = req->dirtied_page ? c->bi.page_budget : 0; in calc_dd_growth()
[all …]
/fs/fuse/
Ddev.c43 static void fuse_request_init(struct fuse_req *req) in fuse_request_init() argument
45 INIT_LIST_HEAD(&req->list); in fuse_request_init()
46 INIT_LIST_HEAD(&req->intr_entry); in fuse_request_init()
47 init_waitqueue_head(&req->waitq); in fuse_request_init()
48 refcount_set(&req->count, 1); in fuse_request_init()
49 __set_bit(FR_PENDING, &req->flags); in fuse_request_init()
54 struct fuse_req *req = kmem_cache_zalloc(fuse_req_cachep, flags); in fuse_request_alloc() local
55 if (req) in fuse_request_alloc()
56 fuse_request_init(req); in fuse_request_alloc()
58 return req; in fuse_request_alloc()
[all …]
Dvirtio_fs.c59 struct fuse_req *req, bool in_flight);
248 void *req; in virtio_fs_hiprio_done_work() local
252 while ((req = virtqueue_get_buf(vq, &len)) != NULL) { in virtio_fs_hiprio_done_work()
253 kfree(req); in virtio_fs_hiprio_done_work()
262 struct fuse_req *req; in virtio_fs_request_dispatch_work() local
271 req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req, in virtio_fs_request_dispatch_work()
273 if (!req) { in virtio_fs_request_dispatch_work()
278 list_del_init(&req->list); in virtio_fs_request_dispatch_work()
280 fuse_request_end(fc, req); in virtio_fs_request_dispatch_work()
286 req = list_first_entry_or_null(&fsvq->queued_reqs, in virtio_fs_request_dispatch_work()
[all …]
/fs/nilfs2/
Ddat.c41 struct nilfs_palloc_req *req, int create) in nilfs_dat_prepare_entry() argument
43 return nilfs_palloc_get_entry_block(dat, req->pr_entry_nr, in nilfs_dat_prepare_entry()
44 create, &req->pr_entry_bh); in nilfs_dat_prepare_entry()
48 struct nilfs_palloc_req *req) in nilfs_dat_commit_entry() argument
50 mark_buffer_dirty(req->pr_entry_bh); in nilfs_dat_commit_entry()
52 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry()
56 struct nilfs_palloc_req *req) in nilfs_dat_abort_entry() argument
58 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry()
61 int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req) in nilfs_dat_prepare_alloc() argument
65 ret = nilfs_palloc_prepare_alloc_entry(dat, req); in nilfs_dat_prepare_alloc()
[all …]
Difile.c55 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local
58 req.pr_entry_nr = 0; /* in nilfs_ifile_create_inode()
62 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode()
64 ret = nilfs_palloc_prepare_alloc_entry(ifile, &req); in nilfs_ifile_create_inode()
66 ret = nilfs_palloc_get_entry_block(ifile, req.pr_entry_nr, 1, in nilfs_ifile_create_inode()
67 &req.pr_entry_bh); in nilfs_ifile_create_inode()
69 nilfs_palloc_abort_alloc_entry(ifile, &req); in nilfs_ifile_create_inode()
72 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode()
75 nilfs_palloc_commit_alloc_entry(ifile, &req); in nilfs_ifile_create_inode()
76 mark_buffer_dirty(req.pr_entry_bh); in nilfs_ifile_create_inode()
[all …]
Dalloc.c499 struct nilfs_palloc_req *req) in nilfs_palloc_prepare_alloc_entry() argument
514 group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); in nilfs_palloc_prepare_alloc_entry()
521 maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, in nilfs_palloc_prepare_alloc_entry()
548 req->pr_entry_nr = in nilfs_palloc_prepare_alloc_entry()
553 req->pr_desc_bh = desc_bh; in nilfs_palloc_prepare_alloc_entry()
554 req->pr_bitmap_bh = bitmap_bh; in nilfs_palloc_prepare_alloc_entry()
583 struct nilfs_palloc_req *req) in nilfs_palloc_commit_alloc_entry() argument
585 mark_buffer_dirty(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry()
586 mark_buffer_dirty(req->pr_desc_bh); in nilfs_palloc_commit_alloc_entry()
589 brelse(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry()
[all …]
/fs/lockd/
Dclntproc.c126 static void nlmclnt_setlockargs(struct nlm_rqst *req, struct file_lock *fl) in nlmclnt_setlockargs() argument
128 struct nlm_args *argp = &req->a_args; in nlmclnt_setlockargs()
130 char *nodename = req->a_host->h_rpcclnt->cl_nodename; in nlmclnt_setlockargs()
135 lock->oh.data = req->a_owner; in nlmclnt_setlockargs()
136 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs()
145 static void nlmclnt_release_lockargs(struct nlm_rqst *req) in nlmclnt_release_lockargs() argument
147 WARN_ON_ONCE(req->a_args.lock.fl.fl_ops != NULL); in nlmclnt_release_lockargs()
261 nlmclnt_call(const struct cred *cred, struct nlm_rqst *req, u32 proc) in nlmclnt_call() argument
263 struct nlm_host *host = req->a_host; in nlmclnt_call()
265 struct nlm_args *argp = &req->a_args; in nlmclnt_call()
[all …]
/fs/
Dio_uring.c366 static void __io_free_req(struct io_kiocb *req);
427 struct io_kiocb *req) in __io_sequence_defer() argument
429 return req->sequence != ctx->cached_cq_tail + ctx->cached_sq_dropped in __io_sequence_defer()
434 struct io_kiocb *req) in io_sequence_defer() argument
436 if ((req->flags & (REQ_F_IO_DRAIN|REQ_F_IO_DRAINED)) != REQ_F_IO_DRAIN) in io_sequence_defer()
439 return __io_sequence_defer(ctx, req); in io_sequence_defer()
444 struct io_kiocb *req; in io_get_deferred_req() local
446 req = list_first_entry_or_null(&ctx->defer_list, struct io_kiocb, list); in io_get_deferred_req()
447 if (req && !io_sequence_defer(ctx, req)) { in io_get_deferred_req()
448 list_del_init(&req->list); in io_get_deferred_req()
[all …]
Daio.c563 struct aio_kiocb *req = container_of(iocb, struct aio_kiocb, rw); in kiocb_set_cancel_fn() local
564 struct kioctx *ctx = req->ki_ctx; in kiocb_set_cancel_fn()
567 if (WARN_ON_ONCE(!list_empty(&req->ki_list))) in kiocb_set_cancel_fn()
571 list_add_tail(&req->ki_list, &ctx->active_reqs); in kiocb_set_cancel_fn()
572 req->ki_cancel = cancel; in kiocb_set_cancel_fn()
616 struct aio_kiocb *req; in free_ioctx_users() local
621 req = list_first_entry(&ctx->active_reqs, in free_ioctx_users()
623 req->ki_cancel(&req->rw); in free_ioctx_users()
624 list_del_init(&req->ki_list); in free_ioctx_users()
1028 struct aio_kiocb *req; in aio_get_req() local
[all …]
/fs/ceph/
Dexport.c134 struct ceph_mds_request *req; in __lookup_inode() local
137 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_LOOKUPINO, in __lookup_inode()
139 if (IS_ERR(req)) in __lookup_inode()
140 return ERR_CAST(req); in __lookup_inode()
145 req->r_args.lookupino.mask = cpu_to_le32(mask); in __lookup_inode()
147 req->r_ino1 = vino; in __lookup_inode()
148 req->r_num_caps = 1; in __lookup_inode()
149 err = ceph_mdsc_do_request(mdsc, NULL, req); in __lookup_inode()
150 inode = req->r_target_inode; in __lookup_inode()
153 ceph_mdsc_put_request(req); in __lookup_inode()
[all …]
Ddir.c358 struct ceph_mds_request *req; in ceph_readdir() local
380 req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS); in ceph_readdir()
381 if (IS_ERR(req)) in ceph_readdir()
382 return PTR_ERR(req); in ceph_readdir()
383 err = ceph_alloc_readdir_reply_buffer(req, inode); in ceph_readdir()
385 ceph_mdsc_put_request(req); in ceph_readdir()
389 req->r_direct_mode = USE_AUTH_MDS; in ceph_readdir()
391 req->r_direct_hash = ceph_frag_value(frag); in ceph_readdir()
392 __set_bit(CEPH_MDS_R_DIRECT_IS_HASH, &req->r_req_flags); in ceph_readdir()
393 req->r_inode_drop = CEPH_CAP_FILE_EXCL; in ceph_readdir()
[all …]
Dmds_client.c688 static void put_request_session(struct ceph_mds_request *req) in put_request_session() argument
690 if (req->r_session) { in put_request_session()
691 ceph_put_mds_session(req->r_session); in put_request_session()
692 req->r_session = NULL; in put_request_session()
698 struct ceph_mds_request *req = container_of(kref, in ceph_mdsc_release_request() local
701 destroy_reply_info(&req->r_reply_info); in ceph_mdsc_release_request()
702 if (req->r_request) in ceph_mdsc_release_request()
703 ceph_msg_put(req->r_request); in ceph_mdsc_release_request()
704 if (req->r_reply) in ceph_mdsc_release_request()
705 ceph_msg_put(req->r_reply); in ceph_mdsc_release_request()
[all …]
Ddebugfs.c53 struct ceph_mds_request *req; in mdsc_show() local
61 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show()
63 if (req->r_request && req->r_session) in mdsc_show()
64 seq_printf(s, "%lld\tmds%d\t", req->r_tid, in mdsc_show()
65 req->r_session->s_mds); in mdsc_show()
66 else if (!req->r_request) in mdsc_show()
67 seq_printf(s, "%lld\t(no request)\t", req->r_tid); in mdsc_show()
69 seq_printf(s, "%lld\t(no session)\t", req->r_tid); in mdsc_show()
71 seq_printf(s, "%s", ceph_mds_op_name(req->r_op)); in mdsc_show()
73 if (test_bit(CEPH_MDS_R_GOT_UNSAFE, &req->r_req_flags)) in mdsc_show()
[all …]
Dlocks.c14 struct ceph_mds_request *req);
67 struct ceph_mds_request *req; in ceph_lock_message() local
86 req = ceph_mdsc_create_request(mdsc, operation, USE_AUTH_MDS); in ceph_lock_message()
87 if (IS_ERR(req)) in ceph_lock_message()
88 return PTR_ERR(req); in ceph_lock_message()
89 req->r_inode = inode; in ceph_lock_message()
91 req->r_num_caps = 1; in ceph_lock_message()
106 req->r_args.filelock_change.rule = lock_type; in ceph_lock_message()
107 req->r_args.filelock_change.type = cmd; in ceph_lock_message()
108 req->r_args.filelock_change.owner = cpu_to_le64(owner); in ceph_lock_message()
[all …]
Dfile.c185 struct ceph_mds_request *req; in prepare_open_request() local
192 req = ceph_mdsc_create_request(mdsc, op, want_auth); in prepare_open_request()
193 if (IS_ERR(req)) in prepare_open_request()
195 req->r_fmode = ceph_flags_to_mode(flags); in prepare_open_request()
196 req->r_args.open.flags = ceph_flags_sys2wire(flags); in prepare_open_request()
197 req->r_args.open.mode = cpu_to_le32(create_mode); in prepare_open_request()
199 return req; in prepare_open_request()
292 struct ceph_mds_request *req; in ceph_renew_caps() local
320 req = prepare_open_request(inode->i_sb, flags, 0); in ceph_renew_caps()
321 if (IS_ERR(req)) { in ceph_renew_caps()
[all …]
/fs/coda/
Dpsdev.c95 struct upc_req *req = NULL; in coda_psdev_write() local
155 req = tmp; in coda_psdev_write()
156 list_del(&req->uc_chain); in coda_psdev_write()
162 if (!req) { in coda_psdev_write()
170 if (req->uc_outSize < nbytes) { in coda_psdev_write()
172 __func__, req->uc_outSize, (long)nbytes, in coda_psdev_write()
174 nbytes = req->uc_outSize; /* don't have more space! */ in coda_psdev_write()
176 if (copy_from_user(req->uc_data, buf, nbytes)) { in coda_psdev_write()
177 req->uc_flags |= CODA_REQ_ABORT; in coda_psdev_write()
178 wake_up(&req->uc_sleep); in coda_psdev_write()
[all …]
/fs/afs/
Dfile.c189 void afs_put_read(struct afs_read *req) in afs_put_read() argument
193 if (refcount_dec_and_test(&req->usage)) { in afs_put_read()
194 if (req->pages) { in afs_put_read()
195 for (i = 0; i < req->nr_pages; i++) in afs_put_read()
196 if (req->pages[i]) in afs_put_read()
197 put_page(req->pages[i]); in afs_put_read()
198 if (req->pages != req->array) in afs_put_read()
199 kfree(req->pages); in afs_put_read()
201 kfree(req); in afs_put_read()
276 struct afs_read *req; in afs_page_filler() local
[all …]
/fs/cifs/
Dsmb2pdu.c554 assemble_neg_contexts(struct smb2_negotiate_req *req, in assemble_neg_contexts() argument
557 char *pneg_ctxt = (char *)req; in assemble_neg_contexts()
572 pneg_ctxt = (*total_len) + (char *)req; in assemble_neg_contexts()
573 req->NegotiateContextOffset = cpu_to_le32(*total_len); in assemble_neg_contexts()
593 req->NegotiateContextCount = cpu_to_le16(5); in assemble_neg_contexts()
595 req->NegotiateContextCount = cpu_to_le16(4); in assemble_neg_contexts()
767 struct smb2_create_req *req = iov[0].iov_base; in add_posix_context() local
776 if (!req->CreateContextsOffset) in add_posix_context()
777 req->CreateContextsOffset = cpu_to_le32( in add_posix_context()
780 le32_add_cpu(&req->CreateContextsLength, sizeof(struct create_posix)); in add_posix_context()
[all …]
/fs/ecryptfs/
Dkthread.c46 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local
58 req = list_first_entry(&ecryptfs_kthread_ctl.req_list, in ecryptfs_threadfn()
61 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn()
62 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn()
64 complete(&req->done); in ecryptfs_threadfn()
91 struct ecryptfs_open_req *req, *tmp; in ecryptfs_destroy_kthread() local
95 list_for_each_entry_safe(req, tmp, &ecryptfs_kthread_ctl.req_list, in ecryptfs_destroy_kthread()
97 list_del(&req->kthread_ctl_list); in ecryptfs_destroy_kthread()
98 *req->lower_file = ERR_PTR(-EIO); in ecryptfs_destroy_kthread()
99 complete(&req->done); in ecryptfs_destroy_kthread()
[all …]
/fs/verity/
Dhash_algs.c108 struct ahash_request *req = NULL; in fsverity_prepare_hash_state() local
122 req = ahash_request_alloc(alg->tfm, GFP_KERNEL); in fsverity_prepare_hash_state()
123 if (!req) { in fsverity_prepare_hash_state()
144 ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_SLEEP | in fsverity_prepare_hash_state()
147 ahash_request_set_crypt(req, &sg, NULL, padded_salt_size); in fsverity_prepare_hash_state()
149 err = crypto_wait_req(crypto_ahash_init(req), &wait); in fsverity_prepare_hash_state()
153 err = crypto_wait_req(crypto_ahash_update(req), &wait); in fsverity_prepare_hash_state()
157 err = crypto_ahash_export(req, hashstate); in fsverity_prepare_hash_state()
161 ahash_request_free(req); in fsverity_prepare_hash_state()
186 struct ahash_request *req, struct page *page, u8 *out) in fsverity_hash_page() argument
[all …]
/fs/ocfs2/
Dioctl.c43 struct ocfs2_info_request __user *req) in o2info_set_request_error() argument
46 (void)put_user(kreq->ir_flags, (__u32 __user *)&(req->ir_flags)); in o2info_set_request_error()
49 static inline void o2info_set_request_filled(struct ocfs2_info_request *req) in o2info_set_request_filled() argument
51 req->ir_flags |= OCFS2_INFO_FL_FILLED; in o2info_set_request_filled()
54 static inline void o2info_clear_request_filled(struct ocfs2_info_request *req) in o2info_clear_request_filled() argument
56 req->ir_flags &= ~OCFS2_INFO_FL_FILLED; in o2info_clear_request_filled()
59 static inline int o2info_coherent(struct ocfs2_info_request *req) in o2info_coherent() argument
61 return (!(req->ir_flags & OCFS2_INFO_FL_NON_COHERENT)); in o2info_coherent()
140 struct ocfs2_info_request __user *req) in ocfs2_info_handle_blocksize() argument
144 if (o2info_from_user(oib, req)) in ocfs2_info_handle_blocksize()
[all …]
/fs/orangefs/
Dorangefs-sysfs.c351 new_op->upcall.req.param.type = ORANGEFS_PARAM_REQUEST_GET; in sysfs_service_op_show()
365 new_op->upcall.req.param.op = in sysfs_service_op_show()
369 new_op->upcall.req.param.op = in sysfs_service_op_show()
373 new_op->upcall.req.param.op = in sysfs_service_op_show()
378 new_op->upcall.req.param.op = in sysfs_service_op_show()
383 new_op->upcall.req.param.op = in sysfs_service_op_show()
388 new_op->upcall.req.param.op = in sysfs_service_op_show()
393 new_op->upcall.req.param.op = in sysfs_service_op_show()
397 new_op->upcall.req.param.op = in sysfs_service_op_show()
401 new_op->upcall.req.param.op = in sysfs_service_op_show()
[all …]

1234