Home
last modified time | relevance | path

Searched refs:req (Results 1 – 25 of 114) sorted by relevance

12345

/fs/nfs/
Dpagelist.c37 const struct nfs_page *req; member
42 const struct nfs_page *req) in nfs_page_iter_page_init() argument
44 i->req = req; in nfs_page_iter_page_init()
50 const struct nfs_page *req = i->req; in nfs_page_iter_page_advance() local
53 i->count = (tmp < req->wb_bytes) ? tmp : req->wb_bytes; in nfs_page_iter_page_advance()
58 const struct nfs_page *req = i->req; in nfs_page_iter_page_get() local
61 if (i->count != req->wb_bytes) { in nfs_page_iter_page_get()
62 size_t base = i->count + req->wb_pgbase; in nfs_page_iter_page_get()
65 page = nfs_page_to_page(req, base); in nfs_page_iter_page_get()
102 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init()
[all …]
Dwrite.c56 static void nfs_redirty_request(struct nfs_page *req);
61 static void nfs_inode_remove_request(struct nfs_page *req);
63 struct nfs_page *req);
151 nfs_page_set_inode_ref(struct nfs_page *req, struct inode *inode) in nfs_page_set_inode_ref() argument
153 if (!test_and_set_bit(PG_INODE_REF, &req->wb_flags)) { in nfs_page_set_inode_ref()
154 kref_get(&req->wb_kref); in nfs_page_set_inode_ref()
160 nfs_cancel_remove_inode(struct nfs_page *req, struct inode *inode) in nfs_cancel_remove_inode() argument
164 if (!test_bit(PG_REMOVE, &req->wb_flags)) in nfs_cancel_remove_inode()
166 ret = nfs_page_group_lock(req); in nfs_cancel_remove_inode()
169 if (test_and_clear_bit(PG_REMOVE, &req->wb_flags)) in nfs_cancel_remove_inode()
[all …]
Ddirect.c122 struct nfs_page *req) in nfs_direct_truncate_request() argument
124 loff_t offs = req_offset(req); in nfs_direct_truncate_request()
276 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() local
277 struct page *page = req->wb_page; in nfs_direct_read_completion()
282 bytes += req->wb_bytes; in nfs_direct_read_completion()
283 nfs_list_remove_request(req); in nfs_direct_read_completion()
284 nfs_release_request(req); in nfs_direct_read_completion()
294 struct nfs_page *req; in nfs_read_sync_pgio_error() local
297 req = nfs_list_entry(head->next); in nfs_read_sync_pgio_error()
298 nfs_list_remove_request(req); in nfs_read_sync_pgio_error()
[all …]
/fs/smb/server/
Dtransport_ipc.c289 static int ipc_server_config_on_startup(struct ksmbd_startup_request *req) in ipc_server_config_on_startup() argument
293 ksmbd_set_fd_limit(req->file_max); in ipc_server_config_on_startup()
294 server_conf.flags = req->flags; in ipc_server_config_on_startup()
295 server_conf.signing = req->signing; in ipc_server_config_on_startup()
296 server_conf.tcp_port = req->tcp_port; in ipc_server_config_on_startup()
297 server_conf.ipc_timeout = req->ipc_timeout * HZ; in ipc_server_config_on_startup()
298 server_conf.deadtime = req->deadtime * SMB_ECHO_INTERVAL; in ipc_server_config_on_startup()
299 server_conf.share_fake_fscaps = req->share_fake_fscaps; in ipc_server_config_on_startup()
300 ksmbd_init_domain(req->sub_auth); in ipc_server_config_on_startup()
302 if (req->smb2_max_read) in ipc_server_config_on_startup()
[all …]
/fs/lockd/
Dclntproc.c129 static void nlmclnt_setlockargs(struct nlm_rqst *req, struct file_lock *fl) in nlmclnt_setlockargs() argument
131 struct nlm_args *argp = &req->a_args; in nlmclnt_setlockargs()
133 char *nodename = req->a_host->h_rpcclnt->cl_nodename; in nlmclnt_setlockargs()
138 lock->oh.data = req->a_owner; in nlmclnt_setlockargs()
139 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs()
148 static void nlmclnt_release_lockargs(struct nlm_rqst *req) in nlmclnt_release_lockargs() argument
150 WARN_ON_ONCE(req->a_args.lock.fl.fl_ops != NULL); in nlmclnt_release_lockargs()
264 nlmclnt_call(const struct cred *cred, struct nlm_rqst *req, u32 proc) in nlmclnt_call() argument
266 struct nlm_host *host = req->a_host; in nlmclnt_call()
268 struct nlm_args *argp = &req->a_args; in nlmclnt_call()
[all …]
/fs/cachefiles/
Dondemand.c12 static inline void cachefiles_req_put(struct cachefiles_req *req) in cachefiles_req_put() argument
14 if (refcount_dec_and_test(&req->ref)) in cachefiles_req_put()
15 kfree(req); in cachefiles_req_put()
25 struct cachefiles_req *req; in cachefiles_ondemand_fd_release() local
43 xas_for_each_marked(&xas, req, ULONG_MAX, CACHEFILES_REQ_NEW) { in cachefiles_ondemand_fd_release()
44 if (req->msg.object_id == object_id && in cachefiles_ondemand_fd_release()
45 req->msg.opcode == CACHEFILES_OP_CLOSE) { in cachefiles_ondemand_fd_release()
46 complete(&req->done); in cachefiles_ondemand_fd_release()
104 struct cachefiles_req *req; in cachefiles_ondemand_fd_ioctl() local
114 req = xas_load(&xas); in cachefiles_ondemand_fd_ioctl()
[all …]
/fs/ubifs/
Dbudget.c359 const struct ubifs_budget_req *req) in calc_idx_growth() argument
363 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth()
364 req->new_dent; in calc_idx_growth()
375 const struct ubifs_budget_req *req) in calc_data_growth() argument
379 data_growth = req->new_ino ? c->bi.inode_budget : 0; in calc_data_growth()
380 if (req->new_page) in calc_data_growth()
382 if (req->new_dent) in calc_data_growth()
384 data_growth += req->new_ino_d; in calc_data_growth()
395 const struct ubifs_budget_req *req) in calc_dd_growth() argument
399 dd_growth = req->dirtied_page ? c->bi.page_budget : 0; in calc_dd_growth()
[all …]
/fs/fuse/
Ddev.c46 static void fuse_request_init(struct fuse_mount *fm, struct fuse_req *req) in fuse_request_init() argument
48 INIT_LIST_HEAD(&req->list); in fuse_request_init()
49 INIT_LIST_HEAD(&req->intr_entry); in fuse_request_init()
50 init_waitqueue_head(&req->waitq); in fuse_request_init()
51 refcount_set(&req->count, 1); in fuse_request_init()
52 __set_bit(FR_PENDING, &req->flags); in fuse_request_init()
53 req->fm = fm; in fuse_request_init()
58 struct fuse_req *req = kmem_cache_zalloc(fuse_req_cachep, flags); in fuse_request_alloc() local
59 if (req) in fuse_request_alloc()
60 fuse_request_init(fm, req); in fuse_request_alloc()
[all …]
Dvirtio_fs.c80 struct virtio_fs_forget_req req; member
84 struct fuse_req *req; member
90 struct fuse_req *req, bool in_flight);
350 void *req; in virtio_fs_hiprio_done_work() local
354 while ((req = virtqueue_get_buf(vq, &len)) != NULL) { in virtio_fs_hiprio_done_work()
355 kfree(req); in virtio_fs_hiprio_done_work()
364 struct fuse_req *req; in virtio_fs_request_dispatch_work() local
372 req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req, in virtio_fs_request_dispatch_work()
374 if (!req) { in virtio_fs_request_dispatch_work()
379 list_del_init(&req->list); in virtio_fs_request_dispatch_work()
[all …]
/fs/nilfs2/
Ddat.c41 struct nilfs_palloc_req *req, int create) in nilfs_dat_prepare_entry() argument
45 ret = nilfs_palloc_get_entry_block(dat, req->pr_entry_nr, in nilfs_dat_prepare_entry()
46 create, &req->pr_entry_bh); in nilfs_dat_prepare_entry()
50 (unsigned long long)req->pr_entry_nr); in nilfs_dat_prepare_entry()
61 struct nilfs_palloc_req *req) in nilfs_dat_commit_entry() argument
63 mark_buffer_dirty(req->pr_entry_bh); in nilfs_dat_commit_entry()
65 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry()
69 struct nilfs_palloc_req *req) in nilfs_dat_abort_entry() argument
71 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry()
74 int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req) in nilfs_dat_prepare_alloc() argument
[all …]
Difile.c55 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local
58 req.pr_entry_nr = NILFS_FIRST_INO(ifile->i_sb); in nilfs_ifile_create_inode()
59 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode()
61 ret = nilfs_palloc_prepare_alloc_entry(ifile, &req, false); in nilfs_ifile_create_inode()
63 ret = nilfs_palloc_get_entry_block(ifile, req.pr_entry_nr, 1, in nilfs_ifile_create_inode()
64 &req.pr_entry_bh); in nilfs_ifile_create_inode()
66 nilfs_palloc_abort_alloc_entry(ifile, &req); in nilfs_ifile_create_inode()
69 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode()
72 nilfs_palloc_commit_alloc_entry(ifile, &req); in nilfs_ifile_create_inode()
73 mark_buffer_dirty(req.pr_entry_bh); in nilfs_ifile_create_inode()
[all …]
Dalloc.c504 struct nilfs_palloc_req *req, bool wrap) in nilfs_palloc_prepare_alloc_entry() argument
519 group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); in nilfs_palloc_prepare_alloc_entry()
526 maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, in nilfs_palloc_prepare_alloc_entry()
559 req->pr_entry_nr = in nilfs_palloc_prepare_alloc_entry()
564 req->pr_desc_bh = desc_bh; in nilfs_palloc_prepare_alloc_entry()
565 req->pr_bitmap_bh = bitmap_bh; in nilfs_palloc_prepare_alloc_entry()
594 struct nilfs_palloc_req *req) in nilfs_palloc_commit_alloc_entry() argument
596 mark_buffer_dirty(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry()
597 mark_buffer_dirty(req->pr_desc_bh); in nilfs_palloc_commit_alloc_entry()
600 brelse(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry()
[all …]
/fs/ceph/
Ddir.c374 struct ceph_mds_request *req; in ceph_readdir() local
396 req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS); in ceph_readdir()
397 if (IS_ERR(req)) in ceph_readdir()
398 return PTR_ERR(req); in ceph_readdir()
400 err = ceph_alloc_readdir_reply_buffer(req, inode); in ceph_readdir()
402 ceph_mdsc_put_request(req); in ceph_readdir()
406 req->r_direct_mode = USE_AUTH_MDS; in ceph_readdir()
408 req->r_direct_hash = ceph_frag_value(frag); in ceph_readdir()
409 __set_bit(CEPH_MDS_R_DIRECT_IS_HASH, &req->r_req_flags); in ceph_readdir()
410 req->r_inode_drop = CEPH_CAP_FILE_EXCL; in ceph_readdir()
[all …]
Dmds_client.c410 struct ceph_mds_request *req, in parse_reply_info_readdir() argument
413 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info_readdir()
443 struct inode *inode = d_inode(req->r_dentry); in parse_reply_info_readdir()
722 struct ceph_mds_request *req, in parse_reply_info_extra() argument
725 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info_extra()
731 return parse_reply_info_readdir(p, end, req, features); in parse_reply_info_extra()
744 struct ceph_mds_request *req, u64 features) in parse_reply_info() argument
746 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info()
768 err = parse_reply_info_extra(&p, p+len, req, features, s); in parse_reply_info()
1045 static void put_request_session(struct ceph_mds_request *req) in put_request_session() argument
[all …]
Dexport.c139 struct ceph_mds_request *req; in __lookup_inode() local
142 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_LOOKUPINO, in __lookup_inode()
144 if (IS_ERR(req)) in __lookup_inode()
145 return ERR_CAST(req); in __lookup_inode()
150 req->r_args.lookupino.mask = cpu_to_le32(mask); in __lookup_inode()
152 req->r_ino1 = vino; in __lookup_inode()
153 req->r_num_caps = 1; in __lookup_inode()
154 err = ceph_mdsc_do_request(mdsc, NULL, req); in __lookup_inode()
155 inode = req->r_target_inode; in __lookup_inode()
158 ceph_mdsc_put_request(req); in __lookup_inode()
[all …]
Dfile.c181 struct ceph_mds_request *req; in prepare_open_request() local
188 req = ceph_mdsc_create_request(mdsc, op, want_auth); in prepare_open_request()
189 if (IS_ERR(req)) in prepare_open_request()
191 req->r_fmode = ceph_flags_to_mode(flags); in prepare_open_request()
192 req->r_args.open.flags = ceph_flags_sys2wire(flags); in prepare_open_request()
193 req->r_args.open.mode = cpu_to_le32(create_mode); in prepare_open_request()
195 return req; in prepare_open_request()
300 struct ceph_mds_request *req; in ceph_renew_caps() local
329 req = prepare_open_request(inode->i_sb, flags, 0); in ceph_renew_caps()
330 if (IS_ERR(req)) { in ceph_renew_caps()
[all …]
Dlocks.c15 struct ceph_mds_request *req);
80 struct ceph_mds_request *req; in ceph_lock_message() local
99 req = ceph_mdsc_create_request(mdsc, operation, USE_AUTH_MDS); in ceph_lock_message()
100 if (IS_ERR(req)) in ceph_lock_message()
101 return PTR_ERR(req); in ceph_lock_message()
102 req->r_inode = inode; in ceph_lock_message()
104 req->r_num_caps = 1; in ceph_lock_message()
119 req->r_args.filelock_change.rule = lock_type; in ceph_lock_message()
120 req->r_args.filelock_change.type = cmd; in ceph_lock_message()
121 req->r_args.filelock_change.owner = cpu_to_le64(owner); in ceph_lock_message()
[all …]
Ddebugfs.c56 struct ceph_mds_request *req; in mdsc_show() local
64 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show()
66 if (req->r_request && req->r_session) in mdsc_show()
67 seq_printf(s, "%lld\tmds%d\t", req->r_tid, in mdsc_show()
68 req->r_session->s_mds); in mdsc_show()
69 else if (!req->r_request) in mdsc_show()
70 seq_printf(s, "%lld\t(no request)\t", req->r_tid); in mdsc_show()
72 seq_printf(s, "%lld\t(no session)\t", req->r_tid); in mdsc_show()
74 seq_printf(s, "%s", ceph_mds_op_name(req->r_op)); in mdsc_show()
76 if (test_bit(CEPH_MDS_R_GOT_UNSAFE, &req->r_req_flags)) in mdsc_show()
[all …]
/fs/erofs/
Dfscache.c42 struct erofs_fscache_request *req; in erofs_fscache_req_alloc() local
44 req = kzalloc(sizeof(struct erofs_fscache_request), GFP_KERNEL); in erofs_fscache_req_alloc()
45 if (!req) in erofs_fscache_req_alloc()
48 req->mapping = mapping; in erofs_fscache_req_alloc()
49 req->start = start; in erofs_fscache_req_alloc()
50 req->len = len; in erofs_fscache_req_alloc()
51 refcount_set(&req->ref, 1); in erofs_fscache_req_alloc()
53 return req; in erofs_fscache_req_alloc()
59 struct erofs_fscache_request *req; in erofs_fscache_req_chain() local
67 req = erofs_fscache_req_alloc(primary->mapping, in erofs_fscache_req_chain()
[all …]
/fs/coda/
Dpsdev.c95 struct upc_req *req = NULL; in coda_psdev_write() local
151 req = tmp; in coda_psdev_write()
152 list_del(&req->uc_chain); in coda_psdev_write()
158 if (!req) { in coda_psdev_write()
166 if (req->uc_outSize < nbytes) { in coda_psdev_write()
168 __func__, req->uc_outSize, (long)nbytes, in coda_psdev_write()
170 nbytes = req->uc_outSize; /* don't have more space! */ in coda_psdev_write()
172 if (copy_from_user(req->uc_data, buf, nbytes)) { in coda_psdev_write()
173 req->uc_flags |= CODA_REQ_ABORT; in coda_psdev_write()
174 wake_up(&req->uc_sleep); in coda_psdev_write()
[all …]
/fs/
Daio.c593 struct aio_kiocb *req; in kiocb_set_cancel_fn() local
604 req = container_of(iocb, struct aio_kiocb, rw); in kiocb_set_cancel_fn()
606 if (WARN_ON_ONCE(!list_empty(&req->ki_list))) in kiocb_set_cancel_fn()
609 ctx = req->ki_ctx; in kiocb_set_cancel_fn()
612 list_add_tail(&req->ki_list, &ctx->active_reqs); in kiocb_set_cancel_fn()
613 req->ki_cancel = cancel; in kiocb_set_cancel_fn()
657 struct aio_kiocb *req; in free_ioctx_users() local
662 req = list_first_entry(&ctx->active_reqs, in free_ioctx_users()
664 req->ki_cancel(&req->rw); in free_ioctx_users()
665 list_del_init(&req->ki_list); in free_ioctx_users()
[all …]
/fs/ecryptfs/
Dkthread.c46 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local
58 req = list_first_entry(&ecryptfs_kthread_ctl.req_list, in ecryptfs_threadfn()
61 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn()
62 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn()
64 complete(&req->done); in ecryptfs_threadfn()
91 struct ecryptfs_open_req *req, *tmp; in ecryptfs_destroy_kthread() local
95 list_for_each_entry_safe(req, tmp, &ecryptfs_kthread_ctl.req_list, in ecryptfs_destroy_kthread()
97 list_del(&req->kthread_ctl_list); in ecryptfs_destroy_kthread()
98 *req->lower_file = ERR_PTR(-EIO); in ecryptfs_destroy_kthread()
99 complete(&req->done); in ecryptfs_destroy_kthread()
[all …]
/fs/ocfs2/
Dioctl.c44 struct ocfs2_info_request __user *req) in o2info_set_request_error() argument
47 (void)put_user(kreq->ir_flags, (__u32 __user *)&(req->ir_flags)); in o2info_set_request_error()
50 static inline void o2info_set_request_filled(struct ocfs2_info_request *req) in o2info_set_request_filled() argument
52 req->ir_flags |= OCFS2_INFO_FL_FILLED; in o2info_set_request_filled()
55 static inline void o2info_clear_request_filled(struct ocfs2_info_request *req) in o2info_clear_request_filled() argument
57 req->ir_flags &= ~OCFS2_INFO_FL_FILLED; in o2info_clear_request_filled()
60 static inline int o2info_coherent(struct ocfs2_info_request *req) in o2info_coherent() argument
62 return (!(req->ir_flags & OCFS2_INFO_FL_NON_COHERENT)); in o2info_coherent()
144 struct ocfs2_info_request __user *req) in ocfs2_info_handle_blocksize() argument
148 if (o2info_from_user(oib, req)) in ocfs2_info_handle_blocksize()
[all …]
/fs/smb/client/
Dsmb2pdu.c694 assemble_neg_contexts(struct smb2_negotiate_req *req, in assemble_neg_contexts() argument
714 pneg_ctxt = (*total_len) + (char *)req; in assemble_neg_contexts()
715 req->NegotiateContextOffset = cpu_to_le32(*total_len); in assemble_neg_contexts()
767 req->NegotiateContextCount = cpu_to_le16(neg_context_count); in assemble_neg_contexts()
1049 struct smb2_negotiate_req *req; in SMB2_negotiate() local
1068 (void **) &req, &total_len); in SMB2_negotiate()
1072 req->hdr.SessionId = 0; in SMB2_negotiate()
1079 req->Dialects[0] = cpu_to_le16(SMB30_PROT_ID); in SMB2_negotiate()
1080 req->Dialects[1] = cpu_to_le16(SMB302_PROT_ID); in SMB2_negotiate()
1081 req->Dialects[2] = cpu_to_le16(SMB311_PROT_ID); in SMB2_negotiate()
[all …]
/fs/orangefs/
Dorangefs-sysfs.c351 new_op->upcall.req.param.type = ORANGEFS_PARAM_REQUEST_GET; in sysfs_service_op_show()
365 new_op->upcall.req.param.op = in sysfs_service_op_show()
369 new_op->upcall.req.param.op = in sysfs_service_op_show()
373 new_op->upcall.req.param.op = in sysfs_service_op_show()
378 new_op->upcall.req.param.op = in sysfs_service_op_show()
383 new_op->upcall.req.param.op = in sysfs_service_op_show()
388 new_op->upcall.req.param.op = in sysfs_service_op_show()
393 new_op->upcall.req.param.op = in sysfs_service_op_show()
397 new_op->upcall.req.param.op = in sysfs_service_op_show()
401 new_op->upcall.req.param.op = in sysfs_service_op_show()
[all …]

12345