/fs/nfs/ |
D | pagelist.c | 37 const struct nfs_page *req; member 42 const struct nfs_page *req) in nfs_page_iter_page_init() argument 44 i->req = req; in nfs_page_iter_page_init() 50 const struct nfs_page *req = i->req; in nfs_page_iter_page_advance() local 53 i->count = (tmp < req->wb_bytes) ? tmp : req->wb_bytes; in nfs_page_iter_page_advance() 58 const struct nfs_page *req = i->req; in nfs_page_iter_page_get() local 61 if (i->count != req->wb_bytes) { in nfs_page_iter_page_get() 62 size_t base = i->count + req->wb_pgbase; in nfs_page_iter_page_get() 65 page = nfs_page_to_page(req, base); in nfs_page_iter_page_get() 102 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init() [all …]
|
D | write.c | 56 static void nfs_redirty_request(struct nfs_page *req); 61 static void nfs_inode_remove_request(struct nfs_page *req); 63 struct nfs_page *req); 151 nfs_page_set_inode_ref(struct nfs_page *req, struct inode *inode) in nfs_page_set_inode_ref() argument 153 if (!test_and_set_bit(PG_INODE_REF, &req->wb_flags)) { in nfs_page_set_inode_ref() 154 kref_get(&req->wb_kref); in nfs_page_set_inode_ref() 160 nfs_cancel_remove_inode(struct nfs_page *req, struct inode *inode) in nfs_cancel_remove_inode() argument 164 if (!test_bit(PG_REMOVE, &req->wb_flags)) in nfs_cancel_remove_inode() 166 ret = nfs_page_group_lock(req); in nfs_cancel_remove_inode() 169 if (test_and_clear_bit(PG_REMOVE, &req->wb_flags)) in nfs_cancel_remove_inode() [all …]
|
D | direct.c | 122 struct nfs_page *req) in nfs_direct_truncate_request() argument 124 loff_t offs = req_offset(req); in nfs_direct_truncate_request() 276 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() local 277 struct page *page = req->wb_page; in nfs_direct_read_completion() 282 bytes += req->wb_bytes; in nfs_direct_read_completion() 283 nfs_list_remove_request(req); in nfs_direct_read_completion() 284 nfs_release_request(req); in nfs_direct_read_completion() 294 struct nfs_page *req; in nfs_read_sync_pgio_error() local 297 req = nfs_list_entry(head->next); in nfs_read_sync_pgio_error() 298 nfs_list_remove_request(req); in nfs_read_sync_pgio_error() [all …]
|
/fs/smb/server/ |
D | transport_ipc.c | 289 static int ipc_server_config_on_startup(struct ksmbd_startup_request *req) in ipc_server_config_on_startup() argument 293 ksmbd_set_fd_limit(req->file_max); in ipc_server_config_on_startup() 294 server_conf.flags = req->flags; in ipc_server_config_on_startup() 295 server_conf.signing = req->signing; in ipc_server_config_on_startup() 296 server_conf.tcp_port = req->tcp_port; in ipc_server_config_on_startup() 297 server_conf.ipc_timeout = req->ipc_timeout * HZ; in ipc_server_config_on_startup() 298 server_conf.deadtime = req->deadtime * SMB_ECHO_INTERVAL; in ipc_server_config_on_startup() 299 server_conf.share_fake_fscaps = req->share_fake_fscaps; in ipc_server_config_on_startup() 300 ksmbd_init_domain(req->sub_auth); in ipc_server_config_on_startup() 302 if (req->smb2_max_read) in ipc_server_config_on_startup() [all …]
|
/fs/lockd/ |
D | clntproc.c | 129 static void nlmclnt_setlockargs(struct nlm_rqst *req, struct file_lock *fl) in nlmclnt_setlockargs() argument 131 struct nlm_args *argp = &req->a_args; in nlmclnt_setlockargs() 133 char *nodename = req->a_host->h_rpcclnt->cl_nodename; in nlmclnt_setlockargs() 138 lock->oh.data = req->a_owner; in nlmclnt_setlockargs() 139 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs() 148 static void nlmclnt_release_lockargs(struct nlm_rqst *req) in nlmclnt_release_lockargs() argument 150 WARN_ON_ONCE(req->a_args.lock.fl.fl_ops != NULL); in nlmclnt_release_lockargs() 264 nlmclnt_call(const struct cred *cred, struct nlm_rqst *req, u32 proc) in nlmclnt_call() argument 266 struct nlm_host *host = req->a_host; in nlmclnt_call() 268 struct nlm_args *argp = &req->a_args; in nlmclnt_call() [all …]
|
/fs/cachefiles/ |
D | ondemand.c | 12 static inline void cachefiles_req_put(struct cachefiles_req *req) in cachefiles_req_put() argument 14 if (refcount_dec_and_test(&req->ref)) in cachefiles_req_put() 15 kfree(req); in cachefiles_req_put() 25 struct cachefiles_req *req; in cachefiles_ondemand_fd_release() local 43 xas_for_each_marked(&xas, req, ULONG_MAX, CACHEFILES_REQ_NEW) { in cachefiles_ondemand_fd_release() 44 if (req->msg.object_id == object_id && in cachefiles_ondemand_fd_release() 45 req->msg.opcode == CACHEFILES_OP_CLOSE) { in cachefiles_ondemand_fd_release() 46 complete(&req->done); in cachefiles_ondemand_fd_release() 104 struct cachefiles_req *req; in cachefiles_ondemand_fd_ioctl() local 114 req = xas_load(&xas); in cachefiles_ondemand_fd_ioctl() [all …]
|
/fs/ubifs/ |
D | budget.c | 359 const struct ubifs_budget_req *req) in calc_idx_growth() argument 363 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 364 req->new_dent; in calc_idx_growth() 375 const struct ubifs_budget_req *req) in calc_data_growth() argument 379 data_growth = req->new_ino ? c->bi.inode_budget : 0; in calc_data_growth() 380 if (req->new_page) in calc_data_growth() 382 if (req->new_dent) in calc_data_growth() 384 data_growth += req->new_ino_d; in calc_data_growth() 395 const struct ubifs_budget_req *req) in calc_dd_growth() argument 399 dd_growth = req->dirtied_page ? c->bi.page_budget : 0; in calc_dd_growth() [all …]
|
/fs/fuse/ |
D | dev.c | 46 static void fuse_request_init(struct fuse_mount *fm, struct fuse_req *req) in fuse_request_init() argument 48 INIT_LIST_HEAD(&req->list); in fuse_request_init() 49 INIT_LIST_HEAD(&req->intr_entry); in fuse_request_init() 50 init_waitqueue_head(&req->waitq); in fuse_request_init() 51 refcount_set(&req->count, 1); in fuse_request_init() 52 __set_bit(FR_PENDING, &req->flags); in fuse_request_init() 53 req->fm = fm; in fuse_request_init() 58 struct fuse_req *req = kmem_cache_zalloc(fuse_req_cachep, flags); in fuse_request_alloc() local 59 if (req) in fuse_request_alloc() 60 fuse_request_init(fm, req); in fuse_request_alloc() [all …]
|
D | virtio_fs.c | 80 struct virtio_fs_forget_req req; member 84 struct fuse_req *req; member 90 struct fuse_req *req, bool in_flight); 350 void *req; in virtio_fs_hiprio_done_work() local 354 while ((req = virtqueue_get_buf(vq, &len)) != NULL) { in virtio_fs_hiprio_done_work() 355 kfree(req); in virtio_fs_hiprio_done_work() 364 struct fuse_req *req; in virtio_fs_request_dispatch_work() local 372 req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req, in virtio_fs_request_dispatch_work() 374 if (!req) { in virtio_fs_request_dispatch_work() 379 list_del_init(&req->list); in virtio_fs_request_dispatch_work() [all …]
|
/fs/nilfs2/ |
D | dat.c | 41 struct nilfs_palloc_req *req, int create) in nilfs_dat_prepare_entry() argument 45 ret = nilfs_palloc_get_entry_block(dat, req->pr_entry_nr, in nilfs_dat_prepare_entry() 46 create, &req->pr_entry_bh); in nilfs_dat_prepare_entry() 50 (unsigned long long)req->pr_entry_nr); in nilfs_dat_prepare_entry() 61 struct nilfs_palloc_req *req) in nilfs_dat_commit_entry() argument 63 mark_buffer_dirty(req->pr_entry_bh); in nilfs_dat_commit_entry() 65 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry() 69 struct nilfs_palloc_req *req) in nilfs_dat_abort_entry() argument 71 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry() 74 int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req) in nilfs_dat_prepare_alloc() argument [all …]
|
D | ifile.c | 55 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local 58 req.pr_entry_nr = NILFS_FIRST_INO(ifile->i_sb); in nilfs_ifile_create_inode() 59 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode() 61 ret = nilfs_palloc_prepare_alloc_entry(ifile, &req, false); in nilfs_ifile_create_inode() 63 ret = nilfs_palloc_get_entry_block(ifile, req.pr_entry_nr, 1, in nilfs_ifile_create_inode() 64 &req.pr_entry_bh); in nilfs_ifile_create_inode() 66 nilfs_palloc_abort_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 69 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode() 72 nilfs_palloc_commit_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 73 mark_buffer_dirty(req.pr_entry_bh); in nilfs_ifile_create_inode() [all …]
|
D | alloc.c | 504 struct nilfs_palloc_req *req, bool wrap) in nilfs_palloc_prepare_alloc_entry() argument 519 group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); in nilfs_palloc_prepare_alloc_entry() 526 maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, in nilfs_palloc_prepare_alloc_entry() 559 req->pr_entry_nr = in nilfs_palloc_prepare_alloc_entry() 564 req->pr_desc_bh = desc_bh; in nilfs_palloc_prepare_alloc_entry() 565 req->pr_bitmap_bh = bitmap_bh; in nilfs_palloc_prepare_alloc_entry() 594 struct nilfs_palloc_req *req) in nilfs_palloc_commit_alloc_entry() argument 596 mark_buffer_dirty(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() 597 mark_buffer_dirty(req->pr_desc_bh); in nilfs_palloc_commit_alloc_entry() 600 brelse(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() [all …]
|
/fs/ceph/ |
D | dir.c | 374 struct ceph_mds_request *req; in ceph_readdir() local 396 req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS); in ceph_readdir() 397 if (IS_ERR(req)) in ceph_readdir() 398 return PTR_ERR(req); in ceph_readdir() 400 err = ceph_alloc_readdir_reply_buffer(req, inode); in ceph_readdir() 402 ceph_mdsc_put_request(req); in ceph_readdir() 406 req->r_direct_mode = USE_AUTH_MDS; in ceph_readdir() 408 req->r_direct_hash = ceph_frag_value(frag); in ceph_readdir() 409 __set_bit(CEPH_MDS_R_DIRECT_IS_HASH, &req->r_req_flags); in ceph_readdir() 410 req->r_inode_drop = CEPH_CAP_FILE_EXCL; in ceph_readdir() [all …]
|
D | mds_client.c | 410 struct ceph_mds_request *req, in parse_reply_info_readdir() argument 413 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info_readdir() 443 struct inode *inode = d_inode(req->r_dentry); in parse_reply_info_readdir() 722 struct ceph_mds_request *req, in parse_reply_info_extra() argument 725 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info_extra() 731 return parse_reply_info_readdir(p, end, req, features); in parse_reply_info_extra() 744 struct ceph_mds_request *req, u64 features) in parse_reply_info() argument 746 struct ceph_mds_reply_info_parsed *info = &req->r_reply_info; in parse_reply_info() 768 err = parse_reply_info_extra(&p, p+len, req, features, s); in parse_reply_info() 1045 static void put_request_session(struct ceph_mds_request *req) in put_request_session() argument [all …]
|
D | export.c | 139 struct ceph_mds_request *req; in __lookup_inode() local 142 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_LOOKUPINO, in __lookup_inode() 144 if (IS_ERR(req)) in __lookup_inode() 145 return ERR_CAST(req); in __lookup_inode() 150 req->r_args.lookupino.mask = cpu_to_le32(mask); in __lookup_inode() 152 req->r_ino1 = vino; in __lookup_inode() 153 req->r_num_caps = 1; in __lookup_inode() 154 err = ceph_mdsc_do_request(mdsc, NULL, req); in __lookup_inode() 155 inode = req->r_target_inode; in __lookup_inode() 158 ceph_mdsc_put_request(req); in __lookup_inode() [all …]
|
D | file.c | 181 struct ceph_mds_request *req; in prepare_open_request() local 188 req = ceph_mdsc_create_request(mdsc, op, want_auth); in prepare_open_request() 189 if (IS_ERR(req)) in prepare_open_request() 191 req->r_fmode = ceph_flags_to_mode(flags); in prepare_open_request() 192 req->r_args.open.flags = ceph_flags_sys2wire(flags); in prepare_open_request() 193 req->r_args.open.mode = cpu_to_le32(create_mode); in prepare_open_request() 195 return req; in prepare_open_request() 300 struct ceph_mds_request *req; in ceph_renew_caps() local 329 req = prepare_open_request(inode->i_sb, flags, 0); in ceph_renew_caps() 330 if (IS_ERR(req)) { in ceph_renew_caps() [all …]
|
D | locks.c | 15 struct ceph_mds_request *req); 80 struct ceph_mds_request *req; in ceph_lock_message() local 99 req = ceph_mdsc_create_request(mdsc, operation, USE_AUTH_MDS); in ceph_lock_message() 100 if (IS_ERR(req)) in ceph_lock_message() 101 return PTR_ERR(req); in ceph_lock_message() 102 req->r_inode = inode; in ceph_lock_message() 104 req->r_num_caps = 1; in ceph_lock_message() 119 req->r_args.filelock_change.rule = lock_type; in ceph_lock_message() 120 req->r_args.filelock_change.type = cmd; in ceph_lock_message() 121 req->r_args.filelock_change.owner = cpu_to_le64(owner); in ceph_lock_message() [all …]
|
D | debugfs.c | 56 struct ceph_mds_request *req; in mdsc_show() local 64 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show() 66 if (req->r_request && req->r_session) in mdsc_show() 67 seq_printf(s, "%lld\tmds%d\t", req->r_tid, in mdsc_show() 68 req->r_session->s_mds); in mdsc_show() 69 else if (!req->r_request) in mdsc_show() 70 seq_printf(s, "%lld\t(no request)\t", req->r_tid); in mdsc_show() 72 seq_printf(s, "%lld\t(no session)\t", req->r_tid); in mdsc_show() 74 seq_printf(s, "%s", ceph_mds_op_name(req->r_op)); in mdsc_show() 76 if (test_bit(CEPH_MDS_R_GOT_UNSAFE, &req->r_req_flags)) in mdsc_show() [all …]
|
/fs/erofs/ |
D | fscache.c | 42 struct erofs_fscache_request *req; in erofs_fscache_req_alloc() local 44 req = kzalloc(sizeof(struct erofs_fscache_request), GFP_KERNEL); in erofs_fscache_req_alloc() 45 if (!req) in erofs_fscache_req_alloc() 48 req->mapping = mapping; in erofs_fscache_req_alloc() 49 req->start = start; in erofs_fscache_req_alloc() 50 req->len = len; in erofs_fscache_req_alloc() 51 refcount_set(&req->ref, 1); in erofs_fscache_req_alloc() 53 return req; in erofs_fscache_req_alloc() 59 struct erofs_fscache_request *req; in erofs_fscache_req_chain() local 67 req = erofs_fscache_req_alloc(primary->mapping, in erofs_fscache_req_chain() [all …]
|
/fs/coda/ |
D | psdev.c | 95 struct upc_req *req = NULL; in coda_psdev_write() local 151 req = tmp; in coda_psdev_write() 152 list_del(&req->uc_chain); in coda_psdev_write() 158 if (!req) { in coda_psdev_write() 166 if (req->uc_outSize < nbytes) { in coda_psdev_write() 168 __func__, req->uc_outSize, (long)nbytes, in coda_psdev_write() 170 nbytes = req->uc_outSize; /* don't have more space! */ in coda_psdev_write() 172 if (copy_from_user(req->uc_data, buf, nbytes)) { in coda_psdev_write() 173 req->uc_flags |= CODA_REQ_ABORT; in coda_psdev_write() 174 wake_up(&req->uc_sleep); in coda_psdev_write() [all …]
|
/fs/ |
D | aio.c | 593 struct aio_kiocb *req; in kiocb_set_cancel_fn() local 604 req = container_of(iocb, struct aio_kiocb, rw); in kiocb_set_cancel_fn() 606 if (WARN_ON_ONCE(!list_empty(&req->ki_list))) in kiocb_set_cancel_fn() 609 ctx = req->ki_ctx; in kiocb_set_cancel_fn() 612 list_add_tail(&req->ki_list, &ctx->active_reqs); in kiocb_set_cancel_fn() 613 req->ki_cancel = cancel; in kiocb_set_cancel_fn() 657 struct aio_kiocb *req; in free_ioctx_users() local 662 req = list_first_entry(&ctx->active_reqs, in free_ioctx_users() 664 req->ki_cancel(&req->rw); in free_ioctx_users() 665 list_del_init(&req->ki_list); in free_ioctx_users() [all …]
|
/fs/ecryptfs/ |
D | kthread.c | 46 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local 58 req = list_first_entry(&ecryptfs_kthread_ctl.req_list, in ecryptfs_threadfn() 61 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn() 62 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn() 64 complete(&req->done); in ecryptfs_threadfn() 91 struct ecryptfs_open_req *req, *tmp; in ecryptfs_destroy_kthread() local 95 list_for_each_entry_safe(req, tmp, &ecryptfs_kthread_ctl.req_list, in ecryptfs_destroy_kthread() 97 list_del(&req->kthread_ctl_list); in ecryptfs_destroy_kthread() 98 *req->lower_file = ERR_PTR(-EIO); in ecryptfs_destroy_kthread() 99 complete(&req->done); in ecryptfs_destroy_kthread() [all …]
|
/fs/ocfs2/ |
D | ioctl.c | 44 struct ocfs2_info_request __user *req) in o2info_set_request_error() argument 47 (void)put_user(kreq->ir_flags, (__u32 __user *)&(req->ir_flags)); in o2info_set_request_error() 50 static inline void o2info_set_request_filled(struct ocfs2_info_request *req) in o2info_set_request_filled() argument 52 req->ir_flags |= OCFS2_INFO_FL_FILLED; in o2info_set_request_filled() 55 static inline void o2info_clear_request_filled(struct ocfs2_info_request *req) in o2info_clear_request_filled() argument 57 req->ir_flags &= ~OCFS2_INFO_FL_FILLED; in o2info_clear_request_filled() 60 static inline int o2info_coherent(struct ocfs2_info_request *req) in o2info_coherent() argument 62 return (!(req->ir_flags & OCFS2_INFO_FL_NON_COHERENT)); in o2info_coherent() 144 struct ocfs2_info_request __user *req) in ocfs2_info_handle_blocksize() argument 148 if (o2info_from_user(oib, req)) in ocfs2_info_handle_blocksize() [all …]
|
/fs/smb/client/ |
D | smb2pdu.c | 694 assemble_neg_contexts(struct smb2_negotiate_req *req, in assemble_neg_contexts() argument 714 pneg_ctxt = (*total_len) + (char *)req; in assemble_neg_contexts() 715 req->NegotiateContextOffset = cpu_to_le32(*total_len); in assemble_neg_contexts() 767 req->NegotiateContextCount = cpu_to_le16(neg_context_count); in assemble_neg_contexts() 1049 struct smb2_negotiate_req *req; in SMB2_negotiate() local 1068 (void **) &req, &total_len); in SMB2_negotiate() 1072 req->hdr.SessionId = 0; in SMB2_negotiate() 1079 req->Dialects[0] = cpu_to_le16(SMB30_PROT_ID); in SMB2_negotiate() 1080 req->Dialects[1] = cpu_to_le16(SMB302_PROT_ID); in SMB2_negotiate() 1081 req->Dialects[2] = cpu_to_le16(SMB311_PROT_ID); in SMB2_negotiate() [all …]
|
/fs/orangefs/ |
D | orangefs-sysfs.c | 351 new_op->upcall.req.param.type = ORANGEFS_PARAM_REQUEST_GET; in sysfs_service_op_show() 365 new_op->upcall.req.param.op = in sysfs_service_op_show() 369 new_op->upcall.req.param.op = in sysfs_service_op_show() 373 new_op->upcall.req.param.op = in sysfs_service_op_show() 378 new_op->upcall.req.param.op = in sysfs_service_op_show() 383 new_op->upcall.req.param.op = in sysfs_service_op_show() 388 new_op->upcall.req.param.op = in sysfs_service_op_show() 393 new_op->upcall.req.param.op = in sysfs_service_op_show() 397 new_op->upcall.req.param.op = in sysfs_service_op_show() 401 new_op->upcall.req.param.op = in sysfs_service_op_show() [all …]
|