/fs/nfs/ |
D | pagelist.c | 49 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init() 51 hdr->cred = nfs_req_openctx(hdr->req)->cred; in nfs_pgheader_init() 52 hdr->io_start = req_offset(hdr->req); in nfs_pgheader_init() 145 nfs_page_group_lock(struct nfs_page *req) in nfs_page_group_lock() argument 147 struct nfs_page *head = req->wb_head; in nfs_page_group_lock() 165 nfs_page_group_unlock(struct nfs_page *req) in nfs_page_group_unlock() argument 167 struct nfs_page *head = req->wb_head; in nfs_page_group_unlock() 185 nfs_page_group_sync_on_bit_locked(struct nfs_page *req, unsigned int bit) in nfs_page_group_sync_on_bit_locked() argument 187 struct nfs_page *head = req->wb_head; in nfs_page_group_sync_on_bit_locked() 191 WARN_ON_ONCE(test_and_set_bit(bit, &req->wb_flags)); in nfs_page_group_sync_on_bit_locked() [all …]
|
D | write.c | 55 static void nfs_redirty_request(struct nfs_page *req); 60 static void nfs_inode_remove_request(struct nfs_page *req); 61 static void nfs_clear_request_commit(struct nfs_page *req); 171 struct nfs_page *req; in nfs_page_find_private_request() local 176 req = nfs_page_private_request(page); in nfs_page_find_private_request() 177 if (req) { in nfs_page_find_private_request() 178 WARN_ON_ONCE(req->wb_head != req); in nfs_page_find_private_request() 179 kref_get(&req->wb_kref); in nfs_page_find_private_request() 182 return req; in nfs_page_find_private_request() 190 struct nfs_page *req = NULL; in nfs_page_find_swap_request() local [all …]
|
D | direct.c | 401 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() local 402 struct page *page = req->wb_page; in nfs_direct_read_completion() 407 bytes += req->wb_bytes; in nfs_direct_read_completion() 408 nfs_list_remove_request(req); in nfs_direct_read_completion() 409 nfs_release_request(req); in nfs_direct_read_completion() 419 struct nfs_page *req; in nfs_read_sync_pgio_error() local 422 req = nfs_list_entry(head->next); in nfs_read_sync_pgio_error() 423 nfs_list_remove_request(req); in nfs_read_sync_pgio_error() 424 nfs_release_request(req); in nfs_read_sync_pgio_error() 478 struct nfs_page *req; in nfs_direct_read_schedule_iovec() local [all …]
|
/fs/ubifs/ |
D | budget.c | 363 const struct ubifs_budget_req *req) in calc_idx_growth() argument 367 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 368 req->new_dent; in calc_idx_growth() 379 const struct ubifs_budget_req *req) in calc_data_growth() argument 383 data_growth = req->new_ino ? c->bi.inode_budget : 0; in calc_data_growth() 384 if (req->new_page) in calc_data_growth() 386 if (req->new_dent) in calc_data_growth() 388 data_growth += req->new_ino_d; in calc_data_growth() 399 const struct ubifs_budget_req *req) in calc_dd_growth() argument 403 dd_growth = req->dirtied_page ? c->bi.page_budget : 0; in calc_dd_growth() [all …]
|
/fs/fuse/ |
D | dev.c | 43 static void fuse_request_init(struct fuse_req *req) in fuse_request_init() argument 45 INIT_LIST_HEAD(&req->list); in fuse_request_init() 46 INIT_LIST_HEAD(&req->intr_entry); in fuse_request_init() 47 init_waitqueue_head(&req->waitq); in fuse_request_init() 48 refcount_set(&req->count, 1); in fuse_request_init() 49 __set_bit(FR_PENDING, &req->flags); in fuse_request_init() 54 struct fuse_req *req = kmem_cache_zalloc(fuse_req_cachep, flags); in fuse_request_alloc() local 55 if (req) in fuse_request_alloc() 56 fuse_request_init(req); in fuse_request_alloc() 58 return req; in fuse_request_alloc() [all …]
|
D | virtio_fs.c | 59 struct fuse_req *req, bool in_flight); 248 void *req; in virtio_fs_hiprio_done_work() local 252 while ((req = virtqueue_get_buf(vq, &len)) != NULL) { in virtio_fs_hiprio_done_work() 253 kfree(req); in virtio_fs_hiprio_done_work() 262 struct fuse_req *req; in virtio_fs_request_dispatch_work() local 271 req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req, in virtio_fs_request_dispatch_work() 273 if (!req) { in virtio_fs_request_dispatch_work() 278 list_del_init(&req->list); in virtio_fs_request_dispatch_work() 280 fuse_request_end(fc, req); in virtio_fs_request_dispatch_work() 286 req = list_first_entry_or_null(&fsvq->queued_reqs, in virtio_fs_request_dispatch_work() [all …]
|
/fs/nilfs2/ |
D | dat.c | 41 struct nilfs_palloc_req *req, int create) in nilfs_dat_prepare_entry() argument 43 return nilfs_palloc_get_entry_block(dat, req->pr_entry_nr, in nilfs_dat_prepare_entry() 44 create, &req->pr_entry_bh); in nilfs_dat_prepare_entry() 48 struct nilfs_palloc_req *req) in nilfs_dat_commit_entry() argument 50 mark_buffer_dirty(req->pr_entry_bh); in nilfs_dat_commit_entry() 52 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry() 56 struct nilfs_palloc_req *req) in nilfs_dat_abort_entry() argument 58 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry() 61 int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req) in nilfs_dat_prepare_alloc() argument 65 ret = nilfs_palloc_prepare_alloc_entry(dat, req); in nilfs_dat_prepare_alloc() [all …]
|
D | ifile.c | 55 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local 58 req.pr_entry_nr = 0; /* in nilfs_ifile_create_inode() 62 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode() 64 ret = nilfs_palloc_prepare_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 66 ret = nilfs_palloc_get_entry_block(ifile, req.pr_entry_nr, 1, in nilfs_ifile_create_inode() 67 &req.pr_entry_bh); in nilfs_ifile_create_inode() 69 nilfs_palloc_abort_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 72 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode() 75 nilfs_palloc_commit_alloc_entry(ifile, &req); in nilfs_ifile_create_inode() 76 mark_buffer_dirty(req.pr_entry_bh); in nilfs_ifile_create_inode() [all …]
|
D | alloc.c | 499 struct nilfs_palloc_req *req) in nilfs_palloc_prepare_alloc_entry() argument 514 group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); in nilfs_palloc_prepare_alloc_entry() 521 maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, in nilfs_palloc_prepare_alloc_entry() 548 req->pr_entry_nr = in nilfs_palloc_prepare_alloc_entry() 553 req->pr_desc_bh = desc_bh; in nilfs_palloc_prepare_alloc_entry() 554 req->pr_bitmap_bh = bitmap_bh; in nilfs_palloc_prepare_alloc_entry() 583 struct nilfs_palloc_req *req) in nilfs_palloc_commit_alloc_entry() argument 585 mark_buffer_dirty(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() 586 mark_buffer_dirty(req->pr_desc_bh); in nilfs_palloc_commit_alloc_entry() 589 brelse(req->pr_bitmap_bh); in nilfs_palloc_commit_alloc_entry() [all …]
|
/fs/lockd/ |
D | clntproc.c | 126 static void nlmclnt_setlockargs(struct nlm_rqst *req, struct file_lock *fl) in nlmclnt_setlockargs() argument 128 struct nlm_args *argp = &req->a_args; in nlmclnt_setlockargs() 130 char *nodename = req->a_host->h_rpcclnt->cl_nodename; in nlmclnt_setlockargs() 135 lock->oh.data = req->a_owner; in nlmclnt_setlockargs() 136 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs() 145 static void nlmclnt_release_lockargs(struct nlm_rqst *req) in nlmclnt_release_lockargs() argument 147 WARN_ON_ONCE(req->a_args.lock.fl.fl_ops != NULL); in nlmclnt_release_lockargs() 261 nlmclnt_call(const struct cred *cred, struct nlm_rqst *req, u32 proc) in nlmclnt_call() argument 263 struct nlm_host *host = req->a_host; in nlmclnt_call() 265 struct nlm_args *argp = &req->a_args; in nlmclnt_call() [all …]
|
/fs/ |
D | io_uring.c | 366 static void __io_free_req(struct io_kiocb *req); 427 struct io_kiocb *req) in __io_sequence_defer() argument 429 return req->sequence != ctx->cached_cq_tail + ctx->cached_sq_dropped in __io_sequence_defer() 434 struct io_kiocb *req) in io_sequence_defer() argument 436 if ((req->flags & (REQ_F_IO_DRAIN|REQ_F_IO_DRAINED)) != REQ_F_IO_DRAIN) in io_sequence_defer() 439 return __io_sequence_defer(ctx, req); in io_sequence_defer() 444 struct io_kiocb *req; in io_get_deferred_req() local 446 req = list_first_entry_or_null(&ctx->defer_list, struct io_kiocb, list); in io_get_deferred_req() 447 if (req && !io_sequence_defer(ctx, req)) { in io_get_deferred_req() 448 list_del_init(&req->list); in io_get_deferred_req() [all …]
|
D | aio.c | 563 struct aio_kiocb *req = container_of(iocb, struct aio_kiocb, rw); in kiocb_set_cancel_fn() local 564 struct kioctx *ctx = req->ki_ctx; in kiocb_set_cancel_fn() 567 if (WARN_ON_ONCE(!list_empty(&req->ki_list))) in kiocb_set_cancel_fn() 571 list_add_tail(&req->ki_list, &ctx->active_reqs); in kiocb_set_cancel_fn() 572 req->ki_cancel = cancel; in kiocb_set_cancel_fn() 616 struct aio_kiocb *req; in free_ioctx_users() local 621 req = list_first_entry(&ctx->active_reqs, in free_ioctx_users() 623 req->ki_cancel(&req->rw); in free_ioctx_users() 624 list_del_init(&req->ki_list); in free_ioctx_users() 1028 struct aio_kiocb *req; in aio_get_req() local [all …]
|
/fs/ceph/ |
D | export.c | 134 struct ceph_mds_request *req; in __lookup_inode() local 137 req = ceph_mdsc_create_request(mdsc, CEPH_MDS_OP_LOOKUPINO, in __lookup_inode() 139 if (IS_ERR(req)) in __lookup_inode() 140 return ERR_CAST(req); in __lookup_inode() 145 req->r_args.lookupino.mask = cpu_to_le32(mask); in __lookup_inode() 147 req->r_ino1 = vino; in __lookup_inode() 148 req->r_num_caps = 1; in __lookup_inode() 149 err = ceph_mdsc_do_request(mdsc, NULL, req); in __lookup_inode() 150 inode = req->r_target_inode; in __lookup_inode() 153 ceph_mdsc_put_request(req); in __lookup_inode() [all …]
|
D | dir.c | 358 struct ceph_mds_request *req; in ceph_readdir() local 380 req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS); in ceph_readdir() 381 if (IS_ERR(req)) in ceph_readdir() 382 return PTR_ERR(req); in ceph_readdir() 383 err = ceph_alloc_readdir_reply_buffer(req, inode); in ceph_readdir() 385 ceph_mdsc_put_request(req); in ceph_readdir() 389 req->r_direct_mode = USE_AUTH_MDS; in ceph_readdir() 391 req->r_direct_hash = ceph_frag_value(frag); in ceph_readdir() 392 __set_bit(CEPH_MDS_R_DIRECT_IS_HASH, &req->r_req_flags); in ceph_readdir() 393 req->r_inode_drop = CEPH_CAP_FILE_EXCL; in ceph_readdir() [all …]
|
D | mds_client.c | 688 static void put_request_session(struct ceph_mds_request *req) in put_request_session() argument 690 if (req->r_session) { in put_request_session() 691 ceph_put_mds_session(req->r_session); in put_request_session() 692 req->r_session = NULL; in put_request_session() 698 struct ceph_mds_request *req = container_of(kref, in ceph_mdsc_release_request() local 701 destroy_reply_info(&req->r_reply_info); in ceph_mdsc_release_request() 702 if (req->r_request) in ceph_mdsc_release_request() 703 ceph_msg_put(req->r_request); in ceph_mdsc_release_request() 704 if (req->r_reply) in ceph_mdsc_release_request() 705 ceph_msg_put(req->r_reply); in ceph_mdsc_release_request() [all …]
|
D | debugfs.c | 53 struct ceph_mds_request *req; in mdsc_show() local 61 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show() 63 if (req->r_request && req->r_session) in mdsc_show() 64 seq_printf(s, "%lld\tmds%d\t", req->r_tid, in mdsc_show() 65 req->r_session->s_mds); in mdsc_show() 66 else if (!req->r_request) in mdsc_show() 67 seq_printf(s, "%lld\t(no request)\t", req->r_tid); in mdsc_show() 69 seq_printf(s, "%lld\t(no session)\t", req->r_tid); in mdsc_show() 71 seq_printf(s, "%s", ceph_mds_op_name(req->r_op)); in mdsc_show() 73 if (test_bit(CEPH_MDS_R_GOT_UNSAFE, &req->r_req_flags)) in mdsc_show() [all …]
|
D | locks.c | 14 struct ceph_mds_request *req); 67 struct ceph_mds_request *req; in ceph_lock_message() local 86 req = ceph_mdsc_create_request(mdsc, operation, USE_AUTH_MDS); in ceph_lock_message() 87 if (IS_ERR(req)) in ceph_lock_message() 88 return PTR_ERR(req); in ceph_lock_message() 89 req->r_inode = inode; in ceph_lock_message() 91 req->r_num_caps = 1; in ceph_lock_message() 106 req->r_args.filelock_change.rule = lock_type; in ceph_lock_message() 107 req->r_args.filelock_change.type = cmd; in ceph_lock_message() 108 req->r_args.filelock_change.owner = cpu_to_le64(owner); in ceph_lock_message() [all …]
|
D | file.c | 185 struct ceph_mds_request *req; in prepare_open_request() local 192 req = ceph_mdsc_create_request(mdsc, op, want_auth); in prepare_open_request() 193 if (IS_ERR(req)) in prepare_open_request() 195 req->r_fmode = ceph_flags_to_mode(flags); in prepare_open_request() 196 req->r_args.open.flags = ceph_flags_sys2wire(flags); in prepare_open_request() 197 req->r_args.open.mode = cpu_to_le32(create_mode); in prepare_open_request() 199 return req; in prepare_open_request() 292 struct ceph_mds_request *req; in ceph_renew_caps() local 320 req = prepare_open_request(inode->i_sb, flags, 0); in ceph_renew_caps() 321 if (IS_ERR(req)) { in ceph_renew_caps() [all …]
|
/fs/coda/ |
D | psdev.c | 95 struct upc_req *req = NULL; in coda_psdev_write() local 155 req = tmp; in coda_psdev_write() 156 list_del(&req->uc_chain); in coda_psdev_write() 162 if (!req) { in coda_psdev_write() 170 if (req->uc_outSize < nbytes) { in coda_psdev_write() 172 __func__, req->uc_outSize, (long)nbytes, in coda_psdev_write() 174 nbytes = req->uc_outSize; /* don't have more space! */ in coda_psdev_write() 176 if (copy_from_user(req->uc_data, buf, nbytes)) { in coda_psdev_write() 177 req->uc_flags |= CODA_REQ_ABORT; in coda_psdev_write() 178 wake_up(&req->uc_sleep); in coda_psdev_write() [all …]
|
/fs/afs/ |
D | file.c | 189 void afs_put_read(struct afs_read *req) in afs_put_read() argument 193 if (refcount_dec_and_test(&req->usage)) { in afs_put_read() 194 if (req->pages) { in afs_put_read() 195 for (i = 0; i < req->nr_pages; i++) in afs_put_read() 196 if (req->pages[i]) in afs_put_read() 197 put_page(req->pages[i]); in afs_put_read() 198 if (req->pages != req->array) in afs_put_read() 199 kfree(req->pages); in afs_put_read() 201 kfree(req); in afs_put_read() 276 struct afs_read *req; in afs_page_filler() local [all …]
|
/fs/cifs/ |
D | smb2pdu.c | 554 assemble_neg_contexts(struct smb2_negotiate_req *req, in assemble_neg_contexts() argument 557 char *pneg_ctxt = (char *)req; in assemble_neg_contexts() 572 pneg_ctxt = (*total_len) + (char *)req; in assemble_neg_contexts() 573 req->NegotiateContextOffset = cpu_to_le32(*total_len); in assemble_neg_contexts() 593 req->NegotiateContextCount = cpu_to_le16(5); in assemble_neg_contexts() 595 req->NegotiateContextCount = cpu_to_le16(4); in assemble_neg_contexts() 767 struct smb2_create_req *req = iov[0].iov_base; in add_posix_context() local 776 if (!req->CreateContextsOffset) in add_posix_context() 777 req->CreateContextsOffset = cpu_to_le32( in add_posix_context() 780 le32_add_cpu(&req->CreateContextsLength, sizeof(struct create_posix)); in add_posix_context() [all …]
|
/fs/ecryptfs/ |
D | kthread.c | 46 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local 58 req = list_first_entry(&ecryptfs_kthread_ctl.req_list, in ecryptfs_threadfn() 61 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn() 62 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn() 64 complete(&req->done); in ecryptfs_threadfn() 91 struct ecryptfs_open_req *req, *tmp; in ecryptfs_destroy_kthread() local 95 list_for_each_entry_safe(req, tmp, &ecryptfs_kthread_ctl.req_list, in ecryptfs_destroy_kthread() 97 list_del(&req->kthread_ctl_list); in ecryptfs_destroy_kthread() 98 *req->lower_file = ERR_PTR(-EIO); in ecryptfs_destroy_kthread() 99 complete(&req->done); in ecryptfs_destroy_kthread() [all …]
|
/fs/verity/ |
D | hash_algs.c | 108 struct ahash_request *req = NULL; in fsverity_prepare_hash_state() local 122 req = ahash_request_alloc(alg->tfm, GFP_KERNEL); in fsverity_prepare_hash_state() 123 if (!req) { in fsverity_prepare_hash_state() 144 ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_SLEEP | in fsverity_prepare_hash_state() 147 ahash_request_set_crypt(req, &sg, NULL, padded_salt_size); in fsverity_prepare_hash_state() 149 err = crypto_wait_req(crypto_ahash_init(req), &wait); in fsverity_prepare_hash_state() 153 err = crypto_wait_req(crypto_ahash_update(req), &wait); in fsverity_prepare_hash_state() 157 err = crypto_ahash_export(req, hashstate); in fsverity_prepare_hash_state() 161 ahash_request_free(req); in fsverity_prepare_hash_state() 186 struct ahash_request *req, struct page *page, u8 *out) in fsverity_hash_page() argument [all …]
|
/fs/ocfs2/ |
D | ioctl.c | 43 struct ocfs2_info_request __user *req) in o2info_set_request_error() argument 46 (void)put_user(kreq->ir_flags, (__u32 __user *)&(req->ir_flags)); in o2info_set_request_error() 49 static inline void o2info_set_request_filled(struct ocfs2_info_request *req) in o2info_set_request_filled() argument 51 req->ir_flags |= OCFS2_INFO_FL_FILLED; in o2info_set_request_filled() 54 static inline void o2info_clear_request_filled(struct ocfs2_info_request *req) in o2info_clear_request_filled() argument 56 req->ir_flags &= ~OCFS2_INFO_FL_FILLED; in o2info_clear_request_filled() 59 static inline int o2info_coherent(struct ocfs2_info_request *req) in o2info_coherent() argument 61 return (!(req->ir_flags & OCFS2_INFO_FL_NON_COHERENT)); in o2info_coherent() 140 struct ocfs2_info_request __user *req) in ocfs2_info_handle_blocksize() argument 144 if (o2info_from_user(oib, req)) in ocfs2_info_handle_blocksize() [all …]
|
/fs/orangefs/ |
D | orangefs-sysfs.c | 351 new_op->upcall.req.param.type = ORANGEFS_PARAM_REQUEST_GET; in sysfs_service_op_show() 365 new_op->upcall.req.param.op = in sysfs_service_op_show() 369 new_op->upcall.req.param.op = in sysfs_service_op_show() 373 new_op->upcall.req.param.op = in sysfs_service_op_show() 378 new_op->upcall.req.param.op = in sysfs_service_op_show() 383 new_op->upcall.req.param.op = in sysfs_service_op_show() 388 new_op->upcall.req.param.op = in sysfs_service_op_show() 393 new_op->upcall.req.param.op = in sysfs_service_op_show() 397 new_op->upcall.req.param.op = in sysfs_service_op_show() 401 new_op->upcall.req.param.op = in sysfs_service_op_show() [all …]
|