/fs/nfsd/ |
D | nfscache.c | 97 struct svc_cacherep *rp; in nfsd_reply_cache_alloc() local 99 rp = kmem_cache_alloc(drc_slab, GFP_KERNEL); in nfsd_reply_cache_alloc() 100 if (rp) { in nfsd_reply_cache_alloc() 101 rp->c_state = RC_UNUSED; in nfsd_reply_cache_alloc() 102 rp->c_type = RC_NOCACHE; in nfsd_reply_cache_alloc() 103 RB_CLEAR_NODE(&rp->c_node); in nfsd_reply_cache_alloc() 104 INIT_LIST_HEAD(&rp->c_lru); in nfsd_reply_cache_alloc() 106 memset(&rp->c_key, 0, sizeof(rp->c_key)); in nfsd_reply_cache_alloc() 107 rp->c_key.k_xid = rqstp->rq_xid; in nfsd_reply_cache_alloc() 108 rp->c_key.k_proc = rqstp->rq_proc; in nfsd_reply_cache_alloc() [all …]
|
D | trace.h | 578 const struct svc_cacherep *rp 580 TP_ARGS(nn, key, rp), 591 __entry->ingress = (__force u32)rp->c_key.k_csum;
|
D | nfs4xdr.c | 5261 struct nfs4_replay *rp = op->replay; in nfsd4_encode_replay() local 5263 p = xdr_reserve_space(xdr, 8 + rp->rp_buflen); in nfsd4_encode_replay() 5269 *p++ = rp->rp_status; /* already xdr'ed */ in nfsd4_encode_replay() 5271 p = xdr_encode_opaque_fixed(p, rp->rp_buf, rp->rp_buflen); in nfsd4_encode_replay()
|
D | nfs4state.c | 4182 static void init_nfs4_replay(struct nfs4_replay *rp) in init_nfs4_replay() argument 4184 rp->rp_status = nfserr_serverfault; in init_nfs4_replay() 4185 rp->rp_buflen = 0; in init_nfs4_replay() 4186 rp->rp_buf = rp->rp_ibuf; in init_nfs4_replay() 4187 mutex_init(&rp->rp_mutex); in init_nfs4_replay()
|
D | nfs4proc.c | 387 struct nfs4_replay *rp = &open->op_openowner->oo_owner.so_replay; in nfsd4_open() local 390 &rp->rp_openfh); in nfsd4_open()
|
/fs/ntfs/ |
D | logfile.c | 39 RESTART_PAGE_HEADER *rp, s64 pos) in ntfs_check_restart_page_header() argument 50 logfile_system_page_size = le32_to_cpu(rp->system_page_size); in ntfs_check_restart_page_header() 51 logfile_log_page_size = le32_to_cpu(rp->log_page_size); in ntfs_check_restart_page_header() 70 if (sle16_to_cpu(rp->major_ver) != 1 || in ntfs_check_restart_page_header() 71 sle16_to_cpu(rp->minor_ver) != 1) { in ntfs_check_restart_page_header() 74 "1.1 only.)", (int)sle16_to_cpu(rp->major_ver), in ntfs_check_restart_page_header() 75 (int)sle16_to_cpu(rp->minor_ver)); in ntfs_check_restart_page_header() 82 if (ntfs_is_chkd_record(rp->magic) && !le16_to_cpu(rp->usa_count)) { in ntfs_check_restart_page_header() 88 if (usa_count != le16_to_cpu(rp->usa_count)) { in ntfs_check_restart_page_header() 94 usa_ofs = le16_to_cpu(rp->usa_ofs); in ntfs_check_restart_page_header() [all …]
|
D | logfile.h | 286 RESTART_PAGE_HEADER **rp); 289 const RESTART_PAGE_HEADER *rp);
|
D | super.c | 1203 RESTART_PAGE_HEADER **rp) in load_and_check_logfile() argument 1215 if (!ntfs_check_logfile(tmp_ino, rp)) { in load_and_check_logfile() 1765 RESTART_PAGE_HEADER *rp; in load_system_files() local 1940 rp = NULL; in load_system_files() 1941 if (!load_and_check_logfile(vol, &rp) || in load_system_files() 1942 !ntfs_is_logfile_clean(vol->logfile_ino, rp)) { in load_system_files() 1958 BUG_ON(!rp); in load_system_files() 1959 ntfs_free(rp); in load_system_files() 1971 ntfs_free(rp); in load_system_files()
|
D | layout.h | 1081 } __attribute__ ((__packed__)) rp; member
|
/fs/cifs/ |
D | cifs_unicode.h | 326 register const struct UniCaseRange *rp; in UniToupper() local 332 rp = CifsUniUpperRange; /* Use range tables */ in UniToupper() 333 while (rp->start) { in UniToupper() 334 if (uc < rp->start) /* Before start of range */ in UniToupper() 336 if (uc <= rp->end) /* In range */ in UniToupper() 337 return uc + rp->table[uc - rp->start]; in UniToupper() 338 rp++; /* Try next range */ in UniToupper() 368 register const struct UniCaseRange *rp; in UniTolower() local 374 rp = CifsUniLowerRange; /* Use range tables */ in UniTolower() 375 while (rp->start) { in UniTolower() [all …]
|
/fs/jfs/ |
D | jfs_unicode.h | 110 UNICASERANGE *rp; in UniToupper() local 115 rp = UniUpperRange; /* Use range tables */ in UniToupper() 116 while (rp->start) { in UniToupper() 117 if (uc < rp->start) /* Before start of range */ in UniToupper() 119 if (uc <= rp->end) /* In range */ in UniToupper() 120 return uc + rp->table[uc - rp->start]; in UniToupper() 121 rp++; /* Try next range */ in UniToupper()
|
D | jfs_dtree.c | 169 static int ciGetLeafPrefixKey(dtpage_t * lp, int li, dtpage_t * rp, 931 dtpage_t *rp; /* new right page split from sp */ in dtSplitUp() local 1089 if ((rc = dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd))) { in dtSplitUp() 1169 switch (rp->header.flag & BT_TYPE) { in dtSplitUp() 1181 rp, 0, &key, in dtSplitUp() 1194 dtGetKey(rp, 0, &key, sbi->mntflag); in dtSplitUp() 1205 dtGetKey(rp, 0, &key, sbi->mntflag); in dtSplitUp() 1242 dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd); in dtSplitUp() 1330 dtpage_t *rp; /* new right page allocated */ in dtSplitPage() local 1380 rp = (dtpage_t *) rmp->data; in dtSplitPage() [all …]
|
D | jfs_xtree.c | 952 xtpage_t *rp; /* new right page allocated */ in xtSplitPage() local 998 rp = (xtpage_t *) rmp->data; in xtSplitPage() 999 rp->header.self = *pxd; in xtSplitPage() 1000 rp->header.flag = sp->header.flag & BT_TYPE; in xtSplitPage() 1001 rp->header.maxentry = sp->header.maxentry; /* little-endian */ in xtSplitPage() 1002 rp->header.nextindex = cpu_to_le16(XTENTRYSTART); in xtSplitPage() 1024 rp->header.next = cpu_to_le64(nextbn); in xtSplitPage() 1025 rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self)); in xtSplitPage() 1050 xad = &rp->xad[XTENTRYSTART]; in xtSplitPage() 1054 rp->header.nextindex = cpu_to_le16(XTENTRYSTART + 1); in xtSplitPage() [all …]
|
/fs/ |
D | binfmt_flat.c | 430 static inline u32 __user *skip_got_header(u32 __user *rp) in skip_got_header() argument 441 if (get_user(rp_val0, rp)) in skip_got_header() 442 return rp; in skip_got_header() 443 if (get_user(rp_val1, rp + 1)) in skip_got_header() 444 return rp; in skip_got_header() 447 rp += 4; in skip_got_header() 449 rp += 2; in skip_got_header() 451 return rp; in skip_got_header() 462 u32 __user *rp; in load_flat_file() local 801 rp = skip_got_header((u32 __user *) datapos); in load_flat_file() [all …]
|
/fs/ceph/ |
D | debugfs.c | 57 struct rb_node *rp; in mdsc_show() local 63 for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) { in mdsc_show() 64 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show()
|
/fs/dlm/ |
D | debug_fs.c | 505 struct dlm_rsb *r, *rp; in table_seq_next() local 517 rp = ri->rsb; in table_seq_next() 518 next = rb_next(&rp->res_hashnode); in table_seq_next() 525 dlm_put_rsb(rp); in table_seq_next() 530 dlm_put_rsb(rp); in table_seq_next()
|
/fs/nfs/blocklayout/ |
D | blocklayout.c | 572 static int decode_sector_number(__be32 **rp, sector_t *sp) in decode_sector_number() argument 576 *rp = xdr_decode_hyper(*rp, &s); in decode_sector_number()
|
/fs/xfs/libxfs/ |
D | xfs_btree.c | 2200 union xfs_btree_rec *rp; in xfs_btree_update() local 2212 rp = xfs_btree_rec_addr(cur, ptr, block); in xfs_btree_update() 2215 xfs_btree_copy_recs(cur, rp, rec, 1); in xfs_btree_update() 3318 union xfs_btree_rec *rp; in xfs_btree_insrec() local 3320 rp = xfs_btree_rec_addr(cur, ptr, block); in xfs_btree_insrec() 3322 xfs_btree_shift_recs(cur, rp, 1, numrecs - ptr + 1); in xfs_btree_insrec() 3325 xfs_btree_copy_recs(cur, rp, rec, 1); in xfs_btree_insrec() 3330 ASSERT(cur->bc_ops->recs_inorder(cur, rp, in xfs_btree_insrec()
|