/fs/nfsd/ |
D | nfscache.c | 124 struct svc_cacherep *rp; in nfsd_reply_cache_alloc() local 126 rp = kmem_cache_alloc(drc_slab, GFP_KERNEL); in nfsd_reply_cache_alloc() 127 if (rp) { in nfsd_reply_cache_alloc() 128 rp->c_state = RC_UNUSED; in nfsd_reply_cache_alloc() 129 rp->c_type = RC_NOCACHE; in nfsd_reply_cache_alloc() 130 INIT_LIST_HEAD(&rp->c_lru); in nfsd_reply_cache_alloc() 132 return rp; in nfsd_reply_cache_alloc() 136 nfsd_reply_cache_free_locked(struct svc_cacherep *rp) in nfsd_reply_cache_free_locked() argument 138 if (rp->c_type == RC_REPLBUFF && rp->c_replvec.iov_base) { in nfsd_reply_cache_free_locked() 139 drc_mem_usage -= rp->c_replvec.iov_len; in nfsd_reply_cache_free_locked() [all …]
|
D | nfs4xdr.c | 4417 struct nfs4_replay *rp = op->replay; in nfsd4_encode_replay() local 4419 BUG_ON(!rp); in nfsd4_encode_replay() 4421 p = xdr_reserve_space(xdr, 8 + rp->rp_buflen); in nfsd4_encode_replay() 4427 *p++ = rp->rp_status; /* already xdr'ed */ in nfsd4_encode_replay() 4429 p = xdr_encode_opaque_fixed(p, rp->rp_buf, rp->rp_buflen); in nfsd4_encode_replay()
|
D | nfs4state.c | 3308 static void init_nfs4_replay(struct nfs4_replay *rp) in init_nfs4_replay() argument 3310 rp->rp_status = nfserr_serverfault; in init_nfs4_replay() 3311 rp->rp_buflen = 0; in init_nfs4_replay() 3312 rp->rp_buf = rp->rp_ibuf; in init_nfs4_replay() 3313 mutex_init(&rp->rp_mutex); in init_nfs4_replay()
|
D | nfs4proc.c | 393 struct nfs4_replay *rp = &open->op_openowner->oo_owner.so_replay; in nfsd4_open() local 396 &rp->rp_openfh); in nfsd4_open()
|
/fs/ntfs/ |
D | logfile.c | 52 RESTART_PAGE_HEADER *rp, s64 pos) in ntfs_check_restart_page_header() argument 63 logfile_system_page_size = le32_to_cpu(rp->system_page_size); in ntfs_check_restart_page_header() 64 logfile_log_page_size = le32_to_cpu(rp->log_page_size); in ntfs_check_restart_page_header() 83 if (sle16_to_cpu(rp->major_ver) != 1 || in ntfs_check_restart_page_header() 84 sle16_to_cpu(rp->minor_ver) != 1) { in ntfs_check_restart_page_header() 87 "1.1 only.)", (int)sle16_to_cpu(rp->major_ver), in ntfs_check_restart_page_header() 88 (int)sle16_to_cpu(rp->minor_ver)); in ntfs_check_restart_page_header() 95 if (ntfs_is_chkd_record(rp->magic) && !le16_to_cpu(rp->usa_count)) { in ntfs_check_restart_page_header() 101 if (usa_count != le16_to_cpu(rp->usa_count)) { in ntfs_check_restart_page_header() 107 usa_ofs = le16_to_cpu(rp->usa_ofs); in ntfs_check_restart_page_header() [all …]
|
D | logfile.h | 300 RESTART_PAGE_HEADER **rp); 303 const RESTART_PAGE_HEADER *rp);
|
D | super.c | 1217 RESTART_PAGE_HEADER **rp) in load_and_check_logfile() argument 1229 if (!ntfs_check_logfile(tmp_ino, rp)) { in load_and_check_logfile() 1779 RESTART_PAGE_HEADER *rp; in load_system_files() local 1954 rp = NULL; in load_system_files() 1955 if (!load_and_check_logfile(vol, &rp) || in load_system_files() 1956 !ntfs_is_logfile_clean(vol->logfile_ino, rp)) { in load_system_files() 1972 BUG_ON(!rp); in load_system_files() 1973 ntfs_free(rp); in load_system_files() 1985 ntfs_free(rp); in load_system_files()
|
D | layout.h | 1095 } __attribute__ ((__packed__)) rp; member
|
/fs/cifs/ |
D | cifs_unicode.h | 343 register const struct UniCaseRange *rp; in UniToupper() local 349 rp = CifsUniUpperRange; /* Use range tables */ in UniToupper() 350 while (rp->start) { in UniToupper() 351 if (uc < rp->start) /* Before start of range */ in UniToupper() 353 if (uc <= rp->end) /* In range */ in UniToupper() 354 return uc + rp->table[uc - rp->start]; in UniToupper() 355 rp++; /* Try next range */ in UniToupper() 385 register const struct UniCaseRange *rp; in UniTolower() local 391 rp = CifsUniLowerRange; /* Use range tables */ in UniTolower() 392 while (rp->start) { in UniTolower() [all …]
|
/fs/jfs/ |
D | jfs_unicode.h | 123 UNICASERANGE *rp; in UniToupper() local 128 rp = UniUpperRange; /* Use range tables */ in UniToupper() 129 while (rp->start) { in UniToupper() 130 if (uc < rp->start) /* Before start of range */ in UniToupper() 132 if (uc <= rp->end) /* In range */ in UniToupper() 133 return uc + rp->table[uc - rp->start]; in UniToupper() 134 rp++; /* Try next range */ in UniToupper()
|
D | jfs_dtree.c | 182 static int ciGetLeafPrefixKey(dtpage_t * lp, int li, dtpage_t * rp, 938 dtpage_t *rp; /* new right page split from sp */ in dtSplitUp() local 1096 if ((rc = dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd))) { in dtSplitUp() 1176 switch (rp->header.flag & BT_TYPE) { in dtSplitUp() 1188 rp, 0, &key, in dtSplitUp() 1201 dtGetKey(rp, 0, &key, sbi->mntflag); in dtSplitUp() 1212 dtGetKey(rp, 0, &key, sbi->mntflag); in dtSplitUp() 1249 dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd); in dtSplitUp() 1337 dtpage_t *rp; /* new right page allocated */ in dtSplitPage() local 1387 rp = (dtpage_t *) rmp->data; in dtSplitPage() [all …]
|
D | jfs_xtree.c | 965 xtpage_t *rp; /* new right page allocated */ in xtSplitPage() local 1011 rp = (xtpage_t *) rmp->data; in xtSplitPage() 1012 rp->header.self = *pxd; in xtSplitPage() 1013 rp->header.flag = sp->header.flag & BT_TYPE; in xtSplitPage() 1014 rp->header.maxentry = sp->header.maxentry; /* little-endian */ in xtSplitPage() 1015 rp->header.nextindex = cpu_to_le16(XTENTRYSTART); in xtSplitPage() 1037 rp->header.next = cpu_to_le64(nextbn); in xtSplitPage() 1038 rp->header.prev = cpu_to_le64(addressPXD(&sp->header.self)); in xtSplitPage() 1063 xad = &rp->xad[XTENTRYSTART]; in xtSplitPage() 1067 rp->header.nextindex = cpu_to_le16(XTENTRYSTART + 1); in xtSplitPage() [all …]
|
/fs/ |
D | binfmt_flat.c | 436 unsigned long *reloc = 0, *rp; in load_flat_file() local 720 for (rp = (unsigned long *)datapos; *rp != 0xffffffff; rp++) { in load_flat_file() 722 if (*rp) { in load_flat_file() 723 addr = calc_reloc(*rp, libinfo, id, 0); in load_flat_file() 728 *rp = addr; in load_flat_file() 756 rp = (unsigned long *) calc_reloc(addr, libinfo, id, 1); in load_flat_file() 757 if (rp == (unsigned long *)RELOC_FAILED) { in load_flat_file() 763 addr = flat_get_addr_from_rp(rp, relval, flags, in load_flat_file() 779 flat_put_addr_at_rp(rp, addr, relval); in load_flat_file()
|
/fs/ceph/ |
D | debugfs.c | 54 struct rb_node *rp; in mdsc_show() local 60 for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) { in mdsc_show() 61 req = rb_entry(rp, struct ceph_mds_request, r_node); in mdsc_show()
|
/fs/dlm/ |
D | debug_fs.c | 508 struct dlm_rsb *r, *rp; in table_seq_next() local 520 rp = ri->rsb; in table_seq_next() 521 next = rb_next(&rp->res_hashnode); in table_seq_next() 528 dlm_put_rsb(rp); in table_seq_next() 533 dlm_put_rsb(rp); in table_seq_next()
|
/fs/nfs/blocklayout/ |
D | blocklayout.c | 526 static int decode_sector_number(__be32 **rp, sector_t *sp) in decode_sector_number() argument 530 *rp = xdr_decode_hyper(*rp, &s); in decode_sector_number()
|
/fs/ncpfs/ |
D | sock.c | 170 static inline int get_conn_number(struct ncp_reply_header *rp) in get_conn_number() argument 172 return rp->conn_low | (rp->conn_high << 8); in get_conn_number()
|
/fs/xfs/libxfs/ |
D | xfs_btree.c | 1928 union xfs_btree_rec *rp; in xfs_btree_update() local 1943 rp = xfs_btree_rec_addr(cur, ptr, block); in xfs_btree_update() 1946 xfs_btree_copy_recs(cur, rp, rec, 1); in xfs_btree_update() 3056 union xfs_btree_rec *rp; in xfs_btree_insrec() local 3058 rp = xfs_btree_rec_addr(cur, ptr, block); in xfs_btree_insrec() 3060 xfs_btree_shift_recs(cur, rp, 1, numrecs - ptr + 1); in xfs_btree_insrec() 3063 xfs_btree_copy_recs(cur, rp, recp, 1); in xfs_btree_insrec() 3068 ASSERT(cur->bc_ops->recs_inorder(cur, rp, in xfs_btree_insrec()
|