/fs/jffs2/ |
D | compr_rubin.c | 39 unsigned long q; member 92 rs->q = 0; in init_rubin() 108 while ((rs->q >= UPPER_BIT_RUBIN) || in encode() 109 ((rs->p + rs->q) <= UPPER_BIT_RUBIN)) { in encode() 112 ret = pushbit(&rs->pp, (rs->q & UPPER_BIT_RUBIN) ? 1 : 0, 0); in encode() 115 rs->q &= LOWER_BITS_RUBIN; in encode() 116 rs->q <<= 1; in encode() 132 rs->q += i0; in encode() 144 pushbit(&rs->pp, (UPPER_BIT_RUBIN & rs->q) ? 1 : 0, 1); in end_rubin() 145 rs->q &= LOWER_BITS_RUBIN; in end_rubin() [all …]
|
/fs/orangefs/ |
D | orangefs-bufmap.c | 13 wait_queue_head_t q; member 20 .q = __WAIT_QUEUE_HEAD_INITIALIZER(rw_map.q) 24 .q = __WAIT_QUEUE_HEAD_INITIALIZER(readdir_map.q) 30 spin_lock(&m->q.lock); in install() 33 wake_up_all_locked(&m->q); in install() 34 spin_unlock(&m->q.lock); in install() 39 spin_lock(&m->q.lock); in mark_killed() 41 spin_unlock(&m->q.lock); in mark_killed() 47 spin_lock(&m->q.lock); in run_down() 51 __add_wait_queue_entry_tail(&m->q, &wait); in run_down() [all …]
|
/fs/hpfs/ |
D | alloc.c | 122 unsigned i, q; in alloc_in_bmp() local 138 q = nr + n; b = 0; in alloc_in_bmp() 139 while ((a = tstbits(bmp, q, n + forward)) != 0) { in alloc_in_bmp() 140 q += a; in alloc_in_bmp() 141 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp() 143 if (q>>5 != nr>>5) { in alloc_in_bmp() 145 q = nr & 0x1f; in alloc_in_bmp() 147 } else if (q > nr) break; in alloc_in_bmp() 150 ret = bs + q; in alloc_in_bmp() 159 q = i<<5; in alloc_in_bmp() [all …]
|
D | ea.c | 289 secno q = hpfs_alloc_sector(s, fno, 1, 0); in hpfs_set_ea() local 290 if (!q) goto bail; in hpfs_set_ea() 291 fnode->ea_secno = cpu_to_le32(q); in hpfs_set_ea()
|
/fs/xfs/ |
D | xfs_qm_syscalls.c | 101 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_scall_quotaoff() local 124 ASSERT(q); in xfs_qm_scall_quotaoff() 125 mutex_lock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 136 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 241 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 249 if ((dqtype & XFS_QMOPT_UQUOTA) && q->qi_uquotaip) { in xfs_qm_scall_quotaoff() 250 xfs_irele(q->qi_uquotaip); in xfs_qm_scall_quotaoff() 251 q->qi_uquotaip = NULL; in xfs_qm_scall_quotaoff() 253 if ((dqtype & XFS_QMOPT_GQUOTA) && q->qi_gquotaip) { in xfs_qm_scall_quotaoff() 254 xfs_irele(q->qi_gquotaip); in xfs_qm_scall_quotaoff() [all …]
|
D | xfs_trans_dquot.c | 291 struct xfs_dqtrx *q) in xfs_trans_dqlockedjoin() argument 293 ASSERT(q[0].qt_dquot != NULL); in xfs_trans_dqlockedjoin() 294 if (q[1].qt_dquot == NULL) { in xfs_trans_dqlockedjoin() 295 xfs_dqlock(q[0].qt_dquot); in xfs_trans_dqlockedjoin() 296 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin() 299 xfs_dqlock2(q[0].qt_dquot, q[1].qt_dquot); in xfs_trans_dqlockedjoin() 300 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin() 301 xfs_trans_dqjoin(tp, q[1].qt_dquot); in xfs_trans_dqlockedjoin() 640 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_trans_dqresv() local 647 defq = xfs_get_defquota(q, xfs_dquot_type(dqp)); in xfs_trans_dqresv() [all …]
|
D | xfs_quotaops.c | 60 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_fs_get_quota_state() local 65 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state() 79 xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, q->qi_uquotaip, in xfs_fs_get_quota_state() 80 mp->m_sb.sb_uquotino, &q->qi_usr_default); in xfs_fs_get_quota_state() 81 xfs_qm_fill_state(&state->s_state[GRPQUOTA], mp, q->qi_gquotaip, in xfs_fs_get_quota_state() 82 mp->m_sb.sb_gquotino, &q->qi_grp_default); in xfs_fs_get_quota_state() 83 xfs_qm_fill_state(&state->s_state[PRJQUOTA], mp, q->qi_pquotaip, in xfs_fs_get_quota_state() 84 mp->m_sb.sb_pquotino, &q->qi_prj_default); in xfs_fs_get_quota_state()
|
D | xfs_discard.c | 155 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim() local 156 unsigned int granularity = q->limits.discard_granularity; in xfs_ioc_trim() 165 if (!blk_queue_discard(q)) in xfs_ioc_trim()
|
D | xfs_dquot.c | 73 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_adjust_dqlimits() local 78 defq = xfs_get_defquota(q, xfs_dquot_type(dq)); in xfs_qm_adjust_dqlimits() 186 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_init_dquot_blk() local 219 curid = id - (id % q->qi_dqperchunk); in xfs_qm_init_dquot_blk() 220 memset(d, 0, BBTOB(q->qi_dqchunklen)); in xfs_qm_init_dquot_blk() 221 for (i = 0; i < q->qi_dqperchunk; i++, d++, curid++) { in xfs_qm_init_dquot_blk() 256 xfs_trans_log_buf(tp, bp, 0, BBTOB(q->qi_dqchunklen) - 1); in xfs_qm_init_dquot_blk()
|
/fs/afs/ |
D | addr_list.c | 136 const char *q, *stop; in afs_parse_text_addrs() local 148 q = memchr(p, ']', end - p); in afs_parse_text_addrs() 150 for (q = p; q < end; q++) in afs_parse_text_addrs() 151 if (*q == '+' || *q == delim) in afs_parse_text_addrs() 155 if (in4_pton(p, q - p, (u8 *)&x[0], -1, &stop)) { in afs_parse_text_addrs() 157 } else if (in6_pton(p, q - p, (u8 *)x, -1, &stop)) { in afs_parse_text_addrs() 164 p = q; in afs_parse_text_addrs() 170 if (q < end && *q == ']') in afs_parse_text_addrs()
|
/fs/proc/ |
D | bootconfig.c | 31 char q; in copy_xbc_key_value_list() local 50 q = '\''; in copy_xbc_key_value_list() 52 q = '"'; in copy_xbc_key_value_list() 54 q, val, q, xbc_node_is_array(vnode) ? ", " : "\n"); in copy_xbc_key_value_list()
|
/fs/erofs/ |
D | zdata.c | 965 struct z_erofs_decompressqueue *q = tagptr_unfold_ptr(t); in z_erofs_decompressqueue_endio() local 979 if (erofs_page_is_managed(EROFS_SB(q->sb), page)) { in z_erofs_decompressqueue_endio() 985 z_erofs_decompress_kickoff(q, tagptr_unfold_tags(t), -1); in z_erofs_decompressqueue_endio() 1350 struct z_erofs_decompressqueue *q; in jobqueue_init() local 1353 q = kvzalloc(sizeof(*q), GFP_KERNEL | __GFP_NOWARN); in jobqueue_init() 1354 if (!q) { in jobqueue_init() 1359 kthread_init_work(&q->u.kthread_work, in jobqueue_init() 1362 INIT_WORK(&q->u.work, z_erofs_decompressqueue_work); in jobqueue_init() 1366 q = fgq; in jobqueue_init() 1370 q->sb = sb; in jobqueue_init() [all …]
|
/fs/ |
D | fs_pin.c | 88 struct hlist_node *q; in group_pin_kill() local 90 q = READ_ONCE(p->first); in group_pin_kill() 91 if (!q) { in group_pin_kill() 95 pin_kill(hlist_entry(q, struct fs_pin, s_list)); in group_pin_kill()
|
D | namespace.c | 1817 struct mount *res, *p, *q, *r, *parent; in copy_tree() local 1825 res = q = clone_mnt(mnt, dentry, flag); in copy_tree() 1826 if (IS_ERR(q)) in copy_tree() 1827 return q; in copy_tree() 1829 q->mnt_mountpoint = mnt->mnt_mountpoint; in copy_tree() 1842 q = ERR_PTR(-EPERM); in copy_tree() 1856 q = q->mnt_parent; in copy_tree() 1859 parent = q; in copy_tree() 1860 q = clone_mnt(p, p->mnt.mnt_root, flag); in copy_tree() 1861 if (IS_ERR(q)) in copy_tree() [all …]
|
D | fs_context.c | 386 char *q = kasprintf(GFP_KERNEL, "%c %s%s%pV\n", level, in logfc() local 400 log->buffer[index] = q ? q : "OOM: Can't store error string"; in logfc() 401 if (q) in logfc()
|
/fs/minix/ |
D | itree_common.c | 215 static inline int all_zeroes(block_t *p, block_t *q) in all_zeroes() argument 217 while (p < q) in all_zeroes() 263 static inline void free_data(struct inode *inode, block_t *p, block_t *q) in free_data() argument 267 for ( ; p < q ; p++) { in free_data() 276 static void free_branches(struct inode *inode, block_t *p, block_t *q, int depth) in free_branches() argument 282 for ( ; p < q ; p++) { in free_branches() 297 free_data(inode, p, q); in free_branches()
|
/fs/efivarfs/ |
D | super.c | 90 struct qstr q; in efivarfs_alloc_dentry() local 93 q.name = name; in efivarfs_alloc_dentry() 94 q.len = strlen(name); in efivarfs_alloc_dentry() 96 err = efivarfs_d_hash(parent, &q); in efivarfs_alloc_dentry() 100 d = d_alloc(parent, &q); in efivarfs_alloc_dentry()
|
/fs/gfs2/ |
D | quota.c | 842 struct gfs2_quota q; in gfs2_adjust_quota() local 852 memset(&q, 0, sizeof(struct gfs2_quota)); in gfs2_adjust_quota() 853 err = gfs2_internal_read(ip, (char *)&q, &loc, sizeof(q)); in gfs2_adjust_quota() 857 loc -= sizeof(q); /* gfs2_internal_read would've advanced the loc ptr */ in gfs2_adjust_quota() 859 be64_add_cpu(&q.qu_value, change); in gfs2_adjust_quota() 860 if (((s64)be64_to_cpu(q.qu_value)) < 0) in gfs2_adjust_quota() 861 q.qu_value = 0; /* Never go negative on quota usage */ in gfs2_adjust_quota() 862 qd->qd_qb.qb_value = q.qu_value; in gfs2_adjust_quota() 865 q.qu_warn = cpu_to_be64(fdq->d_spc_softlimit >> sdp->sd_sb.sb_bsize_shift); in gfs2_adjust_quota() 866 qd->qd_qb.qb_warn = q.qu_warn; in gfs2_adjust_quota() [all …]
|
/fs/ufs/ |
D | inode.c | 131 Indirect chain[4], *q = chain; in ufs_frag_map() local 149 if (!grow_chain32(ufsi, NULL, &ufsi->i_u1.i_data[*p++], chain, q)) in ufs_frag_map() 151 if (!q->key32) in ufs_frag_map() 159 fs32_to_cpu(sb, q->key32) + (n>>shift)); in ufs_frag_map() 163 if (!grow_chain32(ufsi, bh, ptr, chain, ++q)) in ufs_frag_map() 165 if (!q->key32) in ufs_frag_map() 168 res = fs32_to_cpu(sb, q->key32); in ufs_frag_map() 172 if (!grow_chain64(ufsi, NULL, &ufsi->i_u1.u2_i_data[*p++], chain, q)) in ufs_frag_map() 174 if (!q->key64) in ufs_frag_map() 183 fs64_to_cpu(sb, q->key64) + (n>>shift)); in ufs_frag_map() [all …]
|
/fs/sysv/ |
D | itree.c | 273 static inline int all_zeroes(sysv_zone_t *p, sysv_zone_t *q) in all_zeroes() argument 275 while (p < q) in all_zeroes() 330 static inline void free_data(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q) in free_data() argument 332 for ( ; p < q ; p++) { in free_data() 342 static void free_branches(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q, int depth) in free_branches() argument 348 for ( ; p < q ; p++) { in free_branches() 365 free_data(inode, p, q); in free_branches()
|
/fs/jfs/ |
D | ioctl.c | 125 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl() local 132 if (!blk_queue_discard(q)) { in jfs_ioctl() 142 q->limits.discard_granularity); in jfs_ioctl()
|
/fs/nfsd/ |
D | blocklayout.c | 217 struct request_queue *q = bdev->bd_disk->queue; in nfsd4_scsi_identify_device() local 231 if (WARN_ON_ONCE(!blk_queue_scsi_passthrough(q))) in nfsd4_scsi_identify_device() 239 rq = blk_get_request(q, REQ_OP_SCSI_IN, 0); in nfsd4_scsi_identify_device() 246 error = blk_rq_map_kern(q, rq, buf, bufflen, GFP_KERNEL); in nfsd4_scsi_identify_device() 257 blk_execute_rq(rq->q, NULL, rq, 1); in nfsd4_scsi_identify_device()
|
/fs/crypto/ |
D | inline_crypt.c | 221 struct request_queue *q; in fscrypt_derive_raw_secret() local 223 q = bdev_get_queue(sb->s_bdev); in fscrypt_derive_raw_secret() 224 if (!q->ksm) in fscrypt_derive_raw_secret() 227 return blk_ksm_derive_raw_secret(q->ksm, wrapped_key, wrapped_key_size, in fscrypt_derive_raw_secret()
|
/fs/cifs/ |
D | dir.c | 823 static int cifs_ci_hash(const struct dentry *dentry, struct qstr *q) in cifs_ci_hash() argument 831 for (i = 0; i < q->len; i += charlen) { in cifs_ci_hash() 832 charlen = codepage->char2uni(&q->name[i], q->len - i, &c); in cifs_ci_hash() 838 q->hash = end_name_hash(hash); in cifs_ci_hash()
|
/fs/autofs/ |
D | expire.c | 101 struct dentry *q; in get_next_positive_subdir() local 105 q = positive_after(root, prev); in get_next_positive_subdir() 109 return q; in get_next_positive_subdir()
|