/fs/jffs2/ |
D | compr_rubin.c | 39 unsigned long q; member 92 rs->q = 0; in init_rubin() 108 while ((rs->q >= UPPER_BIT_RUBIN) || in encode() 109 ((rs->p + rs->q) <= UPPER_BIT_RUBIN)) { in encode() 112 ret = pushbit(&rs->pp, (rs->q & UPPER_BIT_RUBIN) ? 1 : 0, 0); in encode() 115 rs->q &= LOWER_BITS_RUBIN; in encode() 116 rs->q <<= 1; in encode() 132 rs->q += i0; in encode() 144 pushbit(&rs->pp, (UPPER_BIT_RUBIN & rs->q) ? 1 : 0, 1); in end_rubin() 145 rs->q &= LOWER_BITS_RUBIN; in end_rubin() [all …]
|
/fs/orangefs/ |
D | orangefs-bufmap.c | 13 wait_queue_head_t q; member 20 .q = __WAIT_QUEUE_HEAD_INITIALIZER(rw_map.q) 24 .q = __WAIT_QUEUE_HEAD_INITIALIZER(readdir_map.q) 30 spin_lock(&m->q.lock); in install() 33 wake_up_all_locked(&m->q); in install() 34 spin_unlock(&m->q.lock); in install() 39 spin_lock(&m->q.lock); in mark_killed() 41 spin_unlock(&m->q.lock); in mark_killed() 47 spin_lock(&m->q.lock); in run_down() 51 __add_wait_queue_entry_tail(&m->q, &wait); in run_down() [all …]
|
/fs/xfs/ |
D | xfs_qm_syscalls.c | 39 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_scall_quotaoff() local 62 ASSERT(q); in xfs_qm_scall_quotaoff() 63 mutex_lock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 74 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 179 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff() 187 if ((dqtype & XFS_QMOPT_UQUOTA) && q->qi_uquotaip) { in xfs_qm_scall_quotaoff() 188 xfs_irele(q->qi_uquotaip); in xfs_qm_scall_quotaoff() 189 q->qi_uquotaip = NULL; in xfs_qm_scall_quotaoff() 191 if ((dqtype & XFS_QMOPT_GQUOTA) && q->qi_gquotaip) { in xfs_qm_scall_quotaoff() 192 xfs_irele(q->qi_gquotaip); in xfs_qm_scall_quotaoff() [all …]
|
D | xfs_quotaops.c | 26 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_fill_state() local 40 tstate->spc_timelimit = q->qi_btimelimit; in xfs_qm_fill_state() 41 tstate->ino_timelimit = q->qi_itimelimit; in xfs_qm_fill_state() 42 tstate->rt_spc_timelimit = q->qi_rtbtimelimit; in xfs_qm_fill_state() 43 tstate->spc_warnlimit = q->qi_bwarnlimit; in xfs_qm_fill_state() 44 tstate->ino_warnlimit = q->qi_iwarnlimit; in xfs_qm_fill_state() 45 tstate->rt_spc_warnlimit = q->qi_rtbwarnlimit; in xfs_qm_fill_state() 60 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_fs_get_quota_state() local 65 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state() 79 xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, q->qi_uquotaip, in xfs_fs_get_quota_state() [all …]
|
D | xfs_trans_dquot.c | 282 struct xfs_dqtrx *q) in xfs_trans_dqlockedjoin() argument 284 ASSERT(q[0].qt_dquot != NULL); in xfs_trans_dqlockedjoin() 285 if (q[1].qt_dquot == NULL) { in xfs_trans_dqlockedjoin() 286 xfs_dqlock(q[0].qt_dquot); in xfs_trans_dqlockedjoin() 287 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin() 290 xfs_dqlock2(q[0].qt_dquot, q[1].qt_dquot); in xfs_trans_dqlockedjoin() 291 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin() 292 xfs_trans_dqjoin(tp, q[1].qt_dquot); in xfs_trans_dqlockedjoin() 588 xfs_quotainfo_t *q = mp->m_quotainfo; in xfs_trans_dqresv() local 594 defq = xfs_get_defquota(dqp, q); in xfs_trans_dqresv() [all …]
|
D | xfs_discard.c | 150 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim() local 151 unsigned int granularity = q->limits.discard_granularity; in xfs_ioc_trim() 160 if (!blk_queue_discard(q)) in xfs_ioc_trim()
|
D | xfs_dquot.c | 72 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_adjust_dqlimits() local 78 defq = xfs_get_defquota(dq, q); in xfs_qm_adjust_dqlimits() 214 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_init_dquot_blk() local 227 curid = id - (id % q->qi_dqperchunk); in xfs_qm_init_dquot_blk() 228 memset(d, 0, BBTOB(q->qi_dqchunklen)); in xfs_qm_init_dquot_blk() 229 for (i = 0; i < q->qi_dqperchunk; i++, d++, curid++) { in xfs_qm_init_dquot_blk() 245 xfs_trans_log_buf(tp, bp, 0, BBTOB(q->qi_dqchunklen) - 1); in xfs_qm_init_dquot_blk()
|
/fs/hpfs/ |
D | alloc.c | 122 unsigned i, q; in alloc_in_bmp() local 138 q = nr + n; b = 0; in alloc_in_bmp() 139 while ((a = tstbits(bmp, q, n + forward)) != 0) { in alloc_in_bmp() 140 q += a; in alloc_in_bmp() 141 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp() 143 if (q>>5 != nr>>5) { in alloc_in_bmp() 145 q = nr & 0x1f; in alloc_in_bmp() 147 } else if (q > nr) break; in alloc_in_bmp() 150 ret = bs + q; in alloc_in_bmp() 159 q = i<<5; in alloc_in_bmp() [all …]
|
D | ea.c | 289 secno q = hpfs_alloc_sector(s, fno, 1, 0); in hpfs_set_ea() local 290 if (!q) goto bail; in hpfs_set_ea() 291 fnode->ea_secno = cpu_to_le32(q); in hpfs_set_ea()
|
/fs/afs/ |
D | addr_list.c | 136 const char *q, *stop; in afs_parse_text_addrs() local 148 q = memchr(p, ']', end - p); in afs_parse_text_addrs() 150 for (q = p; q < end; q++) in afs_parse_text_addrs() 151 if (*q == '+' || *q == delim) in afs_parse_text_addrs() 155 if (in4_pton(p, q - p, (u8 *)&x[0], -1, &stop)) { in afs_parse_text_addrs() 157 } else if (in6_pton(p, q - p, (u8 *)x, -1, &stop)) { in afs_parse_text_addrs() 164 p = q; in afs_parse_text_addrs() 170 if (q < end && *q == ']') in afs_parse_text_addrs()
|
/fs/sdcardfs/ |
D | packagelist.c | 57 static inline void qstr_init(struct qstr *q, const char *name) in qstr_init() argument 59 q->name = name; in qstr_init() 60 q->len = strlen(q->name); in qstr_init() 61 q->hash = full_name_case_hash(0, q->name, q->len); in qstr_init() 92 struct qstr q; in get_appid() local 94 qstr_init(&q, key); in get_appid() 95 return __get_appid(&q); in get_appid() 118 struct qstr q; in get_ext_gid() local 120 qstr_init(&q, key); in get_ext_gid() 121 return __get_ext_gid(&q); in get_ext_gid() [all …]
|
/fs/ |
D | fs_pin.c | 88 struct hlist_node *q; in group_pin_kill() local 90 q = READ_ONCE(p->first); in group_pin_kill() 91 if (!q) { in group_pin_kill() 95 pin_kill(hlist_entry(q, struct fs_pin, s_list)); in group_pin_kill()
|
D | fs_context.c | 394 char *q; in logfc() local 407 q = kvasprintf(GFP_KERNEL, fmt, va); in logfc() 409 if (!q) in logfc() 420 q = kstrdup(p, GFP_KERNEL); in logfc() 426 q = (char *)p; in logfc() 432 printk(KERN_WARNING "%s\n", q + 2); in logfc() 435 printk(KERN_ERR "%s\n", q + 2); in logfc() 438 printk(KERN_NOTICE "%s\n", q + 2); in logfc() 442 kfree(q); in logfc() 457 log->buffer[index] = q; in logfc()
|
D | namespace.c | 1774 struct mount *res, *p, *q, *r, *parent; in copy_tree() local 1782 res = q = clone_mnt(mnt, dentry, flag); in copy_tree() 1783 if (IS_ERR(q)) in copy_tree() 1784 return q; in copy_tree() 1786 q->mnt_mountpoint = mnt->mnt_mountpoint; in copy_tree() 1799 q = ERR_PTR(-EPERM); in copy_tree() 1813 q = q->mnt_parent; in copy_tree() 1816 parent = q; in copy_tree() 1817 q = clone_mnt(p, p->mnt.mnt_root, flag); in copy_tree() 1818 if (IS_ERR(q)) in copy_tree() [all …]
|
/fs/efivarfs/ |
D | super.c | 89 struct qstr q; in efivarfs_alloc_dentry() local 92 q.name = name; in efivarfs_alloc_dentry() 93 q.len = strlen(name); in efivarfs_alloc_dentry() 95 err = efivarfs_d_hash(parent, &q); in efivarfs_alloc_dentry() 99 d = d_alloc(parent, &q); in efivarfs_alloc_dentry()
|
/fs/minix/ |
D | itree_common.c | 209 static inline int all_zeroes(block_t *p, block_t *q) in all_zeroes() argument 211 while (p < q) in all_zeroes() 257 static inline void free_data(struct inode *inode, block_t *p, block_t *q) in free_data() argument 261 for ( ; p < q ; p++) { in free_data() 270 static void free_branches(struct inode *inode, block_t *p, block_t *q, int depth) in free_branches() argument 276 for ( ; p < q ; p++) { in free_branches() 291 free_data(inode, p, q); in free_branches()
|
/fs/gfs2/ |
D | quota.c | 816 struct gfs2_quota q; in gfs2_adjust_quota() local 826 memset(&q, 0, sizeof(struct gfs2_quota)); in gfs2_adjust_quota() 827 err = gfs2_internal_read(ip, (char *)&q, &loc, sizeof(q)); in gfs2_adjust_quota() 831 loc -= sizeof(q); /* gfs2_internal_read would've advanced the loc ptr */ in gfs2_adjust_quota() 833 be64_add_cpu(&q.qu_value, change); in gfs2_adjust_quota() 834 if (((s64)be64_to_cpu(q.qu_value)) < 0) in gfs2_adjust_quota() 835 q.qu_value = 0; /* Never go negative on quota usage */ in gfs2_adjust_quota() 836 qd->qd_qb.qb_value = q.qu_value; in gfs2_adjust_quota() 839 q.qu_warn = cpu_to_be64(fdq->d_spc_softlimit >> sdp->sd_sb.sb_bsize_shift); in gfs2_adjust_quota() 840 qd->qd_qb.qb_warn = q.qu_warn; in gfs2_adjust_quota() [all …]
|
/fs/ufs/ |
D | inode.c | 131 Indirect chain[4], *q = chain; in ufs_frag_map() local 149 if (!grow_chain32(ufsi, NULL, &ufsi->i_u1.i_data[*p++], chain, q)) in ufs_frag_map() 151 if (!q->key32) in ufs_frag_map() 159 fs32_to_cpu(sb, q->key32) + (n>>shift)); in ufs_frag_map() 163 if (!grow_chain32(ufsi, bh, ptr, chain, ++q)) in ufs_frag_map() 165 if (!q->key32) in ufs_frag_map() 168 res = fs32_to_cpu(sb, q->key32); in ufs_frag_map() 172 if (!grow_chain64(ufsi, NULL, &ufsi->i_u1.u2_i_data[*p++], chain, q)) in ufs_frag_map() 174 if (!q->key64) in ufs_frag_map() 183 fs64_to_cpu(sb, q->key64) + (n>>shift)); in ufs_frag_map() [all …]
|
/fs/jfs/ |
D | ioctl.c | 125 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl() local 132 if (!blk_queue_discard(q)) { in jfs_ioctl() 142 q->limits.discard_granularity); in jfs_ioctl()
|
/fs/sysv/ |
D | itree.c | 269 static inline int all_zeroes(sysv_zone_t *p, sysv_zone_t *q) in all_zeroes() argument 271 while (p < q) in all_zeroes() 326 static inline void free_data(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q) in free_data() argument 328 for ( ; p < q ; p++) { in free_data() 338 static void free_branches(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q, int depth) in free_branches() argument 344 for ( ; p < q ; p++) { in free_branches() 361 free_data(inode, p, q); in free_branches()
|
/fs/crypto/ |
D | inline_crypt.c | 145 struct request_queue *q; in fscrypt_derive_raw_secret() local 147 q = sb->s_bdev->bd_queue; in fscrypt_derive_raw_secret() 148 if (!q->ksm) in fscrypt_derive_raw_secret() 151 return keyslot_manager_derive_raw_secret(q->ksm, in fscrypt_derive_raw_secret()
|
/fs/nfsd/ |
D | blocklayout.c | 217 struct request_queue *q = bdev->bd_disk->queue; in nfsd4_scsi_identify_device() local 231 if (WARN_ON_ONCE(!blk_queue_scsi_passthrough(q))) in nfsd4_scsi_identify_device() 239 rq = blk_get_request(q, REQ_OP_SCSI_IN, 0); in nfsd4_scsi_identify_device() 246 error = blk_rq_map_kern(q, rq, buf, bufflen, GFP_KERNEL); in nfsd4_scsi_identify_device() 257 blk_execute_rq(rq->q, NULL, rq, 1); in nfsd4_scsi_identify_device()
|
/fs/erofs/ |
D | zdata.c | 1151 struct z_erofs_unzip_io *q[], in jobqueueset_init() argument 1159 q[JQ_BYPASS] = jobqueue_init(sb, fgq + JQ_BYPASS, true); in jobqueueset_init() 1160 qtail[JQ_BYPASS] = &q[JQ_BYPASS]->head; in jobqueueset_init() 1162 q[JQ_SUBMIT] = jobqueue_init(sb, fgq + JQ_SUBMIT, forcefg); in jobqueueset_init() 1163 qtail[JQ_SUBMIT] = &q[JQ_SUBMIT]->head; in jobqueueset_init() 1165 return tagptr_cast_ptr(tagptr_fold(tagptr1_t, q[JQ_SUBMIT], !forcefg)); in jobqueueset_init() 1187 static bool postsubmit_is_all_bypassed(struct z_erofs_unzip_io *q[], in postsubmit_is_all_bypassed() argument 1198 kvfree(container_of(q[JQ_SUBMIT], struct z_erofs_unzip_io_sb, io)); in postsubmit_is_all_bypassed() 1210 struct z_erofs_unzip_io *q[NR_JOBQUEUES]; in z_erofs_vle_submit_all() local 1224 bi_private = jobqueueset_init(sb, qtail, q, fgq, force_fg); in z_erofs_vle_submit_all() [all …]
|
/fs/cifs/ |
D | dir.c | 808 static int cifs_ci_hash(const struct dentry *dentry, struct qstr *q) in cifs_ci_hash() argument 816 for (i = 0; i < q->len; i += charlen) { in cifs_ci_hash() 817 charlen = codepage->char2uni(&q->name[i], q->len - i, &c); in cifs_ci_hash() 823 q->hash = end_name_hash(hash); in cifs_ci_hash()
|
/fs/autofs/ |
D | expire.c | 101 struct dentry *q; in get_next_positive_subdir() local 105 q = positive_after(root, prev); in get_next_positive_subdir() 109 return q; in get_next_positive_subdir()
|