Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 51) sorted by relevance

123

/fs/jffs2/
Dcompr_rubin.c39 unsigned long q; member
92 rs->q = 0; in init_rubin()
108 while ((rs->q >= UPPER_BIT_RUBIN) || in encode()
109 ((rs->p + rs->q) <= UPPER_BIT_RUBIN)) { in encode()
112 ret = pushbit(&rs->pp, (rs->q & UPPER_BIT_RUBIN) ? 1 : 0, 0); in encode()
115 rs->q &= LOWER_BITS_RUBIN; in encode()
116 rs->q <<= 1; in encode()
132 rs->q += i0; in encode()
144 pushbit(&rs->pp, (UPPER_BIT_RUBIN & rs->q) ? 1 : 0, 1); in end_rubin()
145 rs->q &= LOWER_BITS_RUBIN; in end_rubin()
[all …]
/fs/orangefs/
Dorangefs-bufmap.c13 wait_queue_head_t q; member
20 .q = __WAIT_QUEUE_HEAD_INITIALIZER(rw_map.q)
24 .q = __WAIT_QUEUE_HEAD_INITIALIZER(readdir_map.q)
30 spin_lock(&m->q.lock); in install()
33 wake_up_all_locked(&m->q); in install()
34 spin_unlock(&m->q.lock); in install()
39 spin_lock(&m->q.lock); in mark_killed()
41 spin_unlock(&m->q.lock); in mark_killed()
47 spin_lock(&m->q.lock); in run_down()
51 __add_wait_queue_entry_tail(&m->q, &wait); in run_down()
[all …]
/fs/xfs/
Dxfs_qm_syscalls.c39 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_scall_quotaoff() local
62 ASSERT(q); in xfs_qm_scall_quotaoff()
63 mutex_lock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff()
74 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff()
179 mutex_unlock(&q->qi_quotaofflock); in xfs_qm_scall_quotaoff()
187 if ((dqtype & XFS_QMOPT_UQUOTA) && q->qi_uquotaip) { in xfs_qm_scall_quotaoff()
188 xfs_irele(q->qi_uquotaip); in xfs_qm_scall_quotaoff()
189 q->qi_uquotaip = NULL; in xfs_qm_scall_quotaoff()
191 if ((dqtype & XFS_QMOPT_GQUOTA) && q->qi_gquotaip) { in xfs_qm_scall_quotaoff()
192 xfs_irele(q->qi_gquotaip); in xfs_qm_scall_quotaoff()
[all …]
Dxfs_quotaops.c26 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_fill_state() local
40 tstate->spc_timelimit = q->qi_btimelimit; in xfs_qm_fill_state()
41 tstate->ino_timelimit = q->qi_itimelimit; in xfs_qm_fill_state()
42 tstate->rt_spc_timelimit = q->qi_rtbtimelimit; in xfs_qm_fill_state()
43 tstate->spc_warnlimit = q->qi_bwarnlimit; in xfs_qm_fill_state()
44 tstate->ino_warnlimit = q->qi_iwarnlimit; in xfs_qm_fill_state()
45 tstate->rt_spc_warnlimit = q->qi_rtbwarnlimit; in xfs_qm_fill_state()
60 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_fs_get_quota_state() local
65 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state()
79 xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, q->qi_uquotaip, in xfs_fs_get_quota_state()
[all …]
Dxfs_trans_dquot.c282 struct xfs_dqtrx *q) in xfs_trans_dqlockedjoin() argument
284 ASSERT(q[0].qt_dquot != NULL); in xfs_trans_dqlockedjoin()
285 if (q[1].qt_dquot == NULL) { in xfs_trans_dqlockedjoin()
286 xfs_dqlock(q[0].qt_dquot); in xfs_trans_dqlockedjoin()
287 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin()
290 xfs_dqlock2(q[0].qt_dquot, q[1].qt_dquot); in xfs_trans_dqlockedjoin()
291 xfs_trans_dqjoin(tp, q[0].qt_dquot); in xfs_trans_dqlockedjoin()
292 xfs_trans_dqjoin(tp, q[1].qt_dquot); in xfs_trans_dqlockedjoin()
588 xfs_quotainfo_t *q = mp->m_quotainfo; in xfs_trans_dqresv() local
594 defq = xfs_get_defquota(dqp, q); in xfs_trans_dqresv()
[all …]
Dxfs_discard.c150 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim() local
151 unsigned int granularity = q->limits.discard_granularity; in xfs_ioc_trim()
160 if (!blk_queue_discard(q)) in xfs_ioc_trim()
Dxfs_dquot.c72 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_adjust_dqlimits() local
78 defq = xfs_get_defquota(dq, q); in xfs_qm_adjust_dqlimits()
214 struct xfs_quotainfo *q = mp->m_quotainfo; in xfs_qm_init_dquot_blk() local
227 curid = id - (id % q->qi_dqperchunk); in xfs_qm_init_dquot_blk()
228 memset(d, 0, BBTOB(q->qi_dqchunklen)); in xfs_qm_init_dquot_blk()
229 for (i = 0; i < q->qi_dqperchunk; i++, d++, curid++) { in xfs_qm_init_dquot_blk()
245 xfs_trans_log_buf(tp, bp, 0, BBTOB(q->qi_dqchunklen) - 1); in xfs_qm_init_dquot_blk()
/fs/hpfs/
Dalloc.c122 unsigned i, q; in alloc_in_bmp() local
138 q = nr + n; b = 0; in alloc_in_bmp()
139 while ((a = tstbits(bmp, q, n + forward)) != 0) { in alloc_in_bmp()
140 q += a; in alloc_in_bmp()
141 if (n != 1) q = ((q-1)&~(n-1))+n; in alloc_in_bmp()
143 if (q>>5 != nr>>5) { in alloc_in_bmp()
145 q = nr & 0x1f; in alloc_in_bmp()
147 } else if (q > nr) break; in alloc_in_bmp()
150 ret = bs + q; in alloc_in_bmp()
159 q = i<<5; in alloc_in_bmp()
[all …]
Dea.c289 secno q = hpfs_alloc_sector(s, fno, 1, 0); in hpfs_set_ea() local
290 if (!q) goto bail; in hpfs_set_ea()
291 fnode->ea_secno = cpu_to_le32(q); in hpfs_set_ea()
/fs/afs/
Daddr_list.c136 const char *q, *stop; in afs_parse_text_addrs() local
148 q = memchr(p, ']', end - p); in afs_parse_text_addrs()
150 for (q = p; q < end; q++) in afs_parse_text_addrs()
151 if (*q == '+' || *q == delim) in afs_parse_text_addrs()
155 if (in4_pton(p, q - p, (u8 *)&x[0], -1, &stop)) { in afs_parse_text_addrs()
157 } else if (in6_pton(p, q - p, (u8 *)x, -1, &stop)) { in afs_parse_text_addrs()
164 p = q; in afs_parse_text_addrs()
170 if (q < end && *q == ']') in afs_parse_text_addrs()
/fs/sdcardfs/
Dpackagelist.c57 static inline void qstr_init(struct qstr *q, const char *name) in qstr_init() argument
59 q->name = name; in qstr_init()
60 q->len = strlen(q->name); in qstr_init()
61 q->hash = full_name_case_hash(0, q->name, q->len); in qstr_init()
92 struct qstr q; in get_appid() local
94 qstr_init(&q, key); in get_appid()
95 return __get_appid(&q); in get_appid()
118 struct qstr q; in get_ext_gid() local
120 qstr_init(&q, key); in get_ext_gid()
121 return __get_ext_gid(&q); in get_ext_gid()
[all …]
/fs/
Dfs_pin.c88 struct hlist_node *q; in group_pin_kill() local
90 q = READ_ONCE(p->first); in group_pin_kill()
91 if (!q) { in group_pin_kill()
95 pin_kill(hlist_entry(q, struct fs_pin, s_list)); in group_pin_kill()
Dfs_context.c394 char *q; in logfc() local
407 q = kvasprintf(GFP_KERNEL, fmt, va); in logfc()
409 if (!q) in logfc()
420 q = kstrdup(p, GFP_KERNEL); in logfc()
426 q = (char *)p; in logfc()
432 printk(KERN_WARNING "%s\n", q + 2); in logfc()
435 printk(KERN_ERR "%s\n", q + 2); in logfc()
438 printk(KERN_NOTICE "%s\n", q + 2); in logfc()
442 kfree(q); in logfc()
457 log->buffer[index] = q; in logfc()
Dnamespace.c1774 struct mount *res, *p, *q, *r, *parent; in copy_tree() local
1782 res = q = clone_mnt(mnt, dentry, flag); in copy_tree()
1783 if (IS_ERR(q)) in copy_tree()
1784 return q; in copy_tree()
1786 q->mnt_mountpoint = mnt->mnt_mountpoint; in copy_tree()
1799 q = ERR_PTR(-EPERM); in copy_tree()
1813 q = q->mnt_parent; in copy_tree()
1816 parent = q; in copy_tree()
1817 q = clone_mnt(p, p->mnt.mnt_root, flag); in copy_tree()
1818 if (IS_ERR(q)) in copy_tree()
[all …]
/fs/efivarfs/
Dsuper.c89 struct qstr q; in efivarfs_alloc_dentry() local
92 q.name = name; in efivarfs_alloc_dentry()
93 q.len = strlen(name); in efivarfs_alloc_dentry()
95 err = efivarfs_d_hash(parent, &q); in efivarfs_alloc_dentry()
99 d = d_alloc(parent, &q); in efivarfs_alloc_dentry()
/fs/minix/
Ditree_common.c209 static inline int all_zeroes(block_t *p, block_t *q) in all_zeroes() argument
211 while (p < q) in all_zeroes()
257 static inline void free_data(struct inode *inode, block_t *p, block_t *q) in free_data() argument
261 for ( ; p < q ; p++) { in free_data()
270 static void free_branches(struct inode *inode, block_t *p, block_t *q, int depth) in free_branches() argument
276 for ( ; p < q ; p++) { in free_branches()
291 free_data(inode, p, q); in free_branches()
/fs/gfs2/
Dquota.c816 struct gfs2_quota q; in gfs2_adjust_quota() local
826 memset(&q, 0, sizeof(struct gfs2_quota)); in gfs2_adjust_quota()
827 err = gfs2_internal_read(ip, (char *)&q, &loc, sizeof(q)); in gfs2_adjust_quota()
831 loc -= sizeof(q); /* gfs2_internal_read would've advanced the loc ptr */ in gfs2_adjust_quota()
833 be64_add_cpu(&q.qu_value, change); in gfs2_adjust_quota()
834 if (((s64)be64_to_cpu(q.qu_value)) < 0) in gfs2_adjust_quota()
835 q.qu_value = 0; /* Never go negative on quota usage */ in gfs2_adjust_quota()
836 qd->qd_qb.qb_value = q.qu_value; in gfs2_adjust_quota()
839 q.qu_warn = cpu_to_be64(fdq->d_spc_softlimit >> sdp->sd_sb.sb_bsize_shift); in gfs2_adjust_quota()
840 qd->qd_qb.qb_warn = q.qu_warn; in gfs2_adjust_quota()
[all …]
/fs/ufs/
Dinode.c131 Indirect chain[4], *q = chain; in ufs_frag_map() local
149 if (!grow_chain32(ufsi, NULL, &ufsi->i_u1.i_data[*p++], chain, q)) in ufs_frag_map()
151 if (!q->key32) in ufs_frag_map()
159 fs32_to_cpu(sb, q->key32) + (n>>shift)); in ufs_frag_map()
163 if (!grow_chain32(ufsi, bh, ptr, chain, ++q)) in ufs_frag_map()
165 if (!q->key32) in ufs_frag_map()
168 res = fs32_to_cpu(sb, q->key32); in ufs_frag_map()
172 if (!grow_chain64(ufsi, NULL, &ufsi->i_u1.u2_i_data[*p++], chain, q)) in ufs_frag_map()
174 if (!q->key64) in ufs_frag_map()
183 fs64_to_cpu(sb, q->key64) + (n>>shift)); in ufs_frag_map()
[all …]
/fs/jfs/
Dioctl.c125 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl() local
132 if (!blk_queue_discard(q)) { in jfs_ioctl()
142 q->limits.discard_granularity); in jfs_ioctl()
/fs/sysv/
Ditree.c269 static inline int all_zeroes(sysv_zone_t *p, sysv_zone_t *q) in all_zeroes() argument
271 while (p < q) in all_zeroes()
326 static inline void free_data(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q) in free_data() argument
328 for ( ; p < q ; p++) { in free_data()
338 static void free_branches(struct inode *inode, sysv_zone_t *p, sysv_zone_t *q, int depth) in free_branches() argument
344 for ( ; p < q ; p++) { in free_branches()
361 free_data(inode, p, q); in free_branches()
/fs/crypto/
Dinline_crypt.c145 struct request_queue *q; in fscrypt_derive_raw_secret() local
147 q = sb->s_bdev->bd_queue; in fscrypt_derive_raw_secret()
148 if (!q->ksm) in fscrypt_derive_raw_secret()
151 return keyslot_manager_derive_raw_secret(q->ksm, in fscrypt_derive_raw_secret()
/fs/nfsd/
Dblocklayout.c217 struct request_queue *q = bdev->bd_disk->queue; in nfsd4_scsi_identify_device() local
231 if (WARN_ON_ONCE(!blk_queue_scsi_passthrough(q))) in nfsd4_scsi_identify_device()
239 rq = blk_get_request(q, REQ_OP_SCSI_IN, 0); in nfsd4_scsi_identify_device()
246 error = blk_rq_map_kern(q, rq, buf, bufflen, GFP_KERNEL); in nfsd4_scsi_identify_device()
257 blk_execute_rq(rq->q, NULL, rq, 1); in nfsd4_scsi_identify_device()
/fs/erofs/
Dzdata.c1151 struct z_erofs_unzip_io *q[], in jobqueueset_init() argument
1159 q[JQ_BYPASS] = jobqueue_init(sb, fgq + JQ_BYPASS, true); in jobqueueset_init()
1160 qtail[JQ_BYPASS] = &q[JQ_BYPASS]->head; in jobqueueset_init()
1162 q[JQ_SUBMIT] = jobqueue_init(sb, fgq + JQ_SUBMIT, forcefg); in jobqueueset_init()
1163 qtail[JQ_SUBMIT] = &q[JQ_SUBMIT]->head; in jobqueueset_init()
1165 return tagptr_cast_ptr(tagptr_fold(tagptr1_t, q[JQ_SUBMIT], !forcefg)); in jobqueueset_init()
1187 static bool postsubmit_is_all_bypassed(struct z_erofs_unzip_io *q[], in postsubmit_is_all_bypassed() argument
1198 kvfree(container_of(q[JQ_SUBMIT], struct z_erofs_unzip_io_sb, io)); in postsubmit_is_all_bypassed()
1210 struct z_erofs_unzip_io *q[NR_JOBQUEUES]; in z_erofs_vle_submit_all() local
1224 bi_private = jobqueueset_init(sb, qtail, q, fgq, force_fg); in z_erofs_vle_submit_all()
[all …]
/fs/cifs/
Ddir.c808 static int cifs_ci_hash(const struct dentry *dentry, struct qstr *q) in cifs_ci_hash() argument
816 for (i = 0; i < q->len; i += charlen) { in cifs_ci_hash()
817 charlen = codepage->char2uni(&q->name[i], q->len - i, &c); in cifs_ci_hash()
823 q->hash = end_name_hash(hash); in cifs_ci_hash()
/fs/autofs/
Dexpire.c101 struct dentry *q; in get_next_positive_subdir() local
105 q = positive_after(root, prev); in get_next_positive_subdir()
109 return q; in get_next_positive_subdir()

123