/fs/affs/ |
D | inode.c | 22 struct affs_tail *tail; in affs_iget() local 50 tail = AFFS_TAIL(sb, bh); in affs_iget() 51 prot = be32_to_cpu(tail->protect); in affs_iget() 76 id = be16_to_cpu(tail->uid); in affs_iget() 84 id = be16_to_cpu(tail->gid); in affs_iget() 92 switch (be32_to_cpu(tail->stype)) { in affs_iget() 98 if (be32_to_cpu(tail->stype) == ST_USERDIR || in affs_iget() 127 size = be32_to_cpu(tail->size); in affs_iget() 136 if (tail->link_chain) in affs_iget() 153 = (be32_to_cpu(tail->change.days) * 86400LL + in affs_iget() [all …]
|
D | super.c | 37 struct affs_root_tail *tail = AFFS_ROOT_TAIL(sb, bh); in affs_commit_super() local 40 affs_secs_to_datestamp(ktime_get_real_seconds(), &tail->disk_change); in affs_commit_super()
|
/fs/9p/ |
D | vfs_dir.c | 41 int tail; member 108 if (rdir->tail == rdir->head) { in v9fs_dir_readdir() 120 rdir->tail = n; in v9fs_dir_readdir() 122 while (rdir->head < rdir->tail) { in v9fs_dir_readdir() 124 rdir->tail - rdir->head, &st); in v9fs_dir_readdir() 166 if (rdir->tail == rdir->head) { in v9fs_dir_readdir_dotl() 173 rdir->tail = err; in v9fs_dir_readdir_dotl() 176 while (rdir->head < rdir->tail) { in v9fs_dir_readdir_dotl() 179 rdir->tail - rdir->head, in v9fs_dir_readdir_dotl()
|
/fs/reiserfs/ |
D | tail_conversion.c | 214 char *tail; in indirect2direct() local 245 tail = (char *)kmap(page); /* this can schedule */ in indirect2direct() 275 tail = tail + (pos & (PAGE_SIZE - 1)); in indirect2direct() 284 tail ? tail : NULL) < 0) { in indirect2direct()
|
/fs/ |
D | aio.c | 63 unsigned tail; member 160 unsigned tail; member 546 ring->head = ring->tail = 0; in aio_setup_ring() 957 unsigned tail) in refill_reqs_available() argument 963 if (head <= tail) in refill_reqs_available() 964 events_in_ring = tail - head; in refill_reqs_available() 966 events_in_ring = ctx->nr_events - (head - tail); in refill_reqs_available() 1005 refill_reqs_available(ctx, head, ctx->tail); in user_refill_reqs_available() 1093 unsigned tail, pos, head; in aio_complete() local 1103 tail = ctx->tail; in aio_complete() [all …]
|
D | pipe.c | 1111 unsigned int tail; in pipe_set_size() local 1114 tail = pipe->curbuf + pipe->nrbufs; in pipe_set_size() 1115 if (tail < pipe->buffers) in pipe_set_size() 1116 tail = 0; in pipe_set_size() 1118 tail &= (pipe->buffers - 1); in pipe_set_size() 1120 head = pipe->nrbufs - tail; in pipe_set_size() 1123 if (tail) in pipe_set_size() 1124 memcpy(bufs + head, pipe->bufs, tail * sizeof(struct pipe_buffer)); in pipe_set_size()
|
D | fsopen.c | 39 if (log->head == log->tail) { in fscontext_read() 44 index = log->tail & (logsize - 1); in fscontext_read() 49 log->tail++; in fscontext_read()
|
D | direct-io.c | 112 unsigned tail; /* last valid page + 1 */ member 162 return sdio->tail - sdio->head; in dio_pages_present() 187 sdio->tail = 1; in dio_refill_pages() 197 sdio->tail = (ret + PAGE_SIZE - 1) / PAGE_SIZE; in dio_refill_pages() 498 while (sdio->head < sdio->tail) in dio_cleanup() 984 to = (sdio->head == sdio->tail - 1) ? sdio->to : PAGE_SIZE; in do_direct_IO()
|
D | io_uring.c | 84 u32 tail ____cacheline_aligned_in_smp; 476 if (ctx->cached_cq_tail != READ_ONCE(rings->cq.tail)) { in __io_commit_cqring() 478 smp_store_release(&rings->cq.tail, ctx->cached_cq_tail); in __io_commit_cqring() 550 unsigned tail; in io_get_cqring() local 552 tail = ctx->cached_cq_tail; in io_get_cqring() 558 if (tail - READ_ONCE(rings->cq.head) == rings->cq_ring_entries) in io_get_cqring() 562 return &rings->cqes[tail & ctx->cq_mask]; in io_get_cqring() 745 return READ_ONCE(rings->cq.tail) - READ_ONCE(rings->cq.head); in io_cqring_events() 753 return smp_load_acquire(&rings->sq.tail) - ctx->cached_sq_head; in io_sqring_entries() 2607 if (head == smp_load_acquire(&rings->sq.tail)) in io_get_sqring() [all …]
|
D | fs_context.c | 449 sizeof(log->tail) != sizeof(u8)); in logfc() 450 if ((u8)(log->head - log->tail) == logsize) { in logfc() 454 log->tail++; in logfc()
|
D | buffer.c | 868 struct buffer_head *bh, *tail; in link_dev_buffers() local 872 tail = bh; in link_dev_buffers() 875 tail->b_this_page = head; in link_dev_buffers() 1517 struct buffer_head *bh, *head, *tail; in create_empty_buffers() local 1523 tail = bh; in create_empty_buffers() 1526 tail->b_this_page = head; in create_empty_buffers()
|
D | fs-writeback.c | 1178 struct inode *tail; in redirty_tail() local 1180 tail = wb_inode(wb->b_dirty.next); in redirty_tail() 1181 if (time_before(inode->dirtied_when, tail->dirtied_when)) in redirty_tail()
|
/fs/gfs2/ |
D | log.c | 484 unsigned int tail; in current_tail() local 489 tail = sdp->sd_log_head; in current_tail() 493 tail = tr->tr_first; in current_tail() 498 return tail; in current_tail() 688 u64 seq, u32 tail, u32 lblock, u32 flags, in gfs2_write_log_header() argument 709 lh->lh_tail = cpu_to_be32(tail); in gfs2_write_log_header() 764 unsigned int tail; in log_write_header() local 769 tail = current_tail(sdp); in log_write_header() 776 sdp->sd_log_idle = (tail == sdp->sd_log_flush_head); in log_write_header() 777 gfs2_write_log_header(sdp, sdp->sd_jdesc, sdp->sd_log_sequence++, tail, in log_write_header() [all …]
|
D | log.h | 70 u64 seq, u32 tail, u32 lblock, u32 flags,
|
/fs/nfsd/ |
D | nfs3xdr.c | 403 struct kvec *tail = rqstp->rq_arg.tail; in nfs3svc_decode_writeargs() local 426 dlen = head->iov_len + rqstp->rq_arg.page_len + tail->iov_len - hdr; in nfs3svc_decode_writeargs() 505 rqstp->rq_arg.tail[0].iov_len; in nfs3svc_decode_symlinkargs() 715 rqstp->rq_res.tail[0].iov_base = p; in nfs3svc_encode_readlinkres() 717 rqstp->rq_res.tail[0].iov_len = 4 - (resp->len&3); in nfs3svc_encode_readlinkres() 740 rqstp->rq_res.tail[0].iov_base = p; in nfs3svc_encode_readres() 742 rqstp->rq_res.tail[0].iov_len = 4 - (resp->count & 3); in nfs3svc_encode_readres() 823 rqstp->rq_res.tail[0].iov_base = p; in nfs3svc_encode_readdirres() 826 rqstp->rq_res.tail[0].iov_len = 2<<2; in nfs3svc_encode_readdirres()
|
D | nfsxdr.c | 395 p = rqstp->rq_arg.tail[0].iov_base; in nfssvc_decode_symlinkargs() 461 rqstp->rq_res.tail[0].iov_base = p; in nfssvc_encode_readlinkres() 463 rqstp->rq_res.tail[0].iov_len = 4 - (resp->len&3); in nfssvc_encode_readlinkres() 481 rqstp->rq_res.tail[0].iov_base = p; in nfssvc_encode_readres() 483 rqstp->rq_res.tail[0].iov_len = 4 - (resp->count&3); in nfssvc_encode_readres()
|
D | nfs4xdr.c | 164 struct kvec *vec = &argp->rqstp->rq_arg.tail[0]; in read_buf() 166 if (!argp->tail) { in read_buf() 167 argp->tail = true; in read_buf() 3496 buf->tail[0].iov_base = xdr->p; in nfsd4_encode_splice_read() 3497 buf->tail[0].iov_len = 0; in nfsd4_encode_splice_read() 3498 xdr->iov = buf->tail; in nfsd4_encode_splice_read() 3504 buf->tail[0].iov_base += maxcount&3; in nfsd4_encode_splice_read() 3505 buf->tail[0].iov_len = pad; in nfsd4_encode_splice_read() 4565 args->tail = false; in nfs4svc_decode_compoundargs() 4584 buf->tail[0].iov_len); in nfs4svc_encode_compoundres()
|
/fs/jbd2/ |
D | recovery.c | 176 struct jbd2_journal_block_tail *tail; in jbd2_descriptor_block_csum_verify() local 183 tail = (struct jbd2_journal_block_tail *)(buf + j->j_blocksize - in jbd2_descriptor_block_csum_verify() 185 provided = tail->t_checksum; in jbd2_descriptor_block_csum_verify() 186 tail->t_checksum = 0; in jbd2_descriptor_block_csum_verify() 188 tail->t_checksum = provided; in jbd2_descriptor_block_csum_verify()
|
D | journal.c | 857 struct jbd2_journal_block_tail *tail; in jbd2_descriptor_block_csum_set() local 863 tail = (struct jbd2_journal_block_tail *)(bh->b_data + j->j_blocksize - in jbd2_descriptor_block_csum_set() 865 tail->t_checksum = 0; in jbd2_descriptor_block_csum_set() 867 tail->t_checksum = cpu_to_be32(csum); in jbd2_descriptor_block_csum_set()
|
/fs/jfs/ |
D | jfs_logmgr.c | 2025 struct lbuf *tail; in lbmWrite() local 2047 tail = log->wqueue; in lbmWrite() 2052 if (tail == NULL) { in lbmWrite() 2057 bp->l_wqnext = tail->l_wqnext; in lbmWrite() 2058 tail->l_wqnext = bp; in lbmWrite() 2061 tail = bp; in lbmWrite() 2065 if ((bp != tail->l_wqnext) || !(flag & lbmWRITE)) { in lbmWrite() 2182 struct lbuf *nextbp, *tail; in lbmIODone() local 2242 tail = log->wqueue; in lbmIODone() 2245 if (bp == tail) { in lbmIODone() [all …]
|
/fs/ntfs/ |
D | runlist.c | 373 int tail; /* Start of tail of @dst. */ in ntfs_rl_replace() local 412 tail = loc + right + 1; in ntfs_rl_replace() 423 ntfs_rl_mm(dst, marker, tail, dsize - tail); in ntfs_rl_replace() 427 if (dsize - tail > 0 && dst[marker].lcn == LCN_ENOENT) in ntfs_rl_replace()
|
D | aops.c | 1730 struct buffer_head *tail; in mark_ntfs_record_dirty() local 1734 tail = bh; in mark_ntfs_record_dirty() 1737 tail->b_this_page = head; in mark_ntfs_record_dirty()
|
D | mft.c | 498 struct buffer_head *tail; in ntfs_sync_mft_mirror() local 503 tail = bh; in ntfs_sync_mft_mirror() 506 tail->b_this_page = head; in ntfs_sync_mft_mirror()
|
/fs/btrfs/ |
D | volumes.c | 505 struct bio *head, struct bio *tail) in requeue_list() argument 512 if (pending_bios->tail) in requeue_list() 513 tail->bi_next = old_head; in requeue_list() 515 pending_bios->tail = tail; in requeue_list() 535 struct bio *tail; in run_scheduled_bios() local 575 tail = pending_bios->tail; in run_scheduled_bios() 576 WARN_ON(pending && !tail); in run_scheduled_bios() 596 pending_bios->tail = NULL; in run_scheduled_bios() 612 requeue_list(pending_bios, pending, tail); in run_scheduled_bios() 679 requeue_list(pending_bios, pending, tail); in run_scheduled_bios() [all …]
|
D | volumes.h | 23 struct bio *tail; member
|