Home
last modified time | relevance | path

Searched refs:ap (Results 1 – 24 of 24) sorted by relevance

/fs/fuse/
Dioctl.c186 struct fuse_args_pages ap = {}; in fuse_do_ioctl() local
204 ap.pages = fuse_pages_alloc(fm->fc->max_pages, GFP_KERNEL, &ap.descs); in fuse_do_ioctl()
206 if (!ap.pages || !iov_page) in fuse_do_ioctl()
209 fuse_page_descs_length_init(ap.descs, 0, fm->fc->max_pages); in fuse_do_ioctl()
247 while (ap.num_pages < max_pages) { in fuse_do_ioctl()
248 ap.pages[ap.num_pages] = alloc_page(GFP_KERNEL | __GFP_HIGHMEM); in fuse_do_ioctl()
249 if (!ap.pages[ap.num_pages]) in fuse_do_ioctl()
251 ap.num_pages++; in fuse_do_ioctl()
256 ap.args.opcode = FUSE_IOCTL; in fuse_do_ioctl()
257 ap.args.nodeid = ff->nodeid; in fuse_do_ioctl()
[all …]
Dfile.c449 if (idx_from >= curr_index + wpa->ia.ap.num_pages) in fuse_find_writeback()
664 struct fuse_args *args = &ia->ap.args; in fuse_read_args_fill()
680 static void fuse_release_user_pages(struct fuse_args_pages *ap, in fuse_release_user_pages() argument
685 for (i = 0; i < ap->num_pages; i++) { in fuse_release_user_pages()
687 set_page_dirty_lock(ap->pages[i]); in fuse_release_user_pages()
688 put_page(ap->pages[i]); in fuse_release_user_pages()
766 ia->ap.pages = fuse_pages_alloc(npages, GFP_KERNEL, in fuse_io_alloc()
767 &ia->ap.descs); in fuse_io_alloc()
768 if (!ia->ap.pages) { in fuse_io_alloc()
778 kfree(ia->ap.pages); in fuse_io_free()
[all …]
Dcuse.c303 struct fuse_args_pages ap; member
321 struct cuse_init_args *ia = container_of(args, typeof(*ia), ap.args); in cuse_process_init_reply()
322 struct fuse_args_pages *ap = &ia->ap; in cuse_process_init_reply() local
325 struct page *page = ap->pages[0]; in cuse_process_init_reply()
342 rc = cuse_parse_devinfo(page_address(page), ap->args.out_args[1].size, in cuse_process_init_reply()
431 struct fuse_args_pages *ap; in cuse_send_init() local
444 ap = &ia->ap; in cuse_send_init()
448 ap->args.opcode = CUSE_INIT; in cuse_send_init()
449 ap->args.in_numargs = 1; in cuse_send_init()
450 ap->args.in_args[0].size = sizeof(ia->in); in cuse_send_init()
[all …]
Dvirtio_fs.c563 struct fuse_args_pages *ap; in virtio_fs_request_complete() local
576 ap = container_of(args, typeof(*ap), args); in virtio_fs_request_complete()
577 for (i = 0; i < ap->num_pages; i++) { in virtio_fs_request_complete()
578 thislen = ap->descs[i].length; in virtio_fs_request_complete()
580 WARN_ON(ap->descs[i].offset); in virtio_fs_request_complete()
581 page = ap->pages[i]; in virtio_fs_request_complete()
1042 struct fuse_args_pages *ap = container_of(args, typeof(*ap), args); in sg_count_fuse_req() local
1050 total_sgs += sg_count_fuse_pages(ap->descs, ap->num_pages, in sg_count_fuse_req()
1064 total_sgs += sg_count_fuse_pages(ap->descs, ap->num_pages, in sg_count_fuse_req()
1100 struct fuse_args_pages *ap = container_of(req->args, typeof(*ap), args); in sg_init_fuse_args() local
[all …]
Ddev.c988 struct fuse_args_pages *ap = container_of(req->args, typeof(*ap), args); in fuse_copy_pages() local
991 for (i = 0; i < ap->num_pages && (nbytes || zeroing); i++) { in fuse_copy_pages()
993 unsigned int offset = ap->descs[i].offset; in fuse_copy_pages()
994 unsigned int count = min(nbytes, ap->descs[i].length); in fuse_copy_pages()
996 err = fuse_copy_page(cs, &ap->pages[i], offset, count, zeroing); in fuse_copy_pages()
1654 struct fuse_args_pages ap; member
1662 container_of(args, typeof(*ra), ap.args); in fuse_retrieve_end()
1664 release_pages(ra->ap.pages, ra->ap.num_pages); in fuse_retrieve_end()
1682 struct fuse_args_pages *ap; in fuse_retrieve() local
1697 args_size += num_pages * (sizeof(ap->pages[0]) + sizeof(ap->descs[0])); in fuse_retrieve()
[all …]
Dreaddir.c340 struct fuse_args_pages *ap = &ia.ap; in fuse_readdir_uncached() local
350 ap->args.out_pages = true; in fuse_readdir_uncached()
351 ap->num_pages = 1; in fuse_readdir_uncached()
352 ap->pages = &page; in fuse_readdir_uncached()
353 ap->descs = &desc; in fuse_readdir_uncached()
363 res = fuse_simple_request(fm, &ap->args); in fuse_readdir_uncached()
Ddir.c1715 struct fuse_args_pages ap = { in fuse_readlink_page() local
1723 ap.args.opcode = FUSE_READLINK; in fuse_readlink_page()
1724 ap.args.nodeid = get_node_id(inode); in fuse_readlink_page()
1725 ap.args.out_pages = true; in fuse_readlink_page()
1726 ap.args.out_argvar = true; in fuse_readlink_page()
1727 ap.args.page_zeroing = true; in fuse_readlink_page()
1728 ap.args.out_numargs = 1; in fuse_readlink_page()
1729 ap.args.out_args[0].size = desc.length; in fuse_readlink_page()
1730 res = fuse_simple_request(fm, &ap.args); in fuse_readlink_page()
Dfuse_i.h1113 struct fuse_args_pages ap; member
/fs/xfs/libxfs/
Dxfs_bmap.c3025 struct xfs_bmalloca *ap) /* bmap alloc argument struct */ in xfs_bmap_adjacent() argument
3040 mp = ap->ip->i_mount; in xfs_bmap_adjacent()
3041 nullfb = ap->tp->t_firstblock == NULLFSBLOCK; in xfs_bmap_adjacent()
3042 rt = XFS_IS_REALTIME_INODE(ap->ip) && in xfs_bmap_adjacent()
3043 (ap->datatype & XFS_ALLOC_USERDATA); in xfs_bmap_adjacent()
3045 ap->tp->t_firstblock); in xfs_bmap_adjacent()
3050 if (ap->eof && ap->prev.br_startoff != NULLFILEOFF && in xfs_bmap_adjacent()
3051 !isnullstartblock(ap->prev.br_startblock) && in xfs_bmap_adjacent()
3052 ISVALID(ap->prev.br_startblock + ap->prev.br_blockcount, in xfs_bmap_adjacent()
3053 ap->prev.br_startblock)) { in xfs_bmap_adjacent()
[all …]
/fs/xfs/
Dxfs_bmap_util.c74 struct xfs_bmalloca *ap) in xfs_bmap_rtalloc() argument
76 struct xfs_mount *mp = ap->ip->i_mount; in xfs_bmap_rtalloc()
77 xfs_fileoff_t orig_offset = ap->offset; in xfs_bmap_rtalloc()
83 xfs_extlen_t orig_length = ap->length; in xfs_bmap_rtalloc()
90 align = xfs_get_extsz_hint(ap->ip); in xfs_bmap_rtalloc()
93 error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev, in xfs_bmap_rtalloc()
94 align, 1, ap->eof, 0, in xfs_bmap_rtalloc()
95 ap->conv, &ap->offset, &ap->length); in xfs_bmap_rtalloc()
98 ASSERT(ap->length); in xfs_bmap_rtalloc()
99 ASSERT(ap->length % mp->m_sb.sb_rextsize == 0); in xfs_bmap_rtalloc()
[all …]
Dxfs_bmap_util.h20 int xfs_bmap_rtalloc(struct xfs_bmalloca *ap);
27 xfs_bmap_rtalloc(struct xfs_bmalloca *ap) in xfs_bmap_rtalloc() argument
50 void xfs_bmap_adjacent(struct xfs_bmalloca *ap);
Dxfs_filestream.c322 struct xfs_bmalloca *ap, in xfs_filestream_new_ag() argument
325 struct xfs_inode *ip = ap->ip, *pip; in xfs_filestream_new_ag()
327 xfs_extlen_t minlen = ap->length; in xfs_filestream_new_ag()
346 if (ap->datatype & XFS_ALLOC_USERDATA) in xfs_filestream_new_ag()
348 if (ap->tp->t_flags & XFS_TRANS_LOWMODE) in xfs_filestream_new_ag()
Dxfs_filestream.h17 int xfs_filestream_new_ag(struct xfs_bmalloca *ap, xfs_agnumber_t *agp);
Dxfs_buf.c2123 struct xfs_buf *ap = container_of(a, struct xfs_buf, b_list); in xfs_buf_cmp() local
2127 diff = ap->b_maps[0].bm_bn - bp->b_maps[0].bm_bn; in xfs_buf_cmp()
/fs/gfs2/
Dquota.h27 struct gfs2_alloc_parms *ap);
41 struct gfs2_alloc_parms *ap) in gfs2_quota_lock_check() argument
46 ap->allowed = UINT_MAX; /* Assume we are permitted a whole lot */ in gfs2_quota_lock_check()
55 ret = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid, ap); in gfs2_quota_lock_check()
Dinode.c390 struct gfs2_alloc_parms ap = { .target = *dblocks, .aflags = flags, }; in alloc_dinode() local
393 error = gfs2_quota_lock_check(ip, &ap); in alloc_dinode()
397 error = gfs2_inplace_reserve(ip, &ap); in alloc_dinode()
531 struct gfs2_alloc_parms ap = { .target = da->nr_blocks, }; in link_dinode() local
535 error = gfs2_quota_lock_check(dip, &ap); in link_dinode()
539 error = gfs2_inplace_reserve(dip, &ap); in link_dinode()
989 struct gfs2_alloc_parms ap = { .target = da.nr_blocks, }; in gfs2_link() local
990 error = gfs2_quota_lock_check(dip, &ap); in gfs2_link()
994 error = gfs2_inplace_reserve(dip, &ap); in gfs2_link()
1547 struct gfs2_alloc_parms ap = { .target = da.nr_blocks, }; in gfs2_rename() local
[all …]
Dfile.c420 struct gfs2_alloc_parms ap = { .aflags = 0, }; in gfs2_page_mkwrite() local
481 ap.target = data_blocks + ind_blocks; in gfs2_page_mkwrite()
482 err = gfs2_quota_lock_check(ip, &ap); in gfs2_page_mkwrite()
487 err = gfs2_inplace_reserve(ip, &ap); in gfs2_page_mkwrite()
1248 struct gfs2_alloc_parms ap = { .aflags = 0, }; in __gfs2_fallocate() local
1273 ap.min_target = data_blocks + ind_blocks; in __gfs2_fallocate()
1295 ap.target = data_blocks + ind_blocks; in __gfs2_fallocate()
1297 error = gfs2_quota_lock_check(ip, &ap); in __gfs2_fallocate()
1303 if (ap.allowed) in __gfs2_fallocate()
1304 max_blks = ap.allowed; in __gfs2_fallocate()
[all …]
Dquota.c895 struct gfs2_alloc_parms ap = { .aflags = 0, }; in do_sync() local
950 ap.target = reserved; in do_sync()
951 error = gfs2_inplace_reserve(ip, &ap); in do_sync()
1218 struct gfs2_alloc_parms *ap) in gfs2_quota_check() argument
1226 ap->allowed = UINT_MAX; /* Assume we are permitted a whole lot */ in gfs2_quota_check()
1244 if (limit > 0 && (limit - value) < ap->allowed) in gfs2_quota_check()
1245 ap->allowed = limit - value; in gfs2_quota_check()
1247 if (limit && limit < (value + (s64)ap->target)) { in gfs2_quota_check()
1250 if (!ap->min_target || ap->min_target > ap->allowed) { in gfs2_quota_check()
1720 struct gfs2_alloc_parms ap = { .aflags = 0, }; in gfs2_set_dqblk() local
[all …]
Drgrp.c1563 const struct gfs2_alloc_parms *ap) in rg_mblk_search() argument
1585 extlen = max_t(u32, atomic_read(&ip->i_sizehint), ap->target); in rg_mblk_search()
2045 int gfs2_inplace_reserve(struct gfs2_inode *ip, struct gfs2_alloc_parms *ap) in gfs2_inplace_reserve() argument
2053 u32 target = ap->target; in gfs2_inplace_reserve()
2072 if (S_ISDIR(ip->i_inode.i_mode) && (ap->aflags & GFS2_AF_ORLOV)) in gfs2_inplace_reserve()
2129 rg_mblk_search(rs->rs_rgd, ip, ap); in gfs2_inplace_reserve()
2144 rs->rs_reserved = ap->target; in gfs2_inplace_reserve()
2186 if (ap->min_target) in gfs2_inplace_reserve()
2187 target = ap->min_target; in gfs2_inplace_reserve()
Drgrp.h41 struct gfs2_alloc_parms *ap);
Dbmap.c1006 struct gfs2_alloc_parms ap = {}; in gfs2_iomap_begin_write() local
1012 ap.target = data_blocks + ind_blocks; in gfs2_iomap_begin_write()
1013 ret = gfs2_quota_lock_check(ip, &ap); in gfs2_iomap_begin_write()
1017 ret = gfs2_inplace_reserve(ip, &ap); in gfs2_iomap_begin_write()
2056 struct gfs2_alloc_parms ap = { .target = 1, }; in do_grow() local
2062 error = gfs2_quota_lock_check(ip, &ap); in do_grow()
2066 error = gfs2_inplace_reserve(ip, &ap); in do_grow()
Dxattr.c741 struct gfs2_alloc_parms ap = { .target = blks }; in ea_alloc_skeleton() local
748 error = gfs2_quota_lock_check(ip, &ap); in ea_alloc_skeleton()
752 error = gfs2_inplace_reserve(ip, &ap); in ea_alloc_skeleton()
/fs/xfs/scrub/
Dbitmap.c69 struct xbitmap_range *ap; in xbitmap_range_cmp() local
72 ap = container_of(a, struct xbitmap_range, list); in xbitmap_range_cmp()
75 if (ap->start > bp->start) in xbitmap_range_cmp()
77 if (ap->start < bp->start) in xbitmap_range_cmp()
/fs/ext4/
Dioctl.c290 unsigned char *ap, *bp; in memswap() local
292 ap = (unsigned char *)a; in memswap()
295 swap(*ap, *bp); in memswap()
296 ap++; in memswap()