/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvif/ |
D | fifo.c | 32 struct nv_device_info_v1_data runlist[64]; in nvif_fifo_runlists() member 37 if (device->runlist) in nvif_fifo_runlists() 45 for (i = 0; i < ARRAY_SIZE(a->v.runlist); i++) in nvif_fifo_runlists() 46 a->v.runlist[i].mthd = NV_DEVICE_FIFO_RUNLIST_ENGINES(i); in nvif_fifo_runlists() 53 device->runlist = kcalloc(device->runlists, sizeof(*device->runlist), in nvif_fifo_runlists() 55 if (!device->runlist) { in nvif_fifo_runlists() 62 device->runlist[i].engines = a->v.runlist[i].data; in nvif_fifo_runlists() 93 if (device->runlist[i].engines & a.v.engine.data) in nvif_fifo_runlist()
|
D | device.c | 45 kfree(device->runlist); in nvif_device_dtor() 46 device->runlist = NULL; in nvif_device_dtor() 56 device->runlist = NULL; in nvif_device_ctor()
|
/kernel/linux/linux-5.10/fs/ntfs/ |
D | runlist.h | 43 } runlist; typedef 45 static inline void ntfs_init_runlist(runlist *rl) in ntfs_init_runlist() 81 runlist *const runlist, const s64 new_length); 83 int ntfs_rl_punch_nolock(const ntfs_volume *vol, runlist *const runlist,
|
D | runlist.c | 1485 int ntfs_rl_truncate_nolock(const ntfs_volume *vol, runlist *const runlist, in ntfs_rl_truncate_nolock() argument 1492 BUG_ON(!runlist); in ntfs_rl_truncate_nolock() 1494 rl = runlist->rl; in ntfs_rl_truncate_nolock() 1497 runlist->rl = NULL; in ntfs_rl_truncate_nolock() 1513 runlist->rl = rl; in ntfs_rl_truncate_nolock() 1537 old_size = trl - runlist->rl + 1; in ntfs_rl_truncate_nolock() 1555 int new_size = rl - runlist->rl + 1; in ntfs_rl_truncate_nolock() 1556 rl = ntfs_rl_realloc(runlist->rl, old_size, new_size); in ntfs_rl_truncate_nolock() 1564 runlist->rl = rl; in ntfs_rl_truncate_nolock() 1573 if ((rl > runlist->rl) && ((rl - 1)->lcn == LCN_HOLE)) in ntfs_rl_truncate_nolock() [all …]
|
D | attrib.c | 173 rl = ntfs_mapping_pairs_decompress(ni->vol, a, ni->runlist.rl); in ntfs_map_runlist_nolock() 177 ni->runlist.rl = rl; in ntfs_map_runlist_nolock() 288 down_write(&ni->runlist.lock); in ntfs_map_runlist() 290 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= in ntfs_map_runlist() 293 up_write(&ni->runlist.lock); in ntfs_map_runlist() 340 if (!ni->runlist.rl) { in ntfs_attr_vcn_to_lcn_nolock() 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 362 up_read(&ni->runlist.lock); in ntfs_attr_vcn_to_lcn_nolock() 363 down_write(&ni->runlist.lock); in ntfs_attr_vcn_to_lcn_nolock() 364 if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) != in ntfs_attr_vcn_to_lcn_nolock() [all …]
|
D | aops.c | 186 BUG_ON(!ni->runlist.rl && !ni->mft_no && !NInoAttr(ni)); in ntfs_read_block() 248 down_read(&ni->runlist.lock); in ntfs_read_block() 249 rl = ni->runlist.rl; in ntfs_read_block() 282 up_read(&ni->runlist.lock); in ntfs_read_block() 288 up_read(&ni->runlist.lock); in ntfs_read_block() 328 up_read(&ni->runlist.lock); in ntfs_read_block() 704 down_read(&ni->runlist.lock); in ntfs_write_block() 705 rl = ni->runlist.rl; in ntfs_write_block() 760 up_read(&ni->runlist.lock); in ntfs_write_block() 766 up_read(&ni->runlist.lock); in ntfs_write_block() [all …]
|
D | mft.c | 537 runlist.lock); in ntfs_sync_mft_mirror() 538 rl = NTFS_I(vol->mftmirr_ino)->runlist.rl; in ntfs_sync_mft_mirror() 574 up_read(&NTFS_I(vol->mftmirr_ino)->runlist.lock); in ntfs_sync_mft_mirror() 729 down_read(&NTFS_I(vol->mft_ino)->runlist.lock); in write_mft_record_nolock() 730 rl = NTFS_I(vol->mft_ino)->runlist.rl; in write_mft_record_nolock() 761 up_read(&NTFS_I(vol->mft_ino)->runlist.lock); in write_mft_record_nolock() 1300 down_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock() 1307 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock() 1328 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock() 1354 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock() [all …]
|
D | file.c | 830 down_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 832 rl = ni->runlist.rl; in ntfs_prepare_pages_for_non_resident_write() 859 up_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 862 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 884 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 885 down_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 966 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 967 down_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write() 975 while (--rl2 >= ni->runlist.rl) { in ntfs_prepare_pages_for_non_resident_write() 990 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); in ntfs_prepare_pages_for_non_resident_write() [all …]
|
D | inode.h | 55 runlist runlist; /* If state has the NI_NonResident bit set, member 87 runlist attr_list_rl; /* Run list for the attribute list value. */
|
D | compress.c | 602 down_read(&ni->runlist.lock); in ntfs_read_compressed_block() 603 rl = ni->runlist.rl; in ntfs_read_compressed_block() 629 up_read(&ni->runlist.lock); in ntfs_read_compressed_block() 647 up_read(&ni->runlist.lock); in ntfs_read_compressed_block() 926 up_read(&ni->runlist.lock); in ntfs_read_compressed_block() 932 up_read(&ni->runlist.lock); in ntfs_read_compressed_block()
|
D | logfile.c | 745 down_write(&log_ni->runlist.lock); in ntfs_empty_logfile() 746 rl = log_ni->runlist.rl; in ntfs_empty_logfile() 755 rl = log_ni->runlist.rl; in ntfs_empty_logfile() 821 up_write(&log_ni->runlist.lock); in ntfs_empty_logfile() 844 up_write(&log_ni->runlist.lock); in ntfs_empty_logfile()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | nv50.c | 36 cur = fifo->runlist[fifo->cur_runlist]; in nv50_fifo_runlist_update_locked() 67 false, &fifo->runlist[0]); in nv50_fifo_oneinit() 72 false, &fifo->runlist[1]); in nv50_fifo_oneinit() 103 nvkm_memory_unref(&fifo->runlist[1]); in nv50_fifo_dtor() 104 nvkm_memory_unref(&fifo->runlist[0]); in nv50_fifo_dtor()
|
D | gpfifogv100.c | 131 int runlist = ffs(*runlists) -1, ret, i; in gv100_fifo_gpfifo_new_() local 137 if (!vmm || runlist < 0 || runlist >= fifo->runlist_nr) in gv100_fifo_gpfifo_new_() 139 *runlists = BIT_ULL(runlist); in gv100_fifo_gpfifo_new_() 141 engm = fifo->runlist[runlist].engm; in gv100_fifo_gpfifo_new_() 152 chan->runl = runlist; in gv100_fifo_gpfifo_new_() 239 args->v0.ilength, args->v0.runlist, args->v0.priv); in gv100_fifo_gpfifo_new() 243 &args->v0.runlist, in gv100_fifo_gpfifo_new()
|
D | gk104.c | 181 const struct gk104_fifo_runlist_func *func = fifo->func->runlist; in gk104_fifo_runlist_update() 189 mem = fifo->runlist[runl].mem[fifo->runlist[runl].next]; in gk104_fifo_runlist_update() 190 fifo->runlist[runl].next = !fifo->runlist[runl].next; in gk104_fifo_runlist_update() 193 list_for_each_entry(chan, &fifo->runlist[runl].chan, head) { in gk104_fifo_runlist_update() 197 list_for_each_entry(cgrp, &fifo->runlist[runl].cgrp, head) { in gk104_fifo_runlist_update() 229 list_add_tail(&cgrp->head, &fifo->runlist[chan->runl].cgrp); in gk104_fifo_runlist_insert() 232 list_add_tail(&chan->head, &fifo->runlist[chan->runl].chan); in gk104_fifo_runlist_insert() 335 list_for_each_entry(chan, &fifo->runlist[runl].chan, head) { in gk104_fifo_recover_chid() 342 list_for_each_entry(cgrp, &fifo->runlist[runl].cgrp, head) { in gk104_fifo_recover_chid() 364 unsigned long engn, engm = fifo->runlist[runl].engm; in gk104_fifo_recover_chan() [all …]
|
D | gpfifogk104.c | 249 int runlist = ffs(*runlists) -1, ret, i; in gk104_fifo_gpfifo_new_() local 254 if (!vmm || runlist < 0 || runlist >= fifo->runlist_nr) in gk104_fifo_gpfifo_new_() 256 *runlists = BIT_ULL(runlist); in gk104_fifo_gpfifo_new_() 258 engm = fifo->runlist[runlist].engm; in gk104_fifo_gpfifo_new_() 272 chan->runl = runlist; in gk104_fifo_gpfifo_new_() 345 args->v0.ilength, args->v0.runlist, args->v0.priv); in gk104_fifo_gpfifo_new() 349 &args->v0.runlist, in gk104_fifo_gpfifo_new()
|
D | gf100.c | 61 cur = fifo->runlist.mem[fifo->runlist.active]; in gf100_fifo_runlist_commit() 62 fifo->runlist.active = !fifo->runlist.active; in gf100_fifo_runlist_commit() 85 if (wait_event_timeout(fifo->runlist.wait, in gf100_fifo_runlist_commit() 450 wake_up(&fifo->runlist.wait); in gf100_fifo_intr_runlist() 589 false, &fifo->runlist.mem[0]); in gf100_fifo_oneinit() 594 false, &fifo->runlist.mem[1]); in gf100_fifo_oneinit() 598 init_waitqueue_head(&fifo->runlist.wait); in gf100_fifo_oneinit() 663 nvkm_memory_unref(&fifo->runlist.mem[0]); in gf100_fifo_dtor() 664 nvkm_memory_unref(&fifo->runlist.mem[1]); in gf100_fifo_dtor()
|
D | gpfifotu102.c | 67 args->v0.ilength, args->v0.runlist, args->v0.priv); in tu102_fifo_gpfifo_new() 71 &args->v0.runlist, in tu102_fifo_gpfifo_new()
|
D | gk104.h | 38 } runlist[16]; member 74 } *runlist; member
|
D | nv50.h | 9 struct nvkm_memory *runlist[2]; member
|
D | gk20a.c | 36 .runlist = &gk110_fifo_runlist,
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/top/ |
D | base.c | 35 info->runlist = -1; in nvkm_top_device_new() 150 if (info->engine >= 0 && info->runlist >= 0 && n++ == index) { in nvkm_top_engine() 151 *runl = info->runlist; in nvkm_top_engine()
|
D | gk104.c | 58 info->runlist = (data & 0x01e00000) >> 21; in gk104_top_oneinit() 103 info->addr, info->fault, info->engine, info->runlist, in gk104_top_oneinit()
|
D | priv.h | 19 int runlist; member
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/include/nvif/ |
D | cla06f.h | 11 __u64 runlist; member
|
D | clc36f.h | 11 __u64 runlist; member
|