Home
last modified time | relevance | path

Searched refs:runlist (Results 1 – 25 of 45) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvif/
Dfifo.c32 struct nv_device_info_v1_data runlist[64]; in nvif_fifo_runlists() member
37 if (device->runlist) in nvif_fifo_runlists()
45 for (i = 0; i < ARRAY_SIZE(a->v.runlist); i++) in nvif_fifo_runlists()
46 a->v.runlist[i].mthd = NV_DEVICE_FIFO_RUNLIST_ENGINES(i); in nvif_fifo_runlists()
53 device->runlist = kcalloc(device->runlists, sizeof(*device->runlist), in nvif_fifo_runlists()
55 if (!device->runlist) { in nvif_fifo_runlists()
62 device->runlist[i].engines = a->v.runlist[i].data; in nvif_fifo_runlists()
93 if (device->runlist[i].engines & a.v.engine.data) in nvif_fifo_runlist()
Ddevice.c45 kfree(device->runlist); in nvif_device_dtor()
46 device->runlist = NULL; in nvif_device_dtor()
56 device->runlist = NULL; in nvif_device_ctor()
/kernel/linux/linux-5.10/fs/ntfs/
Drunlist.h43 } runlist; typedef
45 static inline void ntfs_init_runlist(runlist *rl) in ntfs_init_runlist()
81 runlist *const runlist, const s64 new_length);
83 int ntfs_rl_punch_nolock(const ntfs_volume *vol, runlist *const runlist,
Drunlist.c1485 int ntfs_rl_truncate_nolock(const ntfs_volume *vol, runlist *const runlist, in ntfs_rl_truncate_nolock() argument
1492 BUG_ON(!runlist); in ntfs_rl_truncate_nolock()
1494 rl = runlist->rl; in ntfs_rl_truncate_nolock()
1497 runlist->rl = NULL; in ntfs_rl_truncate_nolock()
1513 runlist->rl = rl; in ntfs_rl_truncate_nolock()
1537 old_size = trl - runlist->rl + 1; in ntfs_rl_truncate_nolock()
1555 int new_size = rl - runlist->rl + 1; in ntfs_rl_truncate_nolock()
1556 rl = ntfs_rl_realloc(runlist->rl, old_size, new_size); in ntfs_rl_truncate_nolock()
1564 runlist->rl = rl; in ntfs_rl_truncate_nolock()
1573 if ((rl > runlist->rl) && ((rl - 1)->lcn == LCN_HOLE)) in ntfs_rl_truncate_nolock()
[all …]
Dattrib.c173 rl = ntfs_mapping_pairs_decompress(ni->vol, a, ni->runlist.rl); in ntfs_map_runlist_nolock()
177 ni->runlist.rl = rl; in ntfs_map_runlist_nolock()
288 down_write(&ni->runlist.lock); in ntfs_map_runlist()
290 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= in ntfs_map_runlist()
293 up_write(&ni->runlist.lock); in ntfs_map_runlist()
340 if (!ni->runlist.rl) { in ntfs_attr_vcn_to_lcn_nolock()
350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock()
362 up_read(&ni->runlist.lock); in ntfs_attr_vcn_to_lcn_nolock()
363 down_write(&ni->runlist.lock); in ntfs_attr_vcn_to_lcn_nolock()
364 if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) != in ntfs_attr_vcn_to_lcn_nolock()
[all …]
Daops.c186 BUG_ON(!ni->runlist.rl && !ni->mft_no && !NInoAttr(ni)); in ntfs_read_block()
248 down_read(&ni->runlist.lock); in ntfs_read_block()
249 rl = ni->runlist.rl; in ntfs_read_block()
282 up_read(&ni->runlist.lock); in ntfs_read_block()
288 up_read(&ni->runlist.lock); in ntfs_read_block()
328 up_read(&ni->runlist.lock); in ntfs_read_block()
704 down_read(&ni->runlist.lock); in ntfs_write_block()
705 rl = ni->runlist.rl; in ntfs_write_block()
760 up_read(&ni->runlist.lock); in ntfs_write_block()
766 up_read(&ni->runlist.lock); in ntfs_write_block()
[all …]
Dmft.c537 runlist.lock); in ntfs_sync_mft_mirror()
538 rl = NTFS_I(vol->mftmirr_ino)->runlist.rl; in ntfs_sync_mft_mirror()
574 up_read(&NTFS_I(vol->mftmirr_ino)->runlist.lock); in ntfs_sync_mft_mirror()
729 down_read(&NTFS_I(vol->mft_ino)->runlist.lock); in write_mft_record_nolock()
730 rl = NTFS_I(vol->mft_ino)->runlist.rl; in write_mft_record_nolock()
761 up_read(&NTFS_I(vol->mft_ino)->runlist.lock); in write_mft_record_nolock()
1300 down_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock()
1307 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock()
1328 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock()
1354 up_write(&mftbmp_ni->runlist.lock); in ntfs_mft_bitmap_extend_allocation_nolock()
[all …]
Dfile.c830 down_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
832 rl = ni->runlist.rl; in ntfs_prepare_pages_for_non_resident_write()
859 up_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
862 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
884 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
885 down_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
966 up_read(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
967 down_write(&ni->runlist.lock); in ntfs_prepare_pages_for_non_resident_write()
975 while (--rl2 >= ni->runlist.rl) { in ntfs_prepare_pages_for_non_resident_write()
990 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); in ntfs_prepare_pages_for_non_resident_write()
[all …]
Dinode.h55 runlist runlist; /* If state has the NI_NonResident bit set, member
87 runlist attr_list_rl; /* Run list for the attribute list value. */
Dcompress.c602 down_read(&ni->runlist.lock); in ntfs_read_compressed_block()
603 rl = ni->runlist.rl; in ntfs_read_compressed_block()
629 up_read(&ni->runlist.lock); in ntfs_read_compressed_block()
647 up_read(&ni->runlist.lock); in ntfs_read_compressed_block()
926 up_read(&ni->runlist.lock); in ntfs_read_compressed_block()
932 up_read(&ni->runlist.lock); in ntfs_read_compressed_block()
Dlogfile.c745 down_write(&log_ni->runlist.lock); in ntfs_empty_logfile()
746 rl = log_ni->runlist.rl; in ntfs_empty_logfile()
755 rl = log_ni->runlist.rl; in ntfs_empty_logfile()
821 up_write(&log_ni->runlist.lock); in ntfs_empty_logfile()
844 up_write(&log_ni->runlist.lock); in ntfs_empty_logfile()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dnv50.c36 cur = fifo->runlist[fifo->cur_runlist]; in nv50_fifo_runlist_update_locked()
67 false, &fifo->runlist[0]); in nv50_fifo_oneinit()
72 false, &fifo->runlist[1]); in nv50_fifo_oneinit()
103 nvkm_memory_unref(&fifo->runlist[1]); in nv50_fifo_dtor()
104 nvkm_memory_unref(&fifo->runlist[0]); in nv50_fifo_dtor()
Dgpfifogv100.c131 int runlist = ffs(*runlists) -1, ret, i; in gv100_fifo_gpfifo_new_() local
137 if (!vmm || runlist < 0 || runlist >= fifo->runlist_nr) in gv100_fifo_gpfifo_new_()
139 *runlists = BIT_ULL(runlist); in gv100_fifo_gpfifo_new_()
141 engm = fifo->runlist[runlist].engm; in gv100_fifo_gpfifo_new_()
152 chan->runl = runlist; in gv100_fifo_gpfifo_new_()
239 args->v0.ilength, args->v0.runlist, args->v0.priv); in gv100_fifo_gpfifo_new()
243 &args->v0.runlist, in gv100_fifo_gpfifo_new()
Dgk104.c181 const struct gk104_fifo_runlist_func *func = fifo->func->runlist; in gk104_fifo_runlist_update()
189 mem = fifo->runlist[runl].mem[fifo->runlist[runl].next]; in gk104_fifo_runlist_update()
190 fifo->runlist[runl].next = !fifo->runlist[runl].next; in gk104_fifo_runlist_update()
193 list_for_each_entry(chan, &fifo->runlist[runl].chan, head) { in gk104_fifo_runlist_update()
197 list_for_each_entry(cgrp, &fifo->runlist[runl].cgrp, head) { in gk104_fifo_runlist_update()
229 list_add_tail(&cgrp->head, &fifo->runlist[chan->runl].cgrp); in gk104_fifo_runlist_insert()
232 list_add_tail(&chan->head, &fifo->runlist[chan->runl].chan); in gk104_fifo_runlist_insert()
335 list_for_each_entry(chan, &fifo->runlist[runl].chan, head) { in gk104_fifo_recover_chid()
342 list_for_each_entry(cgrp, &fifo->runlist[runl].cgrp, head) { in gk104_fifo_recover_chid()
364 unsigned long engn, engm = fifo->runlist[runl].engm; in gk104_fifo_recover_chan()
[all …]
Dgpfifogk104.c249 int runlist = ffs(*runlists) -1, ret, i; in gk104_fifo_gpfifo_new_() local
254 if (!vmm || runlist < 0 || runlist >= fifo->runlist_nr) in gk104_fifo_gpfifo_new_()
256 *runlists = BIT_ULL(runlist); in gk104_fifo_gpfifo_new_()
258 engm = fifo->runlist[runlist].engm; in gk104_fifo_gpfifo_new_()
272 chan->runl = runlist; in gk104_fifo_gpfifo_new_()
345 args->v0.ilength, args->v0.runlist, args->v0.priv); in gk104_fifo_gpfifo_new()
349 &args->v0.runlist, in gk104_fifo_gpfifo_new()
Dgf100.c61 cur = fifo->runlist.mem[fifo->runlist.active]; in gf100_fifo_runlist_commit()
62 fifo->runlist.active = !fifo->runlist.active; in gf100_fifo_runlist_commit()
85 if (wait_event_timeout(fifo->runlist.wait, in gf100_fifo_runlist_commit()
450 wake_up(&fifo->runlist.wait); in gf100_fifo_intr_runlist()
589 false, &fifo->runlist.mem[0]); in gf100_fifo_oneinit()
594 false, &fifo->runlist.mem[1]); in gf100_fifo_oneinit()
598 init_waitqueue_head(&fifo->runlist.wait); in gf100_fifo_oneinit()
663 nvkm_memory_unref(&fifo->runlist.mem[0]); in gf100_fifo_dtor()
664 nvkm_memory_unref(&fifo->runlist.mem[1]); in gf100_fifo_dtor()
Dgpfifotu102.c67 args->v0.ilength, args->v0.runlist, args->v0.priv); in tu102_fifo_gpfifo_new()
71 &args->v0.runlist, in tu102_fifo_gpfifo_new()
Dgk104.h38 } runlist[16]; member
74 } *runlist; member
Dnv50.h9 struct nvkm_memory *runlist[2]; member
Dgk20a.c36 .runlist = &gk110_fifo_runlist,
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/top/
Dbase.c35 info->runlist = -1; in nvkm_top_device_new()
150 if (info->engine >= 0 && info->runlist >= 0 && n++ == index) { in nvkm_top_engine()
151 *runl = info->runlist; in nvkm_top_engine()
Dgk104.c58 info->runlist = (data & 0x01e00000) >> 21; in gk104_top_oneinit()
103 info->addr, info->fault, info->engine, info->runlist, in gk104_top_oneinit()
Dpriv.h19 int runlist; member
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/include/nvif/
Dcla06f.h11 __u64 runlist; member
Dclc36f.h11 __u64 runlist; member

12