/kernel/linux/linux-5.10/Documentation/scsi/scsi_transport_srp/ |
D | rport_state_diagram.dot | 2 node [shape = doublecircle]; running lost; 7 running_rta [ label = "running;\nreconnect\ntimer\nactive" ]; 9 running [ label = "running;\nreconnect\ntimer\nstopped" ]; 14 …running -> running_rta [ label = "fast_io_fail_tmo = off and\ndev_loss_tmo = off;\nsrp_start_tl_f… 15 …running_rta -> running [ label = "fast_io_fail_tmo = off and\ndev_loss_tmo = off;\nreconnecting s… 16 …running -> blocked [ label = "fast_io_fail_tmo >= 0 or\ndev_loss_tmo >= 0;\nsrp_start_tl_fail_tim… 17 …running -> failfast [ label = "fast_io_fail_tmo = off and\ndev_loss_tmo = off;\nreconnecting fail… 21 blocked -> running [ label = "reconnecting\nsucceeded" ]; 23 failfast -> running [ label = "reconnecting\nsucceeded" ]; 24 running -> lost [ label = "srp_stop_rport_timers()" ];
|
/kernel/linux/linux-5.10/Documentation/ABI/stable/ |
D | sysfs-hypervisor-xen | 5 Description: If running under Xen: 14 Description: If running under Xen: 23 Description: If running under Xen: 32 Description: If running under Xen: 54 Description: If running under Xen: 63 Description: If running under Xen: 71 Description: If running under Xen: 80 Description: If running under Xen: 87 Description: If running under Xen: 95 Description: If running under Xen: [all …]
|
/kernel/linux/linux-5.10/kernel/sched/ |
D | pelt.c | 107 unsigned long load, unsigned long runnable, int running) in accumulate_sum() argument 133 * runnable = running = 0; in accumulate_sum() 149 if (running) in accumulate_sum() 185 unsigned long load, unsigned long runnable, int running) in ___update_load_sum() argument 210 * running is a subset of runnable (weight) so running can't be set if in ___update_load_sum() 213 * This means that weight will be 0 but not running for a sched_entity in ___update_load_sum() 221 runnable = running = 0; in ___update_load_sum() 230 if (!accumulate_sum(delta, sa, load, runnable, running)) in ___update_load_sum() 350 int update_rt_rq_load_avg(u64 now, struct rq *rq, int running) in update_rt_rq_load_avg() argument 353 running, in update_rt_rq_load_avg() [all …]
|
D | pelt.h | 7 int update_rt_rq_load_avg(u64 now, struct rq *rq, int running); 8 int update_dl_rq_load_avg(u64 now, struct rq *rq, int running); 31 int update_irq_load_avg(struct rq *rq, u64 running); 34 update_irq_load_avg(struct rq *rq, u64 running) in update_irq_load_avg() argument 66 * computation done during the running delta time but then sync back to 89 * Running longer results in stealing idle time that will in update_rq_clock_pelt() 168 update_rt_rq_load_avg(u64 now, struct rq *rq, int running) in update_rt_rq_load_avg() argument 174 update_dl_rq_load_avg(u64 now, struct rq *rq, int running) in update_dl_rq_load_avg() argument 191 update_irq_load_avg(struct rq *rq, u64 running) in update_irq_load_avg() argument
|
/kernel/linux/linux-5.10/include/uapi/linux/ |
D | membarrier.h | 34 * @MEMBARRIER_CMD_GLOBAL: Execute a memory barrier on all running threads. 36 * is ensured that all running threads have passed 40 * (non-running threads are de facto in such a 42 * running on the system. This command returns 0. 44 * Execute a memory barrier on all running threads 48 * is ensured that all running threads have passed 52 * (non-running threads are de facto in such a 66 * Execute a memory barrier on each running 69 * caller thread is ensured that all its running 74 * (non-running threads are de facto in such a [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/vm/ |
D | run_vmtests | 73 echo "running hugepage-mmap" 88 echo "running hugepage-shm" 101 echo "running map_hugetlb" 116 echo "running map_fixed_noreplace" 127 echo "running 'gup_benchmark -U' (normal/slow gup)" 138 echo "running gup_benchmark -b (pin_user_pages)" 149 echo "running userfaultfd" 160 echo "running userfaultfd_hugetlb" 174 echo "running userfaultfd_shmem" 190 echo "running compaction_test" [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gvt/ |
D | execlist.c | 66 gvt_dbg_el("[before] running slot %d/context %x pending slot %d\n", in switch_virtual_execlist_slot() 79 gvt_dbg_el("[after] running slot %d/context %x pending slot %d\n", in switch_virtual_execlist_slot() 90 struct intel_vgpu_execlist_slot *running = execlist->running_slot; in emulate_execlist_status() local 101 if (running) { in emulate_execlist_status() 102 status.current_execlist_pointer = !!running->index; in emulate_execlist_status() 103 status.execlist_write_pointer = !!!running->index; in emulate_execlist_status() 105 !!!(running->index); in emulate_execlist_status() 107 !!(running->index); in emulate_execlist_status() 185 struct intel_vgpu_execlist_slot *running = execlist->running_slot; in emulate_execlist_ctx_schedule_out() local 187 struct execlist_ctx_descriptor_format *ctx0 = &running->ctx[0]; in emulate_execlist_ctx_schedule_out() [all …]
|
/kernel/linux/linux-5.10/Documentation/networking/devlink/ |
D | nfp.rst | 46 - stored, running 49 - stored, running 52 - stored, running 55 - stored, running 58 - stored, running 61 - stored, running 64 - stored, running
|
D | bnxt.rst | 64 - stored, running 67 - stored, running 70 - stored, running 73 - running 76 - stored, running 79 - stored, running
|
D | ice.rst | 27 - running 32 - running 38 - running 42 - running 52 - running 56 - running 61 - running 67 - running 73 - running 79 - running [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/media_tests/ |
D | media_dev_allocator.sh | 19 echo "Running unbind of $MDEV from $MDRIVER" 28 echo "Running unbind of $ADEV from $ADRIVER" 38 echo "Running bind of $MDEV from $MDRIVER" 44 echo "Running bind of $ADEV from $ADRIVER" 56 echo "Running unbind of $MDEV from $MDRIVER" 64 echo "Running bind of $MDEV from $MDRIVER" 70 echo "Running unbind of $ADEV from $ADRIVER" 78 echo "Running bind of $ADEV from $ADRIVER"
|
/kernel/linux/linux-5.10/Documentation/virt/kvm/ |
D | running-nested-guests.rst | 2 Running nested guests with KVM 29 - L0 – level-0; the bare metal host, running KVM 31 - L1 – level-1 guest; a VM running on L0; also called the "guest 32 hypervisor", as it itself is capable of running KVM. 34 - L2 – level-2 guest; a VM running on L1, this is the "nested guest" 41 hypervisor running on bare metal, adding another layer and 43 metal, running the LPAR hypervisor), L1 (host hypervisor), L2 61 multiple nested guests (level-2 guests), running different OSes, on 81 In case you are running a Linux kernel older than v4.19, to enable 137 .. note:: If you suspect your L2 (i.e. nested guest) is running slower, [all …]
|
/kernel/linux/linux-5.10/Documentation/driver-api/thermal/ |
D | cpu-idle-cooling.rst | 83 running 95 cycle (aka the cooling device state), the running duration can be 111 running 125 running 160 (Ptarget) resulting in an amount of time running at full power on a 164 P(opp)target = ((Trunning x (P(opp)running) + (Tidle x P(opp)idle)) / 169 Tidle = Trunning x ((P(opp)running / P(opp)target) - 1) 171 At this point if we know the running period for the CPU, that gives us 173 injection duration, we can compute the running duration with:: 175 Trunning = Tidle / ((P(opp)running / P(opp)target) - 1) [all …]
|
/kernel/linux/linux-5.10/sound/isa/sb/ |
D | sb16_csp.c | 159 if (p->running & SNDRV_SB_CSP_ST_RUNNING) in snd_sb_csp_free() 208 info.state = p->running; in snd_sb_csp_ioctl() 217 err = (p->running & SNDRV_SB_CSP_ST_RUNNING ? in snd_sb_csp_ioctl() 221 err = (p->running & SNDRV_SB_CSP_ST_RUNNING ? in snd_sb_csp_ioctl() 225 /* change CSP running state */ in snd_sb_csp_ioctl() 356 p->running = 0; in snd_sb_csp_riff_load() 440 p->running = SNDRV_SB_CSP_ST_LOADED; /* set LOADED flag */ in snd_sb_csp_riff_load() 453 if (p->running & SNDRV_SB_CSP_ST_RUNNING) in snd_sb_csp_unload() 455 if (!(p->running & SNDRV_SB_CSP_ST_LOADED)) in snd_sb_csp_unload() 466 p->running = 0; in snd_sb_csp_unload() [all …]
|
/kernel/linux/linux-5.10/sound/pci/ctxfi/ |
D | cttimer.c | 40 unsigned int running:1; member 55 unsigned int running:1; /* global timer running */ member 86 if (ti->running) in ct_systimer_callback() 102 ti->running = 1; in ct_systimer_start() 114 ti->running = 0; in ct_systimer_stop() 150 if (!atimer->running) in ct_xfitimer_irq_rearm() 152 atimer->running = 1; in ct_xfitimer_irq_rearm() 157 if (atimer->running) { in ct_xfitimer_irq_stop() 161 atimer->running = 0; in ct_xfitimer_irq_stop() 173 * checks the running instance list and determines the next timer interval. [all …]
|
/kernel/linux/linux-5.10/Documentation/ia64/ |
D | mca.rst | 43 * If an MCA/INIT event occurs while the kernel was running (not user 67 running or not. That is, whether it is on a cpu or is blocked. The 70 INIT event are no longer running, they have been converted to blocked 72 rendezvous interrupt are still running on their normal kernel stacks! 78 no longer running on that cpu. That gives us a decent chance of 83 failed, not on the task that was originally running. Again this 85 own stack as running on that cpu. Then a recursive error gets a 144 * i386 backtrace is not very sensitive to whether a process is running 146 running or not. 150 What happens when MCA/INIT is delivered what a cpu is running user [all …]
|
/kernel/linux/linux-5.10/tools/perf/arch/x86/tests/ |
D | rdpmc.c | 41 u64 count, cyc = 0, time_offset = 0, enabled, running, delta; in mmap_read_self() local 48 running = pc->time_running; in mmap_read_self() 50 if (enabled != running) { in mmap_read_self() 65 if (enabled != running) { in mmap_read_self() 75 running += delta; in mmap_read_self() 77 quot = count / running; in mmap_read_self() 78 rem = count % running; in mmap_read_self() 79 count = quot * enabled + (rem * enabled) / running; in mmap_read_self()
|
/kernel/linux/linux-5.10/net/core/ |
D | gen_estimator.c | 45 seqcount_t *running; member 69 __gnet_stats_copy_basic(e->running, b, e->cpu_bstats, e->bstats); in est_fetch_counters() 112 * @running: qdisc running seqcount 128 seqcount_t *running, in gen_new_estimator() argument 157 est->running = running; in gen_new_estimator() 217 * @running: qdisc running seqcount (might be NULL) 229 seqcount_t *running, struct nlattr *opt) in gen_replace_estimator() argument 232 lock, running, opt); in gen_replace_estimator()
|
/kernel/linux/linux-5.10/Documentation/dev-tools/kunit/ |
D | faq.rst | 25 Does KUnit support running on architectures other than UML? 71 1. Try running ``./tools/testing/kunit/kunit.py run`` with the ``--raw_output`` 74 2. Instead of running ``kunit.py run``, try running ``kunit.py config``, 78 3. Running the UML kernel directly can often reveal issues or error messages 79 kunit_tool ignores. This should be as simple as running ``./vmlinux`` after 87 around, so you can see what config was used after running ``kunit.py run``. 91 5. Try to run ``make ARCH=um defconfig`` before running ``kunit.py run``. This 93 6. Finally, try running KUnit outside UML. KUnit and KUnit tests can be
|
/kernel/linux/linux-5.10/tools/testing/selftests/firmware/ |
D | fw_run_tests.sh | 30 echo "Running kernel configuration test 1 -- rare" 41 echo "Running kernel configuration test 2 -- distro" 53 echo "Running kernel configuration test 3 -- android" 64 echo "Running namespace test: " 73 echo "Running basic kernel configuration, working with your config"
|
/kernel/linux/linux-5.10/tools/testing/selftests/powerpc/math/ |
D | fpu_preempt.c | 38 int running; variable 40 extern void preempt_fpu(double *darray, int *threads_starting, int *running); 50 preempt_fpu(darray, &threads_starting, &running); in preempt_fpu_c() 64 running = true; in test_preempt_fpu() 85 * r5 will have loaded the value of running. in test_preempt_fpu() 87 running = 0; in test_preempt_fpu()
|
/kernel/linux/linux-5.10/arch/powerpc/kvm/ |
D | Kconfig | 58 Support running unmodified book3s_32 guest kernels 74 Support running unmodified book3s_64 and book3s_32 guest kernels 89 Support running unmodified book3s_64 guest kernels in 108 Support running guest kernels in virtual machines on processors 109 without using hypervisor mode in the host, by running the 154 Support running unmodified E500 guest kernels in virtual machines on 170 Support running unmodified E500MC/E5500/E6500 guest kernels in
|
/kernel/linux/linux-5.10/sound/core/seq/oss/ |
D | seq_oss_timer.c | 44 rec->running = 0; in snd_seq_oss_timer_new() 152 if (timer->running) in snd_seq_oss_timer_start() 162 timer->running = 1; in snd_seq_oss_timer_start() 174 if (! timer->running) in snd_seq_oss_timer_stop() 177 timer->running = 0; in snd_seq_oss_timer_stop() 188 if (timer->running) in snd_seq_oss_timer_continue() 191 timer->running = 1; in snd_seq_oss_timer_continue() 208 if (timer->running) in snd_seq_oss_timer_tempo()
|
/kernel/linux/linux-5.10/drivers/thermal/ |
D | cpuidle_cooling.c | 32 * cpuidle_cooling_runtime - Running time computation 36 * The running duration is computed from the idle injection duration 38 * means the running duration is zero. If we have a 50% ratio 40 * running duration. 44 * running = idle x ((100 / ratio) - 1) 48 * running = (idle x 100) / ratio - idle 51 * with 10ms of idle injection and 10ms of running duration. 78 * Depending on the configuration or the hardware, the running in cpuidle_cooling_get_max_state() 85 * means for 10ms of idle injection, we have 10ms of running in cpuidle_cooling_get_max_state()
|
/kernel/linux/linux-5.10/Documentation/driver-api/dmaengine/ |
D | pxa_dma.rst | 10 is queued even on a running DMA channel. 15 stop and restart, but is submitted on a "running channel". The other 25 c) Channel running state 26 A driver should be able to query if a channel is running or not. For the 30 know if a channel is in running or stopped state. 71 Suppose the running chain is: 107 this specific case if the DMA is already running in aligned mode. 117 any lock to find out what is the latest completed transfer in a running 139 currently running descriptor. 150 - a driver issued tx1+tx2 => channel is running in aligned mode
|