| /tools/testing/selftests/mqueue/ |
| D | mq_open_tests.c | 59 mqd_t queue = -1; variable 90 if (queue != -1) in shutdown() 91 if (mq_close(queue)) in shutdown() 205 if ((queue = mq_open(queue_path, flags, perms, attr)) == -1) in test_queue() 207 if (mq_getattr(queue, result)) in test_queue() 209 if (mq_close(queue)) in test_queue() 211 queue = -1; in test_queue() 227 if ((queue = mq_open(queue_path, flags, perms, attr)) == -1) in test_queue_fail() 229 if (mq_getattr(queue, result)) in test_queue_fail() 231 if (mq_close(queue)) in test_queue_fail() [all …]
|
| D | mq_perf_tests.c | 99 mqd_t queue = -1; variable 192 if (queue != -1) in shutdown() 193 if (mq_close(queue)) in shutdown() 297 queue = mq_open(queue_path, flags, perms, attr); in open_queue() 298 if (queue == -1) in open_queue() 300 if (mq_getattr(queue, &result)) in open_queue() 335 while (mq_send(queue, buff, sizeof(buff), 0) == 0) in cont_thread() 337 mq_receive(queue, buff, sizeof(buff), &priority); in cont_thread() 342 while (mq_receive(queue, buff, MSG_SIZE, &prio_in) == MSG_SIZE) 346 if (mq_send(queue, buff, MSG_SIZE, prio_out)) \ [all …]
|
| /tools/testing/selftests/net/tcp_ao/lib/ |
| D | repair.c | 42 static void test_sock_checkpoint_queue(int sk, int queue, int qlen, in test_sock_checkpoint_queue() argument 48 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_checkpoint_queue() 127 static void test_sock_restore_seq(int sk, int queue, uint32_t seq) in test_sock_restore_seq() argument 129 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_restore_seq() 136 static void test_sock_restore_queue(int sk, int queue, void *buf, int len) in test_sock_restore_queue() argument 144 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_restore_queue()
|
| /tools/testing/selftests/sched_ext/ |
| D | dsp_local_on.bpf.c | 17 } queue SEC(".maps"); 30 if (bpf_map_push_elem(&queue, &pid, 0)) in BPF_STRUCT_OPS() 39 if (bpf_map_pop_elem(&queue, &pid)) in BPF_STRUCT_OPS()
|
| /tools/perf/util/ |
| D | intel-bts.c | 167 struct auxtrace_queue *queue, in intel_bts_setup_queue() argument 170 struct intel_bts_queue *btsq = queue->priv; in intel_bts_setup_queue() 172 if (list_empty(&queue->head)) in intel_bts_setup_queue() 179 queue->priv = btsq; in intel_bts_setup_queue() 181 if (queue->cpu != -1) in intel_bts_setup_queue() 182 btsq->cpu = queue->cpu; in intel_bts_setup_queue() 183 btsq->tid = queue->tid; in intel_bts_setup_queue() 192 btsq->buffer = auxtrace_buffer__next(queue, NULL); in intel_bts_setup_queue() 248 static int intel_bts_do_fix_overlap(struct auxtrace_queue *queue, in intel_bts_do_fix_overlap() argument 254 if (b->list.prev == &queue->head) in intel_bts_do_fix_overlap() [all …]
|
| D | arm-spe.c | 154 struct auxtrace_queue *queue; in arm_spe_get_trace() local 156 queue = &speq->spe->queues.queue_array[speq->queue_nr]; in arm_spe_get_trace() 158 buffer = auxtrace_buffer__next(queue, buffer); in arm_spe_get_trace() 241 struct auxtrace_queue *queue) in arm_spe_set_pid_tid_cpu() argument 243 struct arm_spe_queue *speq = queue->priv; in arm_spe_set_pid_tid_cpu() 251 speq->tid = queue->tid; in arm_spe_set_pid_tid_cpu() 260 if (queue->cpu == -1) in arm_spe_set_pid_tid_cpu() 723 struct auxtrace_queue *queue, in arm_spe__setup_queue() argument 726 struct arm_spe_queue *speq = queue->priv; in arm_spe__setup_queue() 729 if (list_empty(&queue->head) || speq) in arm_spe__setup_queue() [all …]
|
| D | s390-cpumsf.c | 698 struct auxtrace_queue *queue; in s390_cpumsf_run_decoder() local 701 queue = &sfq->sf->queues.queue_array[sfq->queue_nr]; in s390_cpumsf_run_decoder() 710 sfq->buffer = buffer = auxtrace_buffer__next(queue, in s390_cpumsf_run_decoder() 800 struct auxtrace_queue *queue, in s390_cpumsf_setup_queue() argument 803 struct s390_cpumsf_queue *sfq = queue->priv; in s390_cpumsf_setup_queue() 805 if (list_empty(&queue->head)) in s390_cpumsf_setup_queue() 812 queue->priv = sfq; in s390_cpumsf_setup_queue() 814 if (queue->cpu != -1) in s390_cpumsf_setup_queue() 815 sfq->cpu = queue->cpu; in s390_cpumsf_setup_queue() 850 struct auxtrace_queue *queue; in s390_cpumsf_process_queues() local [all …]
|
| D | cs-etm.c | 561 static void cs_etm__clear_packet_queue(struct cs_etm_packet_queue *queue) in cs_etm__clear_packet_queue() argument 565 queue->head = 0; in cs_etm__clear_packet_queue() 566 queue->tail = 0; in cs_etm__clear_packet_queue() 567 queue->packet_count = 0; in cs_etm__clear_packet_queue() 569 queue->packet_buffer[i].isa = CS_ETM_ISA_UNKNOWN; in cs_etm__clear_packet_queue() 570 queue->packet_buffer[i].start_addr = CS_ETM_INVAL_ADDR; in cs_etm__clear_packet_queue() 571 queue->packet_buffer[i].end_addr = CS_ETM_INVAL_ADDR; in cs_etm__clear_packet_queue() 572 queue->packet_buffer[i].instr_count = 0; in cs_etm__clear_packet_queue() 573 queue->packet_buffer[i].last_instr_taken_branch = false; in cs_etm__clear_packet_queue() 574 queue->packet_buffer[i].last_instr_size = 0; in cs_etm__clear_packet_queue() [all …]
|
| D | intel-pt.c | 458 struct auxtrace_queue *queue; in intel_pt_lookahead() local 461 queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; in intel_pt_lookahead() 466 buffer = auxtrace_buffer__next(queue, buffer); in intel_pt_lookahead() 503 struct auxtrace_queue *queue; in intel_pt_get_trace() local 511 queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; in intel_pt_get_trace() 513 buffer = auxtrace_buffer__next(queue, buffer); in intel_pt_get_trace() 1423 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; in intel_pt_first_timestamp() local 1424 struct intel_pt_queue *ptq = queue->priv; in intel_pt_first_timestamp() 1481 struct auxtrace_queue *queue) in intel_pt_set_pid_tid_cpu() argument 1483 struct intel_pt_queue *ptq = queue->priv; in intel_pt_set_pid_tid_cpu() [all …]
|
| D | annotate.c | 764 int max_lines, struct annotation_line *queue, int addr_fmt_width, in annotation_line__print() argument 795 if (queue != NULL) { in annotation_line__print() 796 list_for_each_entry_from(queue, ¬es->src->source, node) { in annotation_line__print() 797 if (queue == al) in annotation_line__print() 799 annotation_line__print(queue, sym, start, evsel, len, in annotation_line__print() 846 if (queue) in annotation_line__print() 1182 struct annotation_line *pos, *queue = NULL; in symbol__annotate_printf() local 1227 if (context && queue == NULL) { in symbol__annotate_printf() 1228 queue = pos; in symbol__annotate_printf() 1234 queue, addr_fmt_width, opts->percent_type); in symbol__annotate_printf() [all …]
|
| D | auxtrace.c | 296 struct auxtrace_queue *queue; in auxtrace_queues__queue_buffer() local 305 queue = &queues->queue_array[idx]; in auxtrace_queues__queue_buffer() 307 if (!queue->set) { in auxtrace_queues__queue_buffer() 308 queue->set = true; in auxtrace_queues__queue_buffer() 309 queue->tid = buffer->tid; in auxtrace_queues__queue_buffer() 310 queue->cpu = buffer->cpu.cpu; in auxtrace_queues__queue_buffer() 315 list_add_tail(&buffer->list, &queue->head); in auxtrace_queues__queue_buffer() 1026 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue, in auxtrace_buffer__next() argument 1030 if (list_is_last(&buffer->list, &queue->head)) in auxtrace_buffer__next() 1035 if (list_empty(&queue->head)) in auxtrace_buffer__next() [all …]
|
| D | auxtrace.h | 542 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
|
| /tools/testing/selftests/net/netfilter/ |
| D | nft_queue.sh | 91 ip protocol icmp queue bypass 92 icmpv6 type { "echo-request", "echo-reply" } queue num 1 bypass 104 tcp dport 12345 queue num 2 109 tcp dport 12345 queue num 3 110 tcp sport 23456 queue num 3 181 queue num 600 321 tcp dport 12345 limit rate 1/second burst 1 packets counter queue num 0 325 tcp dport 12345 limit rate 1/second burst 1 packets counter queue num 0 366 meta oifname "tvrf" icmp type echo-request counter queue num 1 367 meta oifname "eth0" icmp type echo-request counter queue num 1 [all …]
|
| D | br_netfilter_queue.sh | 65 ip protocol icmp counter queue num 0 bypass
|
| /tools/perf/util/cs-etm-decoder/ |
| D | cs-etm-decoder.c | 501 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_discontinuity() argument 508 cs_etm_decoder__reset_timestamp(queue); in cs_etm_decoder__buffer_discontinuity() 509 return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_discontinuity() 515 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_exception() argument 521 ret = cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_exception() 526 packet = &queue->packet_buffer[queue->tail]; in cs_etm_decoder__buffer_exception() 534 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_exception_ret() argument 537 return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_exception_ret()
|
| /tools/testing/selftests/net/packetdrill/ |
| D | tcp_zerocopy_fastopen-client.pkt | 22 // Read from error queue: no zerocopy notification 43 // Read from error queue: now has first zerocopy notification
|
| D | tcp_zerocopy_fastopen-server.pkt | 29 // Read from error queue: now has first zerocopy notification
|
| /tools/perf/Documentation/ |
| D | perfconfig.example | 36 queue-size = 0
|
| D | perf-config.txt | 123 queue-size = 0 152 % perf config report.queue-size call-graph.order report.children 511 report.queue-size:: 513 event queue for ordering events. Default is 0, meaning no limit. 674 convert.queue-size:: 675 Limit the size of ordered_events queue, so we could control
|
| /tools/tracing/latency/ |
| D | latency-collector.c | 221 struct queue { struct 231 struct queue printqueue; argument 787 static void init_queue(struct queue *q) in init_queue() 797 static __always_inline int queue_len(const struct queue *q) in queue_len() 805 static __always_inline int queue_nr_free(const struct queue *q) in queue_nr_free() 823 static __always_inline void queue_push_to_back(struct queue *q, in queue_push_to_back() 830 static __always_inline struct entry queue_pop_from_front(struct queue *q) in queue_pop_from_front() 838 static __always_inline void queue_cond_signal(struct queue *q) in queue_cond_signal() 843 static __always_inline void queue_cond_wait(struct queue *q) in queue_cond_wait() 848 static __always_inline int queue_try_to_add_entry(struct queue *q, in queue_try_to_add_entry() [all …]
|
| /tools/testing/selftests/net/ |
| D | toeplitz.c | 132 int queue = rx_hash % cfg_num_queues; in verify_rss() local 134 log_verbose(" rxq %d (cpu %d)", queue, rx_irq_cpus[queue]); in verify_rss() 135 if (rx_irq_cpus[queue] != cpu) { in verify_rss()
|
| /tools/testing/selftests/x86/ |
| D | lam.c | 98 } queue; member 429 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring() 432 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring() 445 cring->queue.cqes = cq_ptr + p.cq_off.cqes; in mmap_io_uring() 482 cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask]; in handle_uring_cq() 540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
|
| /tools/bpf/bpftool/bash-completion/ |
| D | bpftool | 151 printf "queue" 183 _bpftool_get_map_ids_for_type queue 229 _bpftool_get_map_names_for_type queue 661 _bpftool_get_map_ids_for_type queue 667 _bpftool_get_map_ids_for_type queue 679 _bpftool_get_map_names_for_type queue 685 _bpftool_get_map_names_for_type queue 758 queue|stack) 809 queue|stack)
|
| /tools/bpf/bpftool/Documentation/ |
| D | bpftool-map.rst | 57 | | **queue** | **stack** | **sk_storage** | **struct_ops** | **ringbuf** | **inode_storage** 137 Peek next value in the queue or stack. 146 Enqueue *VALUE* into the queue. 149 Dequeue and print value from the queue.
|
| /tools/testing/selftests/bpf/progs/ |
| D | map_ptr_kern.c | 587 struct bpf_queue_stack *queue = (struct bpf_queue_stack *)&m_queue; in check_queue() local 590 VERIFY(check(&queue->map, map, 0, sizeof(__u32), MAX_ENTRIES)); in check_queue()
|