Home
last modified time | relevance | path

Searched refs:queue (Results 1 – 25 of 26) sorted by relevance

12

/tools/testing/selftests/mqueue/
Dmq_open_tests.c59 mqd_t queue = -1; variable
90 if (queue != -1) in shutdown()
91 if (mq_close(queue)) in shutdown()
205 if ((queue = mq_open(queue_path, flags, perms, attr)) == -1) in test_queue()
207 if (mq_getattr(queue, result)) in test_queue()
209 if (mq_close(queue)) in test_queue()
211 queue = -1; in test_queue()
227 if ((queue = mq_open(queue_path, flags, perms, attr)) == -1) in test_queue_fail()
229 if (mq_getattr(queue, result)) in test_queue_fail()
231 if (mq_close(queue)) in test_queue_fail()
[all …]
Dmq_perf_tests.c99 mqd_t queue = -1; variable
192 if (queue != -1) in shutdown()
193 if (mq_close(queue)) in shutdown()
297 queue = mq_open(queue_path, flags, perms, attr); in open_queue()
298 if (queue == -1) in open_queue()
300 if (mq_getattr(queue, &result)) in open_queue()
335 while (mq_send(queue, buff, sizeof(buff), 0) == 0) in cont_thread()
337 mq_receive(queue, buff, sizeof(buff), &priority); in cont_thread()
342 while (mq_receive(queue, buff, MSG_SIZE, &prio_in) == MSG_SIZE)
346 if (mq_send(queue, buff, MSG_SIZE, prio_out)) \
[all …]
/tools/testing/selftests/net/tcp_ao/lib/
Drepair.c42 static void test_sock_checkpoint_queue(int sk, int queue, int qlen, in test_sock_checkpoint_queue() argument
48 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_checkpoint_queue()
127 static void test_sock_restore_seq(int sk, int queue, uint32_t seq) in test_sock_restore_seq() argument
129 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_restore_seq()
136 static void test_sock_restore_queue(int sk, int queue, void *buf, int len) in test_sock_restore_queue() argument
144 if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue))) in test_sock_restore_queue()
/tools/testing/selftests/sched_ext/
Ddsp_local_on.bpf.c17 } queue SEC(".maps");
30 if (bpf_map_push_elem(&queue, &pid, 0)) in BPF_STRUCT_OPS()
39 if (bpf_map_pop_elem(&queue, &pid)) in BPF_STRUCT_OPS()
/tools/perf/util/
Dintel-bts.c167 struct auxtrace_queue *queue, in intel_bts_setup_queue() argument
170 struct intel_bts_queue *btsq = queue->priv; in intel_bts_setup_queue()
172 if (list_empty(&queue->head)) in intel_bts_setup_queue()
179 queue->priv = btsq; in intel_bts_setup_queue()
181 if (queue->cpu != -1) in intel_bts_setup_queue()
182 btsq->cpu = queue->cpu; in intel_bts_setup_queue()
183 btsq->tid = queue->tid; in intel_bts_setup_queue()
192 btsq->buffer = auxtrace_buffer__next(queue, NULL); in intel_bts_setup_queue()
248 static int intel_bts_do_fix_overlap(struct auxtrace_queue *queue, in intel_bts_do_fix_overlap() argument
254 if (b->list.prev == &queue->head) in intel_bts_do_fix_overlap()
[all …]
Darm-spe.c154 struct auxtrace_queue *queue; in arm_spe_get_trace() local
156 queue = &speq->spe->queues.queue_array[speq->queue_nr]; in arm_spe_get_trace()
158 buffer = auxtrace_buffer__next(queue, buffer); in arm_spe_get_trace()
241 struct auxtrace_queue *queue) in arm_spe_set_pid_tid_cpu() argument
243 struct arm_spe_queue *speq = queue->priv; in arm_spe_set_pid_tid_cpu()
251 speq->tid = queue->tid; in arm_spe_set_pid_tid_cpu()
260 if (queue->cpu == -1) in arm_spe_set_pid_tid_cpu()
723 struct auxtrace_queue *queue, in arm_spe__setup_queue() argument
726 struct arm_spe_queue *speq = queue->priv; in arm_spe__setup_queue()
729 if (list_empty(&queue->head) || speq) in arm_spe__setup_queue()
[all …]
Ds390-cpumsf.c698 struct auxtrace_queue *queue; in s390_cpumsf_run_decoder() local
701 queue = &sfq->sf->queues.queue_array[sfq->queue_nr]; in s390_cpumsf_run_decoder()
710 sfq->buffer = buffer = auxtrace_buffer__next(queue, in s390_cpumsf_run_decoder()
800 struct auxtrace_queue *queue, in s390_cpumsf_setup_queue() argument
803 struct s390_cpumsf_queue *sfq = queue->priv; in s390_cpumsf_setup_queue()
805 if (list_empty(&queue->head)) in s390_cpumsf_setup_queue()
812 queue->priv = sfq; in s390_cpumsf_setup_queue()
814 if (queue->cpu != -1) in s390_cpumsf_setup_queue()
815 sfq->cpu = queue->cpu; in s390_cpumsf_setup_queue()
850 struct auxtrace_queue *queue; in s390_cpumsf_process_queues() local
[all …]
Dcs-etm.c561 static void cs_etm__clear_packet_queue(struct cs_etm_packet_queue *queue) in cs_etm__clear_packet_queue() argument
565 queue->head = 0; in cs_etm__clear_packet_queue()
566 queue->tail = 0; in cs_etm__clear_packet_queue()
567 queue->packet_count = 0; in cs_etm__clear_packet_queue()
569 queue->packet_buffer[i].isa = CS_ETM_ISA_UNKNOWN; in cs_etm__clear_packet_queue()
570 queue->packet_buffer[i].start_addr = CS_ETM_INVAL_ADDR; in cs_etm__clear_packet_queue()
571 queue->packet_buffer[i].end_addr = CS_ETM_INVAL_ADDR; in cs_etm__clear_packet_queue()
572 queue->packet_buffer[i].instr_count = 0; in cs_etm__clear_packet_queue()
573 queue->packet_buffer[i].last_instr_taken_branch = false; in cs_etm__clear_packet_queue()
574 queue->packet_buffer[i].last_instr_size = 0; in cs_etm__clear_packet_queue()
[all …]
Dintel-pt.c458 struct auxtrace_queue *queue; in intel_pt_lookahead() local
461 queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; in intel_pt_lookahead()
466 buffer = auxtrace_buffer__next(queue, buffer); in intel_pt_lookahead()
503 struct auxtrace_queue *queue; in intel_pt_get_trace() local
511 queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; in intel_pt_get_trace()
513 buffer = auxtrace_buffer__next(queue, buffer); in intel_pt_get_trace()
1423 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; in intel_pt_first_timestamp() local
1424 struct intel_pt_queue *ptq = queue->priv; in intel_pt_first_timestamp()
1481 struct auxtrace_queue *queue) in intel_pt_set_pid_tid_cpu() argument
1483 struct intel_pt_queue *ptq = queue->priv; in intel_pt_set_pid_tid_cpu()
[all …]
Dannotate.c764 int max_lines, struct annotation_line *queue, int addr_fmt_width, in annotation_line__print() argument
795 if (queue != NULL) { in annotation_line__print()
796 list_for_each_entry_from(queue, &notes->src->source, node) { in annotation_line__print()
797 if (queue == al) in annotation_line__print()
799 annotation_line__print(queue, sym, start, evsel, len, in annotation_line__print()
846 if (queue) in annotation_line__print()
1182 struct annotation_line *pos, *queue = NULL; in symbol__annotate_printf() local
1227 if (context && queue == NULL) { in symbol__annotate_printf()
1228 queue = pos; in symbol__annotate_printf()
1234 queue, addr_fmt_width, opts->percent_type); in symbol__annotate_printf()
[all …]
Dauxtrace.c296 struct auxtrace_queue *queue; in auxtrace_queues__queue_buffer() local
305 queue = &queues->queue_array[idx]; in auxtrace_queues__queue_buffer()
307 if (!queue->set) { in auxtrace_queues__queue_buffer()
308 queue->set = true; in auxtrace_queues__queue_buffer()
309 queue->tid = buffer->tid; in auxtrace_queues__queue_buffer()
310 queue->cpu = buffer->cpu.cpu; in auxtrace_queues__queue_buffer()
315 list_add_tail(&buffer->list, &queue->head); in auxtrace_queues__queue_buffer()
1026 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue, in auxtrace_buffer__next() argument
1030 if (list_is_last(&buffer->list, &queue->head)) in auxtrace_buffer__next()
1035 if (list_empty(&queue->head)) in auxtrace_buffer__next()
[all …]
Dauxtrace.h542 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
/tools/testing/selftests/net/netfilter/
Dnft_queue.sh91 ip protocol icmp queue bypass
92 icmpv6 type { "echo-request", "echo-reply" } queue num 1 bypass
104 tcp dport 12345 queue num 2
109 tcp dport 12345 queue num 3
110 tcp sport 23456 queue num 3
181 queue num 600
321 tcp dport 12345 limit rate 1/second burst 1 packets counter queue num 0
325 tcp dport 12345 limit rate 1/second burst 1 packets counter queue num 0
366 meta oifname "tvrf" icmp type echo-request counter queue num 1
367 meta oifname "eth0" icmp type echo-request counter queue num 1
[all …]
Dbr_netfilter_queue.sh65 ip protocol icmp counter queue num 0 bypass
/tools/perf/util/cs-etm-decoder/
Dcs-etm-decoder.c501 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_discontinuity() argument
508 cs_etm_decoder__reset_timestamp(queue); in cs_etm_decoder__buffer_discontinuity()
509 return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_discontinuity()
515 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_exception() argument
521 ret = cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_exception()
526 packet = &queue->packet_buffer[queue->tail]; in cs_etm_decoder__buffer_exception()
534 struct cs_etm_packet_queue *queue, in cs_etm_decoder__buffer_exception_ret() argument
537 return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id, in cs_etm_decoder__buffer_exception_ret()
/tools/testing/selftests/net/packetdrill/
Dtcp_zerocopy_fastopen-client.pkt22 // Read from error queue: no zerocopy notification
43 // Read from error queue: now has first zerocopy notification
Dtcp_zerocopy_fastopen-server.pkt29 // Read from error queue: now has first zerocopy notification
/tools/perf/Documentation/
Dperfconfig.example36 queue-size = 0
Dperf-config.txt123 queue-size = 0
152 % perf config report.queue-size call-graph.order report.children
511 report.queue-size::
513 event queue for ordering events. Default is 0, meaning no limit.
674 convert.queue-size::
675 Limit the size of ordered_events queue, so we could control
/tools/tracing/latency/
Dlatency-collector.c221 struct queue { struct
231 struct queue printqueue; argument
787 static void init_queue(struct queue *q) in init_queue()
797 static __always_inline int queue_len(const struct queue *q) in queue_len()
805 static __always_inline int queue_nr_free(const struct queue *q) in queue_nr_free()
823 static __always_inline void queue_push_to_back(struct queue *q, in queue_push_to_back()
830 static __always_inline struct entry queue_pop_from_front(struct queue *q) in queue_pop_from_front()
838 static __always_inline void queue_cond_signal(struct queue *q) in queue_cond_signal()
843 static __always_inline void queue_cond_wait(struct queue *q) in queue_cond_wait()
848 static __always_inline int queue_try_to_add_entry(struct queue *q, in queue_try_to_add_entry()
[all …]
/tools/testing/selftests/net/
Dtoeplitz.c132 int queue = rx_hash % cfg_num_queues; in verify_rss() local
134 log_verbose(" rxq %d (cpu %d)", queue, rx_irq_cpus[queue]); in verify_rss()
135 if (rx_irq_cpus[queue] != cpu) { in verify_rss()
/tools/testing/selftests/x86/
Dlam.c98 } queue; member
429 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring()
432 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring()
445 cring->queue.cqes = cq_ptr + p.cq_off.cqes; in mmap_io_uring()
482 cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask]; in handle_uring_cq()
540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
/tools/bpf/bpftool/bash-completion/
Dbpftool151 printf "queue"
183 _bpftool_get_map_ids_for_type queue
229 _bpftool_get_map_names_for_type queue
661 _bpftool_get_map_ids_for_type queue
667 _bpftool_get_map_ids_for_type queue
679 _bpftool_get_map_names_for_type queue
685 _bpftool_get_map_names_for_type queue
758 queue|stack)
809 queue|stack)
/tools/bpf/bpftool/Documentation/
Dbpftool-map.rst57 | | **queue** | **stack** | **sk_storage** | **struct_ops** | **ringbuf** | **inode_storage**
137 Peek next value in the queue or stack.
146 Enqueue *VALUE* into the queue.
149 Dequeue and print value from the queue.
/tools/testing/selftests/bpf/progs/
Dmap_ptr_kern.c587 struct bpf_queue_stack *queue = (struct bpf_queue_stack *)&m_queue; in check_queue() local
590 VERIFY(check(&queue->map, map, 0, sizeof(__u32), MAX_ENTRIES)); in check_queue()

12