/kernel/linux/linux-5.10/crypto/ |
D | gf128mul.c | 56 #define gf128mul_dat(q) { \ argument 57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\ 58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\ 59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\ 60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\ 61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\ 62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\ 63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\ 64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\ 65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\ [all …]
|
/kernel/linux/linux-5.10/Documentation/networking/ |
D | tls-offload-layers.svg | 1 …q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
|
/kernel/linux/linux-5.10/drivers/media/v4l2-core/ |
D | videobuf-core.c | 50 #define CALL(q, f, arg...) \ argument 51 ((q->int_ops->f) ? q->int_ops->f(arg) : 0) 52 #define CALLPTR(q, f, arg...) \ argument 53 ((q->int_ops->f) ? q->int_ops->f(arg) : NULL) 55 struct videobuf_buffer *videobuf_alloc_vb(struct videobuf_queue *q) in videobuf_alloc_vb() argument 59 BUG_ON(q->msize < sizeof(*vb)); in videobuf_alloc_vb() 61 if (!q->int_ops || !q->int_ops->alloc_vb) { in videobuf_alloc_vb() 66 vb = q->int_ops->alloc_vb(q->msize); in videobuf_alloc_vb() 76 static int state_neither_active_nor_queued(struct videobuf_queue *q, in state_neither_active_nor_queued() argument 82 spin_lock_irqsave(q->irqlock, flags); in state_neither_active_nor_queued() [all …]
|
/kernel/linux/linux-5.10/sound/core/seq/ |
D | seq_queue.c | 50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument 58 queue_list[i] = q; in queue_list_add() 59 q->queue = i; in queue_list_add() 71 struct snd_seq_queue *q; in queue_list_remove() local 75 q = queue_list[id]; in queue_list_remove() 76 if (q) { in queue_list_remove() 77 spin_lock(&q->owner_lock); in queue_list_remove() 78 if (q->owner == client) { in queue_list_remove() 80 q->klocked = 1; in queue_list_remove() 81 spin_unlock(&q->owner_lock); in queue_list_remove() [all …]
|
/kernel/linux/linux-5.10/sound/core/seq/oss/ |
D | seq_oss_readq.c | 35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local 37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new() 38 if (!q) in snd_seq_oss_readq_new() 41 q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL); in snd_seq_oss_readq_new() 42 if (!q->q) { in snd_seq_oss_readq_new() 43 kfree(q); in snd_seq_oss_readq_new() 47 q->maxlen = maxlen; in snd_seq_oss_readq_new() 48 q->qlen = 0; in snd_seq_oss_readq_new() 49 q->head = q->tail = 0; in snd_seq_oss_readq_new() 50 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new() [all …]
|
D | seq_oss_writeq.c | 27 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local 30 if ((q = kzalloc(sizeof(*q), GFP_KERNEL)) == NULL) in snd_seq_oss_writeq_new() 32 q->dp = dp; in snd_seq_oss_writeq_new() 33 q->maxlen = maxlen; in snd_seq_oss_writeq_new() 34 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new() 35 q->sync_event_put = 0; in snd_seq_oss_writeq_new() 36 q->sync_time = 0; in snd_seq_oss_writeq_new() 37 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new() 46 return q; in snd_seq_oss_writeq_new() 53 snd_seq_oss_writeq_delete(struct seq_oss_writeq *q) in snd_seq_oss_writeq_delete() argument [all …]
|
/kernel/linux/linux-5.10/net/sched/ |
D | sch_choke.c | 76 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument 78 return (q->tail - q->head) & q->tab_mask; in choke_len() 82 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument 84 return q->flags & TC_RED_ECN; in use_ecn() 88 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument 90 return q->flags & TC_RED_HARDDROP; in use_harddrop() 94 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument 97 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() 98 if (q->head == q->tail) in choke_zap_head_holes() 100 } while (q->tab[q->head] == NULL); in choke_zap_head_holes() [all …]
|
D | sch_netem.c | 200 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument 202 struct clgstate *clg = &q->clg; in loss_4state() 265 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument 267 struct clgstate *clg = &q->clg; in loss_gilb_ell() 286 static bool loss_event(struct netem_sched_data *q) in loss_event() argument 288 switch (q->loss_model) { in loss_event() 291 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event() 299 return loss_4state(q); in loss_event() 307 return loss_gilb_ell(q); in loss_event() 345 static u64 packet_time_ns(u64 len, const struct netem_sched_data *q) in packet_time_ns() argument [all …]
|
D | sch_sfq.c | 150 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument 153 return &q->slots[val].dep; in sfq_dep_head() 154 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head() 157 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument 160 return skb_get_hash_perturb(skb, &q->perturbation) & (q->divisor - 1); in sfq_hash() 166 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local 173 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify() 176 fl = rcu_dereference_bh(q->filter_list); in sfq_classify() 178 return sfq_hash(q, skb) + 1; in sfq_classify() 194 if (TC_H_MIN(res.classid) <= q->divisor) in sfq_classify() [all …]
|
D | sch_sfb.c | 123 static void increment_one_qlen(u32 sfbhash, u32 slot, struct sfb_sched_data *q) in increment_one_qlen() argument 126 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in increment_one_qlen() 138 static void increment_qlen(const struct sfb_skb_cb *cb, struct sfb_sched_data *q) in increment_qlen() argument 144 increment_one_qlen(sfbhash, 0, q); in increment_qlen() 148 increment_one_qlen(sfbhash, 1, q); in increment_qlen() 152 struct sfb_sched_data *q) in decrement_one_qlen() argument 155 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in decrement_one_qlen() 167 static void decrement_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in decrement_qlen() argument 173 decrement_one_qlen(sfbhash, 0, q); in decrement_qlen() 177 decrement_one_qlen(sfbhash, 1, q); in decrement_qlen() [all …]
|
D | sch_fq_pie.c | 73 static unsigned int fq_pie_hash(const struct fq_pie_sched_data *q, in fq_pie_hash() argument 76 return reciprocal_scale(skb_get_hash(skb), q->flows_cnt); in fq_pie_hash() 82 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_classify() local 89 TC_H_MIN(skb->priority) <= q->flows_cnt) in fq_pie_classify() 92 filter = rcu_dereference_bh(q->filter_list); in fq_pie_classify() 94 return fq_pie_hash(q, skb) + 1; in fq_pie_classify() 110 if (TC_H_MIN(res.classid) <= q->flows_cnt) in fq_pie_classify() 131 struct fq_pie_sched_data *q = qdisc_priv(sch); in fq_pie_qdisc_enqueue() local 149 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue() 152 memory_limited = q->memory_usage > q->memory_limit + skb->truesize; in fq_pie_qdisc_enqueue() [all …]
|
D | sch_fq.c | 172 static void fq_flow_unset_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_unset_throttled() argument 174 rb_erase(&f->rate_node, &q->delayed); in fq_flow_unset_throttled() 175 q->throttled_flows--; in fq_flow_unset_throttled() 176 fq_flow_add_tail(&q->old_flows, f); in fq_flow_unset_throttled() 179 static void fq_flow_set_throttled(struct fq_sched_data *q, struct fq_flow *f) in fq_flow_set_throttled() argument 181 struct rb_node **p = &q->delayed.rb_node, *parent = NULL; in fq_flow_set_throttled() 194 rb_insert_color(&f->rate_node, &q->delayed); in fq_flow_set_throttled() 195 q->throttled_flows++; in fq_flow_set_throttled() 196 q->stat_throttled++; in fq_flow_set_throttled() 199 if (q->time_next_delayed_flow > f->time_next_packet) in fq_flow_set_throttled() [all …]
|
/kernel/linux/linux-5.10/net/xdp/ |
D | xsk_queue.h | 112 static inline bool xskq_cons_read_addr_unchecked(struct xsk_queue *q, u64 *addr) in xskq_cons_read_addr_unchecked() argument 114 struct xdp_umem_ring *ring = (struct xdp_umem_ring *)q->ring; in xskq_cons_read_addr_unchecked() 116 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked() 117 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_addr_unchecked() 173 static inline bool xskq_cons_is_valid_desc(struct xsk_queue *q, in xskq_cons_is_valid_desc() argument 178 q->invalid_descs++; in xskq_cons_is_valid_desc() 184 static inline bool xskq_cons_read_desc(struct xsk_queue *q, in xskq_cons_read_desc() argument 188 while (q->cached_cons != q->cached_prod) { in xskq_cons_read_desc() 189 struct xdp_rxtx_ring *ring = (struct xdp_rxtx_ring *)q->ring; in xskq_cons_read_desc() 190 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_desc() [all …]
|
/kernel/linux/linux-5.10/drivers/media/common/videobuf2/ |
D | videobuf2-core.c | 37 #define dprintk(q, level, fmt, arg...) \ argument 40 pr_info("[%s] %s: " fmt, (q)->name, __func__, \ 93 #define log_qop(q, op) \ argument 94 dprintk(q, 2, "call_qop(%s)%s\n", #op, \ 95 (q)->ops->op ? "" : " (nop)") 97 #define call_qop(q, op, args...) \ argument 101 log_qop(q, op); \ 102 err = (q)->ops->op ? (q)->ops->op(args) : 0; \ 104 (q)->cnt_ ## op++; \ 108 #define call_void_qop(q, op, args...) \ argument [all …]
|
/kernel/linux/linux-5.10/block/ |
D | blk-sysfs.c | 61 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument 63 return queue_var_show(q->nr_requests, (page)); in queue_requests_show() 67 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument 72 if (!queue_is_mq(q)) in queue_requests_store() 82 err = blk_mq_update_nr_requests(q, nr); in queue_requests_store() 89 static ssize_t queue_ra_show(struct request_queue *q, char *page) in queue_ra_show() argument 91 unsigned long ra_kb = q->backing_dev_info->ra_pages << in queue_ra_show() 98 queue_ra_store(struct request_queue *q, const char *page, size_t count) in queue_ra_store() argument 106 q->backing_dev_info->ra_pages = ra_kb >> (PAGE_SHIFT - 10); in queue_ra_store() 111 static ssize_t queue_max_sectors_show(struct request_queue *q, char *page) in queue_max_sectors_show() argument [all …]
|
D | blk-pm.c | 31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument 33 q->dev = dev; in blk_pm_runtime_init() 34 q->rpm_status = RPM_ACTIVE; in blk_pm_runtime_init() 35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init() 36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init() 61 int blk_pre_runtime_suspend(struct request_queue *q) in blk_pre_runtime_suspend() argument 65 if (!q->dev) in blk_pre_runtime_suspend() 68 WARN_ON_ONCE(q->rpm_status != RPM_ACTIVE); in blk_pre_runtime_suspend() 70 spin_lock_irq(&q->queue_lock); in blk_pre_runtime_suspend() 71 q->rpm_status = RPM_SUSPENDING; in blk_pre_runtime_suspend() [all …]
|
D | elevator.c | 62 struct request_queue *q = rq->q; in elv_iosched_allow_bio_merge() local 63 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() 66 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge() 140 static struct elevator_type *elevator_get(struct request_queue *q, in elevator_get() argument 147 e = elevator_find(name, q->required_elevator_features); in elevator_get() 152 e = elevator_find(name, q->required_elevator_features); in elevator_get() 164 struct elevator_queue *elevator_alloc(struct request_queue *q, in elevator_alloc() argument 169 eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node); in elevator_alloc() 191 void __elevator_exit(struct request_queue *q, struct elevator_queue *e) in __elevator_exit() argument 194 blk_mq_exit_sched(q, e); in __elevator_exit() [all …]
|
/kernel/linux/linux-5.10/drivers/s390/cio/ |
D | qdio_main.c | 112 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument 115 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs() 118 qperf_inc(q, eqbs); in qdio_do_eqbs() 120 if (!q->is_input_q) in qdio_do_eqbs() 121 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs() 123 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs() 133 qperf_inc(q, eqbs_partial); in qdio_do_eqbs() 134 DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "EQBS part:%02x", in qdio_do_eqbs() 139 DBF_DEV_EVENT(DBF_WARN, q->irq_ptr, "EQBS again:%2d", ccq); in qdio_do_eqbs() 142 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_do_eqbs() [all …]
|
/kernel/linux/linux-5.10/drivers/spi/ |
D | spi-fsl-qspi.c | 278 static inline int needs_swap_endian(struct fsl_qspi *q) in needs_swap_endian() argument 280 return q->devtype_data->quirks & QUADSPI_QUIRK_SWAP_ENDIAN; in needs_swap_endian() 283 static inline int needs_4x_clock(struct fsl_qspi *q) in needs_4x_clock() argument 285 return q->devtype_data->quirks & QUADSPI_QUIRK_4X_INT_CLK; in needs_4x_clock() 288 static inline int needs_fill_txfifo(struct fsl_qspi *q) in needs_fill_txfifo() argument 290 return q->devtype_data->quirks & QUADSPI_QUIRK_TKT253890; in needs_fill_txfifo() 293 static inline int needs_wakeup_wait_mode(struct fsl_qspi *q) in needs_wakeup_wait_mode() argument 295 return q->devtype_data->quirks & QUADSPI_QUIRK_TKT245618; in needs_wakeup_wait_mode() 298 static inline int needs_amba_base_offset(struct fsl_qspi *q) in needs_amba_base_offset() argument 300 return !(q->devtype_data->quirks & QUADSPI_QUIRK_BASE_INTERNAL); in needs_amba_base_offset() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/ |
D | rxe_queue.h | 56 void rxe_queue_reset(struct rxe_queue *q); 62 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p, 72 static inline int next_index(struct rxe_queue *q, int index) in next_index() argument 74 return (index + 1) & q->buf->index_mask; in next_index() 77 static inline int queue_empty(struct rxe_queue *q) in queue_empty() argument 79 return ((q->buf->producer_index - q->buf->consumer_index) in queue_empty() 80 & q->index_mask) == 0; in queue_empty() 83 static inline int queue_full(struct rxe_queue *q) in queue_full() argument 85 return ((q->buf->producer_index + 1 - q->buf->consumer_index) in queue_full() 86 & q->index_mask) == 0; in queue_full() [all …]
|
/kernel/linux/linux-5.10/drivers/misc/habanalabs/common/ |
D | hw_queue.c | 31 static inline int queue_free_slots(struct hl_hw_queue *q, u32 queue_len) in queue_free_slots() argument 33 int delta = (q->pi - queue_ci_get(&q->ci, queue_len)); in queue_free_slots() 44 struct hl_hw_queue *q; in hl_int_hw_queue_update_ci() local 50 q = &hdev->kernel_queues[0]; in hl_int_hw_queue_update_ci() 51 for (i = 0 ; i < hdev->asic_prop.max_queues ; i++, q++) { in hl_int_hw_queue_update_ci() 52 if (q->queue_type == QUEUE_TYPE_INT) in hl_int_hw_queue_update_ci() 53 atomic_add(cs->jobs_in_queue_cnt[i], &q->ci); in hl_int_hw_queue_update_ci() 74 struct hl_hw_queue *q, u32 ctl, u32 len, u64 ptr) in ext_and_hw_queue_submit_bd() argument 78 bd = q->kernel_address; in ext_and_hw_queue_submit_bd() 79 bd += hl_pi_2_offset(q->pi); in ext_and_hw_queue_submit_bd() [all …]
|
/kernel/linux/linux-5.10/drivers/net/wireless/broadcom/b43/ |
D | pio.c | 24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument 37 cookie = (((u16)q->index + 1) << 12); in generate_cookie() 49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local 54 q = pio->tx_queue_AC_BK; in parse_cookie() 57 q = pio->tx_queue_AC_BE; in parse_cookie() 60 q = pio->tx_queue_AC_VI; in parse_cookie() 63 q = pio->tx_queue_AC_VO; in parse_cookie() 66 q = pio->tx_queue_mcast; in parse_cookie() 69 if (B43_WARN_ON(!q)) in parse_cookie() 72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_queue.c | 27 void print_queue_properties(struct queue_properties *q) in print_queue_properties() argument 29 if (!q) in print_queue_properties() 33 pr_debug("Queue Type: %u\n", q->type); in print_queue_properties() 34 pr_debug("Queue Size: %llu\n", q->queue_size); in print_queue_properties() 35 pr_debug("Queue percent: %u\n", q->queue_percent); in print_queue_properties() 36 pr_debug("Queue Address: 0x%llX\n", q->queue_address); in print_queue_properties() 37 pr_debug("Queue Id: %u\n", q->queue_id); in print_queue_properties() 38 pr_debug("Queue Process Vmid: %u\n", q->vmid); in print_queue_properties() 39 pr_debug("Queue Read Pointer: 0x%px\n", q->read_ptr); in print_queue_properties() 40 pr_debug("Queue Write Pointer: 0x%px\n", q->write_ptr); in print_queue_properties() [all …]
|
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/ |
D | dma.c | 83 mt76_dma_alloc_queue(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_alloc_queue() argument 90 spin_lock_init(&q->lock); in mt76_dma_alloc_queue() 92 q->regs = dev->mmio.regs + ring_base + idx * MT_RING_SIZE; in mt76_dma_alloc_queue() 93 q->ndesc = n_desc; in mt76_dma_alloc_queue() 94 q->buf_size = bufsize; in mt76_dma_alloc_queue() 95 q->hw_idx = idx; in mt76_dma_alloc_queue() 97 size = q->ndesc * sizeof(struct mt76_desc); in mt76_dma_alloc_queue() 98 q->desc = dmam_alloc_coherent(dev->dev, size, &q->desc_dma, GFP_KERNEL); in mt76_dma_alloc_queue() 99 if (!q->desc) in mt76_dma_alloc_queue() 102 size = q->ndesc * sizeof(*q->entry); in mt76_dma_alloc_queue() [all …]
|
D | sdio.c | 23 struct mt76_queue *q = &dev->q_rx[qid]; in mt76s_alloc_rx_queue() local 25 spin_lock_init(&q->lock); in mt76s_alloc_rx_queue() 26 q->entry = devm_kcalloc(dev->dev, in mt76s_alloc_rx_queue() 27 MT_NUM_RX_ENTRIES, sizeof(*q->entry), in mt76s_alloc_rx_queue() 29 if (!q->entry) in mt76s_alloc_rx_queue() 32 q->ndesc = MT_NUM_RX_ENTRIES; in mt76s_alloc_rx_queue() 33 q->head = q->tail = 0; in mt76s_alloc_rx_queue() 34 q->queued = 0; in mt76s_alloc_rx_queue() 41 struct mt76_queue *q; in mt76s_alloc_tx() local 45 q = devm_kzalloc(dev->dev, sizeof(*q), GFP_KERNEL); in mt76s_alloc_tx() [all …]
|