Home
last modified time | relevance | path

Searched refs:pq (Results 1 – 25 of 34) sorted by relevance

12

/drivers/net/wireless/broadcom/brcm80211/brcmutil/
Dutils.c47 struct sk_buff *brcmu_pktq_penq(struct pktq *pq, int prec, in brcmu_pktq_penq() argument
52 if (pktq_full(pq) || pktq_pfull(pq, prec)) in brcmu_pktq_penq()
55 q = &pq->q[prec].skblist; in brcmu_pktq_penq()
57 pq->len++; in brcmu_pktq_penq()
59 if (pq->hi_prec < prec) in brcmu_pktq_penq()
60 pq->hi_prec = (u8) prec; in brcmu_pktq_penq()
66 struct sk_buff *brcmu_pktq_penq_head(struct pktq *pq, int prec, in brcmu_pktq_penq_head() argument
71 if (pktq_full(pq) || pktq_pfull(pq, prec)) in brcmu_pktq_penq_head()
74 q = &pq->q[prec].skblist; in brcmu_pktq_penq_head()
76 pq->len++; in brcmu_pktq_penq_head()
[all …]
/drivers/net/wireless/broadcom/brcm80211/include/
Dbrcmu_utils.h76 static inline int pktq_plen(struct pktq *pq, int prec) in pktq_plen() argument
78 return pq->q[prec].skblist.qlen; in pktq_plen()
81 static inline int pktq_pavail(struct pktq *pq, int prec) in pktq_pavail() argument
83 return pq->q[prec].max - pq->q[prec].skblist.qlen; in pktq_pavail()
86 static inline bool pktq_pfull(struct pktq *pq, int prec) in pktq_pfull() argument
88 return pq->q[prec].skblist.qlen >= pq->q[prec].max; in pktq_pfull()
91 static inline bool pktq_pempty(struct pktq *pq, int prec) in pktq_pempty() argument
93 return skb_queue_empty(&pq->q[prec].skblist); in pktq_pempty()
96 static inline struct sk_buff *pktq_ppeek(struct pktq *pq, int prec) in pktq_ppeek() argument
98 return skb_peek(&pq->q[prec].skblist); in pktq_ppeek()
[all …]
/drivers/infiniband/hw/qib/
Dqib_user_sdma.c81 struct qib_user_sdma_queue *pq; /* which pq this pkt belongs to */ member
185 struct qib_user_sdma_queue *pq = in qib_user_sdma_queue_create() local
189 if (!pq) in qib_user_sdma_queue_create()
192 pq->counter = 0; in qib_user_sdma_queue_create()
193 pq->sent_counter = 0; in qib_user_sdma_queue_create()
194 pq->num_pending = 0; in qib_user_sdma_queue_create()
195 pq->num_sending = 0; in qib_user_sdma_queue_create()
196 pq->added = 0; in qib_user_sdma_queue_create()
197 pq->sdma_rb_node = NULL; in qib_user_sdma_queue_create()
199 INIT_LIST_HEAD(&pq->sent); in qib_user_sdma_queue_create()
[all …]
Dqib_user_sdma.h38 void qib_user_sdma_queue_destroy(struct qib_user_sdma_queue *pq);
41 struct qib_user_sdma_queue *pq,
46 struct qib_user_sdma_queue *pq);
49 struct qib_user_sdma_queue *pq);
51 u32 qib_user_sdma_complete_counter(const struct qib_user_sdma_queue *pq);
52 u32 qib_user_sdma_inflight_counter(struct qib_user_sdma_queue *pq);
Dqib_file_ops.c1570 fd->pq = qib_user_sdma_queue_create(&dd->pcidev->dev, in do_qib_user_sdma_queue_create()
1574 if (!fd->pq) in do_qib_user_sdma_queue_create()
1814 if (fd->pq) { in qib_close()
1815 qib_user_sdma_queue_drain(rcd->ppd, fd->pq); in qib_close()
1816 qib_user_sdma_queue_destroy(fd->pq); in qib_close()
1909 static int qib_sdma_get_inflight(struct qib_user_sdma_queue *pq, in qib_sdma_get_inflight() argument
1912 const u32 val = qib_user_sdma_inflight_counter(pq); in qib_sdma_get_inflight()
1921 struct qib_user_sdma_queue *pq, in qib_sdma_get_complete() argument
1927 if (!pq) in qib_sdma_get_complete()
1930 err = qib_user_sdma_make_progress(ppd, pq); in qib_sdma_get_complete()
[all …]
Dqib.h1105 struct qib_user_sdma_queue *pq; member
1181 (((struct qib_filedata *)(fp)->private_data)->pq)
/drivers/dma/ioat/
Dprep.c56 struct ioat_pq_descriptor *pq = (struct ioat_pq_descriptor *) descs[0]; in pq_set_src() local
60 pq->coef[idx] = coef; in pq_set_src()
66 struct ioat_pq_descriptor *pq = (struct ioat_pq_descriptor *)desc[0]; in pq16_set_src() local
74 pq->coef[idx] = coef; in pq16_set_src()
280 struct ioat_pq_descriptor *pq = desc->pq; in dump_pq_desc_dbg() local
282 struct ioat_raw_descriptor *descs[] = { (void *) pq, (void *) pq_ex }; in dump_pq_desc_dbg()
283 int src_cnt = src_cnt_to_sw(pq->ctl_f.src_cnt); in dump_pq_desc_dbg()
290 (unsigned long long) (pq_ex ? pq_ex->next : pq->next), in dump_pq_desc_dbg()
291 desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op, in dump_pq_desc_dbg()
292 pq->ctl_f.int_en, pq->ctl_f.compl_write, in dump_pq_desc_dbg()
[all …]
Ddma.c510 struct ioat_pq_descriptor *pq = desc->pq; in desc_has_ext() local
512 if (src_cnt_to_sw(pq->ctl_f.src_cnt) > 3) in desc_has_ext()
565 struct ioat_pq_descriptor *pq = desc->pq; in desc_get_errstat() local
568 if (!pq->dwbes_f.wbes) in desc_get_errstat()
573 if (pq->dwbes_f.p_val_err) in desc_get_errstat()
576 if (pq->dwbes_f.q_val_err) in desc_get_errstat()
Ddma.h181 struct ioat_pq_descriptor *pq; member
374 ioat_prep_pq_val(struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src,
/drivers/infiniband/hw/hfi1/
Duser_sdma.c81 static inline void pq_update(struct hfi1_user_sdma_pkt_q *pq);
90 static inline void set_comp_state(struct hfi1_user_sdma_pkt_q *pq,
128 struct hfi1_user_sdma_pkt_q *pq = in defer_packet_queue() local
132 trace_hfi1_usdma_defer(pq, sde, &pq->busy); in defer_packet_queue()
140 xchg(&pq->state, SDMA_PKT_Q_DEFERRED); in defer_packet_queue()
141 if (list_empty(&pq->busy.list)) { in defer_packet_queue()
142 pq->busy.lock = &sde->waitlock; in defer_packet_queue()
143 iowait_get_priority(&pq->busy); in defer_packet_queue()
144 iowait_queue(pkts_sent, &pq->busy, &sde->dmawait); in defer_packet_queue()
155 struct hfi1_user_sdma_pkt_q *pq = in activate_packet_queue() local
[all …]
Dtrace_tx.h659 TP_PROTO(struct hfi1_user_sdma_pkt_q *pq,
662 TP_ARGS(pq, sde, wait),
663 TP_STRUCT__entry(DD_DEV_ENTRY(pq->dd)
664 __field(struct hfi1_user_sdma_pkt_q *, pq)
670 TP_fast_assign(DD_DEV_ASSIGN(pq->dd);
671 __entry->pq = pq;
679 (unsigned long long)__entry->pq,
688 TP_PROTO(struct hfi1_user_sdma_pkt_q *pq,
691 TP_ARGS(pq, wait, reason),
692 TP_STRUCT__entry(DD_DEV_ENTRY(pq->dd)
[all …]
Duser_sdma.h118 hfi1_cdbg(SDMA, "[%u:%u:%u:%u] " fmt, (req)->pq->dd->unit, \
119 (req)->pq->ctxt, (req)->pq->subctxt, (req)->info.comp_idx, \
147 struct hfi1_user_sdma_pkt_q *pq; member
173 struct hfi1_user_sdma_pkt_q *pq ____cacheline_aligned_in_smp;
Dfile_ops.c303 struct hfi1_user_sdma_pkt_q *pq; in hfi1_write_iter() local
312 pq = srcu_dereference(fd->pq, &fd->pq_srcu); in hfi1_write_iter()
313 if (!cq || !pq) { in hfi1_write_iter()
325 if (atomic_read(&pq->n_reqs) == pq->n_max_reqs) { in hfi1_write_iter()
/drivers/media/v4l2-core/
Dv4l2-jpeg.c317 u8 pq, tq, *qk; in jpeg_parse_quantization_tables() local
325 pq = (pq_tq >> 4) & 0xf; in jpeg_parse_quantization_tables()
331 if (pq != 0 && (pq != 1 || precision != 12)) in jpeg_parse_quantization_tables()
341 ret = jpeg_skip(stream, pq ? 128 : 64); in jpeg_parse_quantization_tables()
347 tables[tq].length = pq ? 128 : 64; in jpeg_parse_quantization_tables()
350 len -= pq ? 129 : 65; in jpeg_parse_quantization_tables()
/drivers/zorro/
Dgen-devlist.c18 pq(FILE *f, const char *c) in pq() function
69 pq(devf, c); in main()
90 pq(devf, c); in main()
/drivers/gpu/drm/amd/amdkfd/
Dkfd_kernel_queue.c84 retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq); in kq_initialize()
90 kq->pq_kernel_addr = kq->pq->cpu_ptr; in kq_initialize()
91 kq->pq_gpu_addr = kq->pq->gpu_addr; in kq_initialize()
190 kfd_gtt_sa_free(dev, kq->pq); in kq_initialize()
222 kfd_gtt_sa_free(kq->dev, kq->pq); in kq_uninitialize()
Dkfd_kernel_queue.h71 struct kfd_mem_obj *pq; member
/drivers/dma/
Diop-adma.c671 iop_adma_prep_dma_pq_val(struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src, in iop_adma_prep_dma_pq_val() argument
1083 struct page *pq[IOP_ADMA_NUM_SRC_TEST+2+2]; in iop_adma_pq_zero_sum_self_test() local
1085 struct page **pq_hw = &pq[IOP_ADMA_NUM_SRC_TEST+2]; in iop_adma_pq_zero_sum_self_test()
1101 for (i = 0; i < ARRAY_SIZE(pq); i++) { in iop_adma_pq_zero_sum_self_test()
1102 pq[i] = alloc_page(GFP_KERNEL); in iop_adma_pq_zero_sum_self_test()
1103 if (!pq[i]) { in iop_adma_pq_zero_sum_self_test()
1105 __free_page(pq[i]); in iop_adma_pq_zero_sum_self_test()
1112 pq_sw[i] = page_address(pq[i]); in iop_adma_pq_zero_sum_self_test()
1115 pq_sw[i] = page_address(pq[i]); in iop_adma_pq_zero_sum_self_test()
1116 pq_sw[i+1] = page_address(pq[i+1]); in iop_adma_pq_zero_sum_self_test()
[all …]
Dfsl_raid.c408 struct fsl_re_pq_cdb *pq; in fsl_re_prep_dma_pq() local
470 pq = desc->cdb_addr; in fsl_re_prep_dma_pq()
471 pq->cdb32 = cdb; in fsl_re_prep_dma_pq()
473 p = pq->gfm_q1; in fsl_re_prep_dma_pq()
/drivers/atm/
Dfirestream.c619 static struct FS_QENTRY pq[60]; variable
640 pq[qp].cmd = cmd; in submit_queue()
641 pq[qp].p0 = p1; in submit_queue()
642 pq[qp].p1 = p2; in submit_queue()
643 pq[qp].p2 = p3; in submit_queue()
1941 i, pq[qp].cmd, pq[qp].p0, pq[qp].p1, pq[qp].p2); in firestream_remove_one()
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
Dfwsignal.c959 brcmf_fws_flow_control_check(struct brcmf_fws_info *fws, struct pktq *pq, in brcmf_fws_flow_control_check() argument
968 pq->len <= BRCMF_FWS_FLOWCONTROL_LOWATER) in brcmf_fws_flow_control_check()
972 pq->len >= BRCMF_FWS_FLOWCONTROL_HIWATER) { in brcmf_fws_flow_control_check()
1240 struct pktq *pq; in brcmf_fws_enq() local
1260 pq = &entry->psq; in brcmf_fws_enq()
1261 if (pktq_full(pq) || pktq_pfull(pq, prec)) { in brcmf_fws_enq()
1265 queue = &pq->q[prec].skblist; in brcmf_fws_enq()
1302 pq->len++; in brcmf_fws_enq()
1303 if (pq->hi_prec < prec) in brcmf_fws_enq()
1304 pq->hi_prec = (u8) prec; in brcmf_fws_enq()
/drivers/mtd/ubi/
Dwl.c281 list_for_each_entry(p, &ubi->pq[i], u.list) in in_pq()
305 list_add_tail(&e->u.list, &ubi->pq[pq_tail]); in prot_queue_add()
510 list_for_each_entry_safe(e, tmp, &ubi->pq[ubi->pq_head], u.list) { in serve_prot_queue()
1793 INIT_LIST_HEAD(&ubi->pq[i]);
1945 list_for_each_entry_safe(e, tmp, &ubi->pq[i], u.list) {
Dubi.h613 struct list_head pq[UBI_PROT_QUEUE_LEN]; member
1032 list_for_each_entry((e), &(ubi->pq[(i)]), u.list)
/drivers/gpu/drm/amd/display/dc/dcn10/
Ddcn10_resource.c1395 dc->caps.color.dpp.dgam_rom_caps.pq = 0; in dcn10_resource_construct()
1405 dc->caps.color.dpp.ogam_rom_caps.pq = 0; in dcn10_resource_construct()
1417 dc->caps.color.mpc.ogam_rom_caps.pq = 0; in dcn10_resource_construct()
/drivers/dma/ppc4xx/
Dadma.c2586 struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src, in ppc440spe_adma_prep_dma_pqzero_sum() argument
2600 pdest = pq[0]; in ppc440spe_adma_prep_dma_pqzero_sum()
2605 qdest = pq[1]; in ppc440spe_adma_prep_dma_pqzero_sum()
2769 dma_addr_t pq[2]; in ppc440spe_adma_prep_dma_xor_zero_sum() local
2772 pq[0] = src[0]; in ppc440spe_adma_prep_dma_xor_zero_sum()
2773 pq[1] = 0; in ppc440spe_adma_prep_dma_xor_zero_sum()
2776 tx = ppc440spe_adma_prep_dma_pqzero_sum(chan, pq, &src[1], in ppc440spe_adma_prep_dma_xor_zero_sum()

12