/drivers/net/wireless/broadcom/brcm80211/brcmutil/ |
D | utils.c | 47 struct sk_buff *brcmu_pktq_penq(struct pktq *pq, int prec, in brcmu_pktq_penq() argument 52 if (pktq_full(pq) || pktq_pfull(pq, prec)) in brcmu_pktq_penq() 55 q = &pq->q[prec].skblist; in brcmu_pktq_penq() 57 pq->len++; in brcmu_pktq_penq() 59 if (pq->hi_prec < prec) in brcmu_pktq_penq() 60 pq->hi_prec = (u8) prec; in brcmu_pktq_penq() 66 struct sk_buff *brcmu_pktq_penq_head(struct pktq *pq, int prec, in brcmu_pktq_penq_head() argument 71 if (pktq_full(pq) || pktq_pfull(pq, prec)) in brcmu_pktq_penq_head() 74 q = &pq->q[prec].skblist; in brcmu_pktq_penq_head() 76 pq->len++; in brcmu_pktq_penq_head() [all …]
|
/drivers/net/wireless/broadcom/brcm80211/include/ |
D | brcmu_utils.h | 76 static inline int pktq_plen(struct pktq *pq, int prec) in pktq_plen() argument 78 return pq->q[prec].skblist.qlen; in pktq_plen() 81 static inline int pktq_pavail(struct pktq *pq, int prec) in pktq_pavail() argument 83 return pq->q[prec].max - pq->q[prec].skblist.qlen; in pktq_pavail() 86 static inline bool pktq_pfull(struct pktq *pq, int prec) in pktq_pfull() argument 88 return pq->q[prec].skblist.qlen >= pq->q[prec].max; in pktq_pfull() 91 static inline bool pktq_pempty(struct pktq *pq, int prec) in pktq_pempty() argument 93 return skb_queue_empty(&pq->q[prec].skblist); in pktq_pempty() 96 static inline struct sk_buff *pktq_ppeek(struct pktq *pq, int prec) in pktq_ppeek() argument 98 return skb_peek(&pq->q[prec].skblist); in pktq_ppeek() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_user_sdma.c | 81 struct qib_user_sdma_queue *pq; /* which pq this pkt belongs to */ member 185 struct qib_user_sdma_queue *pq = in qib_user_sdma_queue_create() local 189 if (!pq) in qib_user_sdma_queue_create() 192 pq->counter = 0; in qib_user_sdma_queue_create() 193 pq->sent_counter = 0; in qib_user_sdma_queue_create() 194 pq->num_pending = 0; in qib_user_sdma_queue_create() 195 pq->num_sending = 0; in qib_user_sdma_queue_create() 196 pq->added = 0; in qib_user_sdma_queue_create() 197 pq->sdma_rb_node = NULL; in qib_user_sdma_queue_create() 199 INIT_LIST_HEAD(&pq->sent); in qib_user_sdma_queue_create() [all …]
|
D | qib_user_sdma.h | 38 void qib_user_sdma_queue_destroy(struct qib_user_sdma_queue *pq); 41 struct qib_user_sdma_queue *pq, 46 struct qib_user_sdma_queue *pq); 49 struct qib_user_sdma_queue *pq); 51 u32 qib_user_sdma_complete_counter(const struct qib_user_sdma_queue *pq); 52 u32 qib_user_sdma_inflight_counter(struct qib_user_sdma_queue *pq);
|
D | qib_file_ops.c | 1570 fd->pq = qib_user_sdma_queue_create(&dd->pcidev->dev, in do_qib_user_sdma_queue_create() 1574 if (!fd->pq) in do_qib_user_sdma_queue_create() 1814 if (fd->pq) { in qib_close() 1815 qib_user_sdma_queue_drain(rcd->ppd, fd->pq); in qib_close() 1816 qib_user_sdma_queue_destroy(fd->pq); in qib_close() 1909 static int qib_sdma_get_inflight(struct qib_user_sdma_queue *pq, in qib_sdma_get_inflight() argument 1912 const u32 val = qib_user_sdma_inflight_counter(pq); in qib_sdma_get_inflight() 1921 struct qib_user_sdma_queue *pq, in qib_sdma_get_complete() argument 1927 if (!pq) in qib_sdma_get_complete() 1930 err = qib_user_sdma_make_progress(ppd, pq); in qib_sdma_get_complete() [all …]
|
D | qib.h | 1105 struct qib_user_sdma_queue *pq; member 1181 (((struct qib_filedata *)(fp)->private_data)->pq)
|
/drivers/dma/ioat/ |
D | prep.c | 56 struct ioat_pq_descriptor *pq = (struct ioat_pq_descriptor *) descs[0]; in pq_set_src() local 60 pq->coef[idx] = coef; in pq_set_src() 66 struct ioat_pq_descriptor *pq = (struct ioat_pq_descriptor *)desc[0]; in pq16_set_src() local 74 pq->coef[idx] = coef; in pq16_set_src() 280 struct ioat_pq_descriptor *pq = desc->pq; in dump_pq_desc_dbg() local 282 struct ioat_raw_descriptor *descs[] = { (void *) pq, (void *) pq_ex }; in dump_pq_desc_dbg() 283 int src_cnt = src_cnt_to_sw(pq->ctl_f.src_cnt); in dump_pq_desc_dbg() 290 (unsigned long long) (pq_ex ? pq_ex->next : pq->next), in dump_pq_desc_dbg() 291 desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op, in dump_pq_desc_dbg() 292 pq->ctl_f.int_en, pq->ctl_f.compl_write, in dump_pq_desc_dbg() [all …]
|
D | dma.c | 510 struct ioat_pq_descriptor *pq = desc->pq; in desc_has_ext() local 512 if (src_cnt_to_sw(pq->ctl_f.src_cnt) > 3) in desc_has_ext() 565 struct ioat_pq_descriptor *pq = desc->pq; in desc_get_errstat() local 568 if (!pq->dwbes_f.wbes) in desc_get_errstat() 573 if (pq->dwbes_f.p_val_err) in desc_get_errstat() 576 if (pq->dwbes_f.q_val_err) in desc_get_errstat()
|
D | dma.h | 181 struct ioat_pq_descriptor *pq; member 374 ioat_prep_pq_val(struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src,
|
/drivers/infiniband/hw/hfi1/ |
D | user_sdma.c | 81 static inline void pq_update(struct hfi1_user_sdma_pkt_q *pq); 90 static inline void set_comp_state(struct hfi1_user_sdma_pkt_q *pq, 128 struct hfi1_user_sdma_pkt_q *pq = in defer_packet_queue() local 132 trace_hfi1_usdma_defer(pq, sde, &pq->busy); in defer_packet_queue() 140 xchg(&pq->state, SDMA_PKT_Q_DEFERRED); in defer_packet_queue() 141 if (list_empty(&pq->busy.list)) { in defer_packet_queue() 142 pq->busy.lock = &sde->waitlock; in defer_packet_queue() 143 iowait_get_priority(&pq->busy); in defer_packet_queue() 144 iowait_queue(pkts_sent, &pq->busy, &sde->dmawait); in defer_packet_queue() 155 struct hfi1_user_sdma_pkt_q *pq = in activate_packet_queue() local [all …]
|
D | trace_tx.h | 659 TP_PROTO(struct hfi1_user_sdma_pkt_q *pq, 662 TP_ARGS(pq, sde, wait), 663 TP_STRUCT__entry(DD_DEV_ENTRY(pq->dd) 664 __field(struct hfi1_user_sdma_pkt_q *, pq) 670 TP_fast_assign(DD_DEV_ASSIGN(pq->dd); 671 __entry->pq = pq; 679 (unsigned long long)__entry->pq, 688 TP_PROTO(struct hfi1_user_sdma_pkt_q *pq, 691 TP_ARGS(pq, wait, reason), 692 TP_STRUCT__entry(DD_DEV_ENTRY(pq->dd) [all …]
|
D | user_sdma.h | 118 hfi1_cdbg(SDMA, "[%u:%u:%u:%u] " fmt, (req)->pq->dd->unit, \ 119 (req)->pq->ctxt, (req)->pq->subctxt, (req)->info.comp_idx, \ 147 struct hfi1_user_sdma_pkt_q *pq; member 173 struct hfi1_user_sdma_pkt_q *pq ____cacheline_aligned_in_smp;
|
D | file_ops.c | 303 struct hfi1_user_sdma_pkt_q *pq; in hfi1_write_iter() local 312 pq = srcu_dereference(fd->pq, &fd->pq_srcu); in hfi1_write_iter() 313 if (!cq || !pq) { in hfi1_write_iter() 325 if (atomic_read(&pq->n_reqs) == pq->n_max_reqs) { in hfi1_write_iter()
|
/drivers/media/v4l2-core/ |
D | v4l2-jpeg.c | 317 u8 pq, tq, *qk; in jpeg_parse_quantization_tables() local 325 pq = (pq_tq >> 4) & 0xf; in jpeg_parse_quantization_tables() 331 if (pq != 0 && (pq != 1 || precision != 12)) in jpeg_parse_quantization_tables() 341 ret = jpeg_skip(stream, pq ? 128 : 64); in jpeg_parse_quantization_tables() 347 tables[tq].length = pq ? 128 : 64; in jpeg_parse_quantization_tables() 350 len -= pq ? 129 : 65; in jpeg_parse_quantization_tables()
|
/drivers/zorro/ |
D | gen-devlist.c | 18 pq(FILE *f, const char *c) in pq() function 69 pq(devf, c); in main() 90 pq(devf, c); in main()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_kernel_queue.c | 84 retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq); in kq_initialize() 90 kq->pq_kernel_addr = kq->pq->cpu_ptr; in kq_initialize() 91 kq->pq_gpu_addr = kq->pq->gpu_addr; in kq_initialize() 190 kfd_gtt_sa_free(dev, kq->pq); in kq_initialize() 222 kfd_gtt_sa_free(kq->dev, kq->pq); in kq_uninitialize()
|
D | kfd_kernel_queue.h | 71 struct kfd_mem_obj *pq; member
|
/drivers/dma/ |
D | iop-adma.c | 671 iop_adma_prep_dma_pq_val(struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src, in iop_adma_prep_dma_pq_val() argument 1083 struct page *pq[IOP_ADMA_NUM_SRC_TEST+2+2]; in iop_adma_pq_zero_sum_self_test() local 1085 struct page **pq_hw = &pq[IOP_ADMA_NUM_SRC_TEST+2]; in iop_adma_pq_zero_sum_self_test() 1101 for (i = 0; i < ARRAY_SIZE(pq); i++) { in iop_adma_pq_zero_sum_self_test() 1102 pq[i] = alloc_page(GFP_KERNEL); in iop_adma_pq_zero_sum_self_test() 1103 if (!pq[i]) { in iop_adma_pq_zero_sum_self_test() 1105 __free_page(pq[i]); in iop_adma_pq_zero_sum_self_test() 1112 pq_sw[i] = page_address(pq[i]); in iop_adma_pq_zero_sum_self_test() 1115 pq_sw[i] = page_address(pq[i]); in iop_adma_pq_zero_sum_self_test() 1116 pq_sw[i+1] = page_address(pq[i+1]); in iop_adma_pq_zero_sum_self_test() [all …]
|
D | fsl_raid.c | 408 struct fsl_re_pq_cdb *pq; in fsl_re_prep_dma_pq() local 470 pq = desc->cdb_addr; in fsl_re_prep_dma_pq() 471 pq->cdb32 = cdb; in fsl_re_prep_dma_pq() 473 p = pq->gfm_q1; in fsl_re_prep_dma_pq()
|
/drivers/atm/ |
D | firestream.c | 619 static struct FS_QENTRY pq[60]; variable 640 pq[qp].cmd = cmd; in submit_queue() 641 pq[qp].p0 = p1; in submit_queue() 642 pq[qp].p1 = p2; in submit_queue() 643 pq[qp].p2 = p3; in submit_queue() 1941 i, pq[qp].cmd, pq[qp].p0, pq[qp].p1, pq[qp].p2); in firestream_remove_one()
|
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | fwsignal.c | 959 brcmf_fws_flow_control_check(struct brcmf_fws_info *fws, struct pktq *pq, in brcmf_fws_flow_control_check() argument 968 pq->len <= BRCMF_FWS_FLOWCONTROL_LOWATER) in brcmf_fws_flow_control_check() 972 pq->len >= BRCMF_FWS_FLOWCONTROL_HIWATER) { in brcmf_fws_flow_control_check() 1240 struct pktq *pq; in brcmf_fws_enq() local 1260 pq = &entry->psq; in brcmf_fws_enq() 1261 if (pktq_full(pq) || pktq_pfull(pq, prec)) { in brcmf_fws_enq() 1265 queue = &pq->q[prec].skblist; in brcmf_fws_enq() 1302 pq->len++; in brcmf_fws_enq() 1303 if (pq->hi_prec < prec) in brcmf_fws_enq() 1304 pq->hi_prec = (u8) prec; in brcmf_fws_enq()
|
/drivers/mtd/ubi/ |
D | wl.c | 281 list_for_each_entry(p, &ubi->pq[i], u.list) in in_pq() 305 list_add_tail(&e->u.list, &ubi->pq[pq_tail]); in prot_queue_add() 510 list_for_each_entry_safe(e, tmp, &ubi->pq[ubi->pq_head], u.list) { in serve_prot_queue() 1793 INIT_LIST_HEAD(&ubi->pq[i]); 1945 list_for_each_entry_safe(e, tmp, &ubi->pq[i], u.list) {
|
D | ubi.h | 613 struct list_head pq[UBI_PROT_QUEUE_LEN]; member 1032 list_for_each_entry((e), &(ubi->pq[(i)]), u.list)
|
/drivers/gpu/drm/amd/display/dc/dcn10/ |
D | dcn10_resource.c | 1395 dc->caps.color.dpp.dgam_rom_caps.pq = 0; in dcn10_resource_construct() 1405 dc->caps.color.dpp.ogam_rom_caps.pq = 0; in dcn10_resource_construct() 1417 dc->caps.color.mpc.ogam_rom_caps.pq = 0; in dcn10_resource_construct()
|
/drivers/dma/ppc4xx/ |
D | adma.c | 2586 struct dma_chan *chan, dma_addr_t *pq, dma_addr_t *src, in ppc440spe_adma_prep_dma_pqzero_sum() argument 2600 pdest = pq[0]; in ppc440spe_adma_prep_dma_pqzero_sum() 2605 qdest = pq[1]; in ppc440spe_adma_prep_dma_pqzero_sum() 2769 dma_addr_t pq[2]; in ppc440spe_adma_prep_dma_xor_zero_sum() local 2772 pq[0] = src[0]; in ppc440spe_adma_prep_dma_xor_zero_sum() 2773 pq[1] = 0; in ppc440spe_adma_prep_dma_xor_zero_sum() 2776 tx = ppc440spe_adma_prep_dma_pqzero_sum(chan, pq, &src[1], in ppc440spe_adma_prep_dma_xor_zero_sum()
|