/drivers/scsi/aic7xxx/ |
D | queue.h | 112 #define SLIST_HEAD_INITIALIZER(head) \ argument 123 #define SLIST_EMPTY(head) ((head)->slh_first == NULL) argument 125 #define SLIST_FIRST(head) ((head)->slh_first) argument 127 #define SLIST_FOREACH(var, head, field) \ argument 128 for ((var) = SLIST_FIRST((head)); \ 132 #define SLIST_INIT(head) do { \ argument 133 SLIST_FIRST((head)) = NULL; \ 141 #define SLIST_INSERT_HEAD(head, elm, field) do { \ argument 142 SLIST_NEXT((elm), field) = SLIST_FIRST((head)); \ 143 SLIST_FIRST((head)) = (elm); \ [all …]
|
/drivers/gpu/drm/nouveau/dispnv04/ |
D | hw.c | 39 NVWriteVgaSeq(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaSeq() argument 41 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVWriteVgaSeq() 42 NVWritePRMVIO(dev, head, NV_PRMVIO_SR, value); in NVWriteVgaSeq() 46 NVReadVgaSeq(struct drm_device *dev, int head, uint8_t index) in NVReadVgaSeq() argument 48 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVReadVgaSeq() 49 return NVReadPRMVIO(dev, head, NV_PRMVIO_SR); in NVReadVgaSeq() 53 NVWriteVgaGr(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaGr() argument 55 NVWritePRMVIO(dev, head, NV_PRMVIO_GRX, index); in NVWriteVgaGr() 56 NVWritePRMVIO(dev, head, NV_PRMVIO_GX, value); in NVWriteVgaGr() 60 NVReadVgaGr(struct drm_device *dev, int head, uint8_t index) in NVReadVgaGr() argument [all …]
|
D | hw.h | 37 void NVWriteVgaSeq(struct drm_device *, int head, uint8_t index, uint8_t value); 38 uint8_t NVReadVgaSeq(struct drm_device *, int head, uint8_t index); 39 void NVWriteVgaGr(struct drm_device *, int head, uint8_t index, uint8_t value); 40 uint8_t NVReadVgaGr(struct drm_device *, int head, uint8_t index); 42 void NVBlankScreen(struct drm_device *, int head, bool blank); 48 void nouveau_hw_save_state(struct drm_device *, int head, 50 void nouveau_hw_load_state(struct drm_device *, int head, 52 void nouveau_hw_load_state_palette(struct drm_device *, int head, 60 int head, uint32_t reg) in NVReadCRTC() argument 64 if (head) in NVReadCRTC() [all …]
|
/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
D | head.c | 34 struct nvkm_head *head; in nvkm_head_find() local 35 list_for_each_entry(head, &disp->head, head) { in nvkm_head_find() 36 if (head->id == id) in nvkm_head_find() 37 return head; in nvkm_head_find() 44 struct nvkm_head *head, void *data, u32 size) in nvkm_head_mthd_scanoutpos() argument 56 head->func->state(head, &head->arm); in nvkm_head_mthd_scanoutpos() 57 args->v0.vtotal = head->arm.vtotal; in nvkm_head_mthd_scanoutpos() 58 args->v0.vblanks = head->arm.vblanks; in nvkm_head_mthd_scanoutpos() 59 args->v0.vblanke = head->arm.vblanke; in nvkm_head_mthd_scanoutpos() 60 args->v0.htotal = head->arm.htotal; in nvkm_head_mthd_scanoutpos() [all …]
|
D | nv50.c | 96 disp->head.nr = func->head.cnt(&disp->base, &disp->head.mask); in nv50_disp_oneinit_() 98 disp->head.nr, disp->head.mask); in nv50_disp_oneinit_() 99 for_each_set_bit(i, &disp->head.mask, disp->head.nr) { in nv50_disp_oneinit_() 100 ret = func->head.new(&disp->base, i); in nv50_disp_oneinit_() 182 nv50_disp_super_iedt(struct nvkm_head *head, struct nvkm_outp *outp, in nv50_disp_super_iedt() argument 186 struct nvkm_bios *bios = head->disp->engine.subdev.device->bios; in nv50_disp_super_iedt() 189 const u16 m = (0x0100 << head->id) | (l << 6) | outp->info.or; in nv50_disp_super_iedt() 197 nv50_disp_super_ied_on(struct nvkm_head *head, in nv50_disp_super_ied_on() argument 200 struct nvkm_subdev *subdev = &head->disp->engine.subdev; in nv50_disp_super_ied_on() 214 data = nv50_disp_super_iedt(head, outp, &ver, &hdr, &cnt, &len, &iedt); in nv50_disp_super_ied_on() [all …]
|
D | gf119.c | 40 struct nvkm_head *head; in gf119_disp_super() local 44 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super() 45 mask[head->id] = nvkm_rd32(device, 0x6101d4 + (head->id * 0x800)); in gf119_disp_super() 46 HEAD_DBG(head, "%08x", mask[head->id]); in gf119_disp_super() 52 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super() 53 if (!(mask[head->id] & 0x00001000)) in gf119_disp_super() 55 nv50_disp_super_1_0(disp, head); in gf119_disp_super() 59 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super() 60 if (!(mask[head->id] & 0x00001000)) in gf119_disp_super() 62 nv50_disp_super_2_0(disp, head); in gf119_disp_super() [all …]
|
D | gv100.c | 46 struct nvkm_head *head; in gv100_disp_super() local 51 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super() 52 mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4)); in gv100_disp_super() 53 HEAD_DBG(head, "%08x", mask[head->id]); in gv100_disp_super() 59 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super() 60 if (!(mask[head->id] & 0x00001000)) in gv100_disp_super() 62 nv50_disp_super_1_0(disp, head); in gv100_disp_super() 66 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super() 67 if (!(mask[head->id] & 0x00001000)) in gv100_disp_super() 69 nv50_disp_super_2_0(disp, head); in gv100_disp_super() [all …]
|
D | vga.c | 27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument 35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport() 41 head = 0; /* CR44 selects head */ in nvkm_rdport() 42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport() 49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument 57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport() 63 head = 0; /* CR44 selects head */ in nvkm_wrport() 64 nvkm_wr08(device, 0x0c0000 + (head * 0x2000) + port, data); in nvkm_wrport() 69 nvkm_rdvgas(struct nvkm_device *device, int head, u8 index) in nvkm_rdvgas() argument 71 nvkm_wrport(device, head, 0x03c4, index); in nvkm_rdvgas() [all …]
|
D | base.c | 46 struct nvkm_head *head = nvkm_head_find(disp, id); in nvkm_disp_vblank_fini() local 47 if (head) in nvkm_disp_vblank_fini() 48 head->func->vblank_put(head); in nvkm_disp_vblank_fini() 55 struct nvkm_head *head = nvkm_head_find(disp, id); in nvkm_disp_vblank_init() local 56 if (head) in nvkm_disp_vblank_init() 57 head->func->vblank_get(head); in nvkm_disp_vblank_init() 73 if (ret = -ENXIO, req->v0.head <= disp->vblank.index_nr) { in nvkm_disp_vblank_ctor() 75 notify->index = req->v0.head; in nvkm_disp_vblank_ctor() 91 nvkm_disp_vblank(struct nvkm_disp *disp, int head) in nvkm_disp_vblank() argument 94 nvkm_event_send(&disp->vblank, 1, head, &rep, sizeof(rep)); in nvkm_disp_vblank() [all …]
|
D | headnv50.c | 27 nv50_head_vblank_put(struct nvkm_head *head) in nv50_head_vblank_put() argument 29 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_vblank_put() 30 nvkm_mask(device, 0x61002c, (4 << head->id), 0); in nv50_head_vblank_put() 34 nv50_head_vblank_get(struct nvkm_head *head) in nv50_head_vblank_get() argument 36 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_vblank_get() 37 nvkm_mask(device, 0x61002c, (4 << head->id), (4 << head->id)); in nv50_head_vblank_get() 41 nv50_head_rgclk(struct nvkm_head *head, int div) in nv50_head_rgclk() argument 43 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_rgclk() 44 nvkm_mask(device, 0x614200 + (head->id * 0x800), 0x0000000f, div); in nv50_head_rgclk() 48 nv50_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline) in nv50_head_rgpos() argument [all …]
|
/drivers/nvme/host/ |
D | multipath.c | 57 if (!ns->head->disk) { in nvme_mpath_set_disk_name() 59 ns->head->instance); in nvme_mpath_set_disk_name() 63 ns->ctrl->instance, ns->head->instance); in nvme_mpath_set_disk_name() 87 spin_lock_irqsave(&ns->head->requeue_lock, flags); in nvme_failover_req() 89 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req() 90 blk_steal_bios(&ns->head->requeue_list, req); in nvme_failover_req() 91 spin_unlock_irqrestore(&ns->head->requeue_lock, flags); in nvme_failover_req() 94 kblockd_schedule_work(&ns->head->requeue_work); in nvme_failover_req() 103 if (ns->head->disk) in nvme_kick_requeue_lists() 104 kblockd_schedule_work(&ns->head->requeue_work); in nvme_kick_requeue_lists() [all …]
|
/drivers/scsi/sym53c8xx_2/ |
D | sym_misc.h | 42 static inline struct sym_quehead *sym_que_first(struct sym_quehead *head) in sym_que_first() argument 44 return (head->flink == head) ? 0 : head->flink; in sym_que_first() 47 static inline struct sym_quehead *sym_que_last(struct sym_quehead *head) in sym_que_last() argument 49 return (head->blink == head) ? 0 : head->blink; in sym_que_last() 69 static inline int sym_que_empty(struct sym_quehead *head) in sym_que_empty() argument 71 return head->flink == head; in sym_que_empty() 75 struct sym_quehead *head) in sym_que_splice() argument 81 struct sym_quehead *at = head->flink; in sym_que_splice() 83 first->blink = head; in sym_que_splice() 84 head->flink = first; in sym_que_splice() [all …]
|
/drivers/rpmsg/ |
D | qcom_glink_smem.c | 40 __le32 *head; member 54 u32 head; in glink_smem_rx_avail() local 70 head = le32_to_cpu(*pipe->head); in glink_smem_rx_avail() 73 if (head < tail) in glink_smem_rx_avail() 74 return pipe->native.length - tail + head; in glink_smem_rx_avail() 76 return head - tail; in glink_smem_rx_avail() 117 u32 head; in glink_smem_tx_avail() local 121 head = le32_to_cpu(*pipe->head); in glink_smem_tx_avail() 124 if (tail <= head) in glink_smem_tx_avail() 125 avail = pipe->native.length - head + tail; in glink_smem_tx_avail() [all …]
|
D | qcom_glink_rpm.c | 51 void __iomem *head; member 59 unsigned int head; in glink_rpm_rx_avail() local 62 head = readl(pipe->head); in glink_rpm_rx_avail() 65 if (head < tail) in glink_rpm_rx_avail() 66 return pipe->native.length - tail + head; in glink_rpm_rx_avail() 68 return head - tail; in glink_rpm_rx_avail() 113 unsigned int head; in glink_rpm_tx_avail() local 116 head = readl(pipe->head); in glink_rpm_tx_avail() 119 if (tail <= head) in glink_rpm_tx_avail() 120 return pipe->native.length - head + tail; in glink_rpm_tx_avail() [all …]
|
/drivers/gpu/drm/nouveau/include/nvif/ |
D | list.h | 159 list_add(struct list_head *entry, struct list_head *head) in list_add() argument 161 __list_add(entry, head, head->next); in list_add() 180 list_add_tail(struct list_head *entry, struct list_head *head) in list_add_tail() argument 182 __list_add(entry, head->prev, head); in list_add_tail() 220 struct list_head *head) in list_move_tail() argument 223 list_add_tail(list, head); in list_move_tail() 235 list_empty(struct list_head *head) in list_empty() argument 237 return head->next == head; in list_empty() 314 #define list_for_each_entry(pos, head, member) \ argument 315 for (pos = __container_of((head)->next, pos, member); \ [all …]
|
/drivers/net/wireless/ath/carl9170/ |
D | fwdesc.h | 113 struct carl9170fw_desc_head head; member 134 struct carl9170fw_desc_head head; member 151 struct carl9170fw_desc_head head; member 160 struct carl9170fw_desc_head head; member 176 struct carl9170fw_desc_head head; member 186 struct carl9170fw_desc_head head; member 196 struct carl9170fw_desc_head head; member 206 struct carl9170fw_desc_head head; member 214 .head = { \ 221 static inline void carl9170fw_fill_desc(struct carl9170fw_desc_head *head, in carl9170fw_fill_desc() argument [all …]
|
/drivers/gpu/drm/nouveau/dispnv50/ |
D | head.c | 40 nv50_head_flush_clr(struct nv50_head *head, in nv50_head_flush_clr() argument 46 if (clr.crc) nv50_crc_atomic_clr(head); in nv50_head_flush_clr() 47 if (clr.olut) head->func->olut_clr(head); in nv50_head_flush_clr() 48 if (clr.core) head->func->core_clr(head); in nv50_head_flush_clr() 49 if (clr.curs) head->func->curs_clr(head); in nv50_head_flush_clr() 53 nv50_head_flush_set_wndw(struct nv50_head *head, struct nv50_head_atom *asyh) in nv50_head_flush_set_wndw() argument 55 if (asyh->set.curs ) head->func->curs_set(head, asyh); in nv50_head_flush_set_wndw() 57 asyh->olut.offset = nv50_lut_load(&head->olut, in nv50_head_flush_set_wndw() 61 head->func->olut_set(head, asyh); in nv50_head_flush_set_wndw() 66 nv50_head_flush_set(struct nv50_head *head, struct nv50_head_atom *asyh) in nv50_head_flush_set() argument [all …]
|
D | crc.c | 70 nv50_crc_program_ctx(struct nv50_head *head, in nv50_crc_program_ctx() argument 73 struct nv50_disp *disp = nv50_disp(head->base.base.dev); in nv50_crc_program_ctx() 77 core->func->crc->set_ctx(head, ctx); in nv50_crc_program_ctx() 85 struct nv50_head *head = container_of(crc, struct nv50_head, crc); in nv50_crc_ctx_flip_work() local 86 struct drm_crtc *crtc = &head->base.base; in nv50_crc_ctx_flip_work() 97 head->base.index); in nv50_crc_ctx_flip_work() 108 nv50_crc_program_ctx(head, NULL); in nv50_crc_ctx_flip_work() 109 nv50_crc_program_ctx(head, &crc->ctx[new_idx]); in nv50_crc_ctx_flip_work() 123 nv50_crc_get_entries(struct nv50_head *head, in nv50_crc_get_entries() argument 127 struct drm_crtc *crtc = &head->base.base; in nv50_crc_get_entries() [all …]
|
D | head507d.c | 30 head507d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_procamp() argument 32 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_procamp() 33 const int i = head->base.index; in head507d_procamp() 49 head507d_dither(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_dither() argument 51 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_dither() 52 const int i = head->base.index; in head507d_dither() 67 head507d_ovly(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_ovly() argument 69 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_ovly() 70 const int i = head->base.index; in head507d_ovly() 95 head507d_base(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_base() argument [all …]
|
D | head907d.c | 37 head907d_or(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_or() argument 39 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_or() 40 const int i = head->base.index; in head907d_or() 52 HEAD_SET_CONTROL(i), 0x31ec6000 | head->base.index << 25 | in head907d_or() 58 head907d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_procamp() argument 60 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_procamp() 61 const int i = head->base.index; in head907d_procamp() 78 head907d_dither(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_dither() argument 80 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_dither() 81 const int i = head->base.index; in head907d_dither() [all …]
|
/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_mbx.c | 16 fifo->head = 0; in fm10k_fifo_init() 28 return fifo->tail - fifo->head; in fm10k_fifo_used() 39 return fifo->size + fifo->head - fifo->tail; in fm10k_fifo_unused() 50 return fifo->head == fifo->tail; in fm10k_fifo_empty() 62 return (fifo->head + offset) & (fifo->size - 1); in fm10k_fifo_head_offset() 85 u32 *head = fifo->buffer + fm10k_fifo_head_offset(fifo, 0); in fm10k_fifo_head_len() local 92 return FM10K_TLV_DWORD_LEN(*head); in fm10k_fifo_head_len() 106 fifo->head += len; in fm10k_fifo_head_drop() 120 fifo->head = fifo->tail; in fm10k_fifo_drop_all() 132 static u16 fm10k_mbx_index_len(struct fm10k_mbx_info *mbx, u16 head, u16 tail) in fm10k_mbx_index_len() argument [all …]
|
/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 32 u32 head; in rvt_cq_enter() local 41 head = RDMA_READ_UAPI_ATOMIC(u_wc->head); in rvt_cq_enter() 46 head = k_wc->head; in rvt_cq_enter() 54 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter() 55 head = cq->ibcq.cqe; in rvt_cq_enter() 58 next = head + 1; in rvt_cq_enter() 78 trace_rvt_cq_enter(cq, entry, head); in rvt_cq_enter() 80 uqueue[head].wr_id = entry->wr_id; in rvt_cq_enter() 81 uqueue[head].status = entry->status; in rvt_cq_enter() 82 uqueue[head].opcode = entry->opcode; in rvt_cq_enter() [all …]
|
/drivers/gpu/drm/ |
D | drm_agpsupport.c | 220 list_add(&entry->head, &dev->agp->memory); in drm_legacy_agp_alloc() 252 list_for_each_entry(entry, &dev->agp->memory, head) { in drm_legacy_agp_lookup_entry() 355 list_del(&entry->head); in drm_legacy_agp_free() 387 struct drm_agp_head *head = NULL; in drm_legacy_agp_init() local 389 head = kzalloc(sizeof(*head), GFP_KERNEL); in drm_legacy_agp_init() 390 if (!head) in drm_legacy_agp_init() 392 head->bridge = agp_find_bridge(pdev); in drm_legacy_agp_init() 393 if (!head->bridge) { in drm_legacy_agp_init() 394 head->bridge = agp_backend_acquire(pdev); in drm_legacy_agp_init() 395 if (!head->bridge) { in drm_legacy_agp_init() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_ras.c | 151 .head = obj->head, in amdgpu_ras_debugfs_read() 246 data->head.block = block_id; in amdgpu_ras_debugfs_ctrl_parse_data() 249 data->head.type = AMDGPU_RAS_ERROR__MULTI_UNCORRECTABLE; in amdgpu_ras_debugfs_ctrl_parse_data() 251 data->head.type = AMDGPU_RAS_ERROR__SINGLE_CORRECTABLE; in amdgpu_ras_debugfs_ctrl_parse_data() 263 data->head.sub_block_index = sub_block; in amdgpu_ras_debugfs_ctrl_parse_data() 383 if (!amdgpu_ras_is_supported(adev, data.head.block)) in amdgpu_ras_debugfs_ctrl_write() 388 ret = amdgpu_ras_feature_enable(adev, &data.head, 0); in amdgpu_ras_debugfs_ctrl_write() 391 ret = amdgpu_ras_feature_enable(adev, &data.head, 1); in amdgpu_ras_debugfs_ctrl_write() 404 if ((data.head.block == AMDGPU_RAS_BLOCK__UMC) && in amdgpu_ras_debugfs_ctrl_write() 503 .head = obj->head, in amdgpu_ras_sysfs_read() [all …]
|
/drivers/usb/core/ |
D | phy.c | 81 struct list_head *head; in usb_phy_roothub_init() local 87 head = &phy_roothub->list; in usb_phy_roothub_init() 89 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_init() 98 list_for_each_entry_continue_reverse(roothub_entry, head, list) in usb_phy_roothub_init() 108 struct list_head *head; in usb_phy_roothub_exit() local 114 head = &phy_roothub->list; in usb_phy_roothub_exit() 116 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_exit() 130 struct list_head *head; in usb_phy_roothub_set_mode() local 136 head = &phy_roothub->list; in usb_phy_roothub_set_mode() 138 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_set_mode() [all …]
|