Home
last modified time | relevance | path

Searched refs:head (Results 1 – 25 of 1323) sorted by relevance

12345678910>>...53

/drivers/scsi/aic7xxx/
Dqueue.h112 #define SLIST_HEAD_INITIALIZER(head) \ argument
123 #define SLIST_EMPTY(head) ((head)->slh_first == NULL) argument
125 #define SLIST_FIRST(head) ((head)->slh_first) argument
127 #define SLIST_FOREACH(var, head, field) \ argument
128 for ((var) = SLIST_FIRST((head)); \
132 #define SLIST_INIT(head) do { \ argument
133 SLIST_FIRST((head)) = NULL; \
141 #define SLIST_INSERT_HEAD(head, elm, field) do { \ argument
142 SLIST_NEXT((elm), field) = SLIST_FIRST((head)); \
143 SLIST_FIRST((head)) = (elm); \
[all …]
/drivers/gpu/drm/nouveau/dispnv04/
Dhw.c39 NVWriteVgaSeq(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaSeq() argument
41 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVWriteVgaSeq()
42 NVWritePRMVIO(dev, head, NV_PRMVIO_SR, value); in NVWriteVgaSeq()
46 NVReadVgaSeq(struct drm_device *dev, int head, uint8_t index) in NVReadVgaSeq() argument
48 NVWritePRMVIO(dev, head, NV_PRMVIO_SRX, index); in NVReadVgaSeq()
49 return NVReadPRMVIO(dev, head, NV_PRMVIO_SR); in NVReadVgaSeq()
53 NVWriteVgaGr(struct drm_device *dev, int head, uint8_t index, uint8_t value) in NVWriteVgaGr() argument
55 NVWritePRMVIO(dev, head, NV_PRMVIO_GRX, index); in NVWriteVgaGr()
56 NVWritePRMVIO(dev, head, NV_PRMVIO_GX, value); in NVWriteVgaGr()
60 NVReadVgaGr(struct drm_device *dev, int head, uint8_t index) in NVReadVgaGr() argument
[all …]
Dhw.h37 void NVWriteVgaSeq(struct drm_device *, int head, uint8_t index, uint8_t value);
38 uint8_t NVReadVgaSeq(struct drm_device *, int head, uint8_t index);
39 void NVWriteVgaGr(struct drm_device *, int head, uint8_t index, uint8_t value);
40 uint8_t NVReadVgaGr(struct drm_device *, int head, uint8_t index);
42 void NVBlankScreen(struct drm_device *, int head, bool blank);
48 void nouveau_hw_save_state(struct drm_device *, int head,
50 void nouveau_hw_load_state(struct drm_device *, int head,
52 void nouveau_hw_load_state_palette(struct drm_device *, int head,
60 int head, uint32_t reg) in NVReadCRTC() argument
64 if (head) in NVReadCRTC()
[all …]
/drivers/gpu/drm/nouveau/nvkm/engine/disp/
Dhead.c34 struct nvkm_head *head; in nvkm_head_find() local
35 list_for_each_entry(head, &disp->head, head) { in nvkm_head_find()
36 if (head->id == id) in nvkm_head_find()
37 return head; in nvkm_head_find()
44 struct nvkm_head *head, void *data, u32 size) in nvkm_head_mthd_scanoutpos() argument
56 head->func->state(head, &head->arm); in nvkm_head_mthd_scanoutpos()
57 args->v0.vtotal = head->arm.vtotal; in nvkm_head_mthd_scanoutpos()
58 args->v0.vblanks = head->arm.vblanks; in nvkm_head_mthd_scanoutpos()
59 args->v0.vblanke = head->arm.vblanke; in nvkm_head_mthd_scanoutpos()
60 args->v0.htotal = head->arm.htotal; in nvkm_head_mthd_scanoutpos()
[all …]
Dnv50.c96 disp->head.nr = func->head.cnt(&disp->base, &disp->head.mask); in nv50_disp_oneinit_()
98 disp->head.nr, disp->head.mask); in nv50_disp_oneinit_()
99 for_each_set_bit(i, &disp->head.mask, disp->head.nr) { in nv50_disp_oneinit_()
100 ret = func->head.new(&disp->base, i); in nv50_disp_oneinit_()
182 nv50_disp_super_iedt(struct nvkm_head *head, struct nvkm_outp *outp, in nv50_disp_super_iedt() argument
186 struct nvkm_bios *bios = head->disp->engine.subdev.device->bios; in nv50_disp_super_iedt()
189 const u16 m = (0x0100 << head->id) | (l << 6) | outp->info.or; in nv50_disp_super_iedt()
197 nv50_disp_super_ied_on(struct nvkm_head *head, in nv50_disp_super_ied_on() argument
200 struct nvkm_subdev *subdev = &head->disp->engine.subdev; in nv50_disp_super_ied_on()
214 data = nv50_disp_super_iedt(head, outp, &ver, &hdr, &cnt, &len, &iedt); in nv50_disp_super_ied_on()
[all …]
Dgf119.c40 struct nvkm_head *head; in gf119_disp_super() local
44 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super()
45 mask[head->id] = nvkm_rd32(device, 0x6101d4 + (head->id * 0x800)); in gf119_disp_super()
46 HEAD_DBG(head, "%08x", mask[head->id]); in gf119_disp_super()
52 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super()
53 if (!(mask[head->id] & 0x00001000)) in gf119_disp_super()
55 nv50_disp_super_1_0(disp, head); in gf119_disp_super()
59 list_for_each_entry(head, &disp->base.head, head) { in gf119_disp_super()
60 if (!(mask[head->id] & 0x00001000)) in gf119_disp_super()
62 nv50_disp_super_2_0(disp, head); in gf119_disp_super()
[all …]
Dgv100.c46 struct nvkm_head *head; in gv100_disp_super() local
51 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super()
52 mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4)); in gv100_disp_super()
53 HEAD_DBG(head, "%08x", mask[head->id]); in gv100_disp_super()
59 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super()
60 if (!(mask[head->id] & 0x00001000)) in gv100_disp_super()
62 nv50_disp_super_1_0(disp, head); in gv100_disp_super()
66 list_for_each_entry(head, &disp->base.head, head) { in gv100_disp_super()
67 if (!(mask[head->id] & 0x00001000)) in gv100_disp_super()
69 nv50_disp_super_2_0(disp, head); in gv100_disp_super()
[all …]
Dvga.c27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument
35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport()
41 head = 0; /* CR44 selects head */ in nvkm_rdport()
42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport()
49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument
57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport()
63 head = 0; /* CR44 selects head */ in nvkm_wrport()
64 nvkm_wr08(device, 0x0c0000 + (head * 0x2000) + port, data); in nvkm_wrport()
69 nvkm_rdvgas(struct nvkm_device *device, int head, u8 index) in nvkm_rdvgas() argument
71 nvkm_wrport(device, head, 0x03c4, index); in nvkm_rdvgas()
[all …]
Dbase.c46 struct nvkm_head *head = nvkm_head_find(disp, id); in nvkm_disp_vblank_fini() local
47 if (head) in nvkm_disp_vblank_fini()
48 head->func->vblank_put(head); in nvkm_disp_vblank_fini()
55 struct nvkm_head *head = nvkm_head_find(disp, id); in nvkm_disp_vblank_init() local
56 if (head) in nvkm_disp_vblank_init()
57 head->func->vblank_get(head); in nvkm_disp_vblank_init()
73 if (ret = -ENXIO, req->v0.head <= disp->vblank.index_nr) { in nvkm_disp_vblank_ctor()
75 notify->index = req->v0.head; in nvkm_disp_vblank_ctor()
91 nvkm_disp_vblank(struct nvkm_disp *disp, int head) in nvkm_disp_vblank() argument
94 nvkm_event_send(&disp->vblank, 1, head, &rep, sizeof(rep)); in nvkm_disp_vblank()
[all …]
Dheadnv50.c27 nv50_head_vblank_put(struct nvkm_head *head) in nv50_head_vblank_put() argument
29 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_vblank_put()
30 nvkm_mask(device, 0x61002c, (4 << head->id), 0); in nv50_head_vblank_put()
34 nv50_head_vblank_get(struct nvkm_head *head) in nv50_head_vblank_get() argument
36 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_vblank_get()
37 nvkm_mask(device, 0x61002c, (4 << head->id), (4 << head->id)); in nv50_head_vblank_get()
41 nv50_head_rgclk(struct nvkm_head *head, int div) in nv50_head_rgclk() argument
43 struct nvkm_device *device = head->disp->engine.subdev.device; in nv50_head_rgclk()
44 nvkm_mask(device, 0x614200 + (head->id * 0x800), 0x0000000f, div); in nv50_head_rgclk()
48 nv50_head_rgpos(struct nvkm_head *head, u16 *hline, u16 *vline) in nv50_head_rgpos() argument
[all …]
/drivers/nvme/host/
Dmultipath.c57 if (!ns->head->disk) { in nvme_mpath_set_disk_name()
59 ns->head->instance); in nvme_mpath_set_disk_name()
63 ns->ctrl->instance, ns->head->instance); in nvme_mpath_set_disk_name()
87 spin_lock_irqsave(&ns->head->requeue_lock, flags); in nvme_failover_req()
89 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req()
90 blk_steal_bios(&ns->head->requeue_list, req); in nvme_failover_req()
91 spin_unlock_irqrestore(&ns->head->requeue_lock, flags); in nvme_failover_req()
94 kblockd_schedule_work(&ns->head->requeue_work); in nvme_failover_req()
103 if (ns->head->disk) in nvme_kick_requeue_lists()
104 kblockd_schedule_work(&ns->head->requeue_work); in nvme_kick_requeue_lists()
[all …]
/drivers/scsi/sym53c8xx_2/
Dsym_misc.h42 static inline struct sym_quehead *sym_que_first(struct sym_quehead *head) in sym_que_first() argument
44 return (head->flink == head) ? 0 : head->flink; in sym_que_first()
47 static inline struct sym_quehead *sym_que_last(struct sym_quehead *head) in sym_que_last() argument
49 return (head->blink == head) ? 0 : head->blink; in sym_que_last()
69 static inline int sym_que_empty(struct sym_quehead *head) in sym_que_empty() argument
71 return head->flink == head; in sym_que_empty()
75 struct sym_quehead *head) in sym_que_splice() argument
81 struct sym_quehead *at = head->flink; in sym_que_splice()
83 first->blink = head; in sym_que_splice()
84 head->flink = first; in sym_que_splice()
[all …]
/drivers/rpmsg/
Dqcom_glink_smem.c40 __le32 *head; member
54 u32 head; in glink_smem_rx_avail() local
70 head = le32_to_cpu(*pipe->head); in glink_smem_rx_avail()
73 if (head < tail) in glink_smem_rx_avail()
74 return pipe->native.length - tail + head; in glink_smem_rx_avail()
76 return head - tail; in glink_smem_rx_avail()
117 u32 head; in glink_smem_tx_avail() local
121 head = le32_to_cpu(*pipe->head); in glink_smem_tx_avail()
124 if (tail <= head) in glink_smem_tx_avail()
125 avail = pipe->native.length - head + tail; in glink_smem_tx_avail()
[all …]
Dqcom_glink_rpm.c51 void __iomem *head; member
59 unsigned int head; in glink_rpm_rx_avail() local
62 head = readl(pipe->head); in glink_rpm_rx_avail()
65 if (head < tail) in glink_rpm_rx_avail()
66 return pipe->native.length - tail + head; in glink_rpm_rx_avail()
68 return head - tail; in glink_rpm_rx_avail()
113 unsigned int head; in glink_rpm_tx_avail() local
116 head = readl(pipe->head); in glink_rpm_tx_avail()
119 if (tail <= head) in glink_rpm_tx_avail()
120 return pipe->native.length - head + tail; in glink_rpm_tx_avail()
[all …]
/drivers/gpu/drm/nouveau/include/nvif/
Dlist.h159 list_add(struct list_head *entry, struct list_head *head) in list_add() argument
161 __list_add(entry, head, head->next); in list_add()
180 list_add_tail(struct list_head *entry, struct list_head *head) in list_add_tail() argument
182 __list_add(entry, head->prev, head); in list_add_tail()
220 struct list_head *head) in list_move_tail() argument
223 list_add_tail(list, head); in list_move_tail()
235 list_empty(struct list_head *head) in list_empty() argument
237 return head->next == head; in list_empty()
314 #define list_for_each_entry(pos, head, member) \ argument
315 for (pos = __container_of((head)->next, pos, member); \
[all …]
/drivers/net/wireless/ath/carl9170/
Dfwdesc.h113 struct carl9170fw_desc_head head; member
134 struct carl9170fw_desc_head head; member
151 struct carl9170fw_desc_head head; member
160 struct carl9170fw_desc_head head; member
176 struct carl9170fw_desc_head head; member
186 struct carl9170fw_desc_head head; member
196 struct carl9170fw_desc_head head; member
206 struct carl9170fw_desc_head head; member
214 .head = { \
221 static inline void carl9170fw_fill_desc(struct carl9170fw_desc_head *head, in carl9170fw_fill_desc() argument
[all …]
/drivers/gpu/drm/nouveau/dispnv50/
Dhead.c40 nv50_head_flush_clr(struct nv50_head *head, in nv50_head_flush_clr() argument
46 if (clr.crc) nv50_crc_atomic_clr(head); in nv50_head_flush_clr()
47 if (clr.olut) head->func->olut_clr(head); in nv50_head_flush_clr()
48 if (clr.core) head->func->core_clr(head); in nv50_head_flush_clr()
49 if (clr.curs) head->func->curs_clr(head); in nv50_head_flush_clr()
53 nv50_head_flush_set_wndw(struct nv50_head *head, struct nv50_head_atom *asyh) in nv50_head_flush_set_wndw() argument
55 if (asyh->set.curs ) head->func->curs_set(head, asyh); in nv50_head_flush_set_wndw()
57 asyh->olut.offset = nv50_lut_load(&head->olut, in nv50_head_flush_set_wndw()
61 head->func->olut_set(head, asyh); in nv50_head_flush_set_wndw()
66 nv50_head_flush_set(struct nv50_head *head, struct nv50_head_atom *asyh) in nv50_head_flush_set() argument
[all …]
Dcrc.c70 nv50_crc_program_ctx(struct nv50_head *head, in nv50_crc_program_ctx() argument
73 struct nv50_disp *disp = nv50_disp(head->base.base.dev); in nv50_crc_program_ctx()
77 core->func->crc->set_ctx(head, ctx); in nv50_crc_program_ctx()
85 struct nv50_head *head = container_of(crc, struct nv50_head, crc); in nv50_crc_ctx_flip_work() local
86 struct drm_crtc *crtc = &head->base.base; in nv50_crc_ctx_flip_work()
97 head->base.index); in nv50_crc_ctx_flip_work()
108 nv50_crc_program_ctx(head, NULL); in nv50_crc_ctx_flip_work()
109 nv50_crc_program_ctx(head, &crc->ctx[new_idx]); in nv50_crc_ctx_flip_work()
123 nv50_crc_get_entries(struct nv50_head *head, in nv50_crc_get_entries() argument
127 struct drm_crtc *crtc = &head->base.base; in nv50_crc_get_entries()
[all …]
Dhead507d.c30 head507d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_procamp() argument
32 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_procamp()
33 const int i = head->base.index; in head507d_procamp()
49 head507d_dither(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_dither() argument
51 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_dither()
52 const int i = head->base.index; in head507d_dither()
67 head507d_ovly(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_ovly() argument
69 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head507d_ovly()
70 const int i = head->base.index; in head507d_ovly()
95 head507d_base(struct nv50_head *head, struct nv50_head_atom *asyh) in head507d_base() argument
[all …]
Dhead907d.c37 head907d_or(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_or() argument
39 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_or()
40 const int i = head->base.index; in head907d_or()
52 HEAD_SET_CONTROL(i), 0x31ec6000 | head->base.index << 25 | in head907d_or()
58 head907d_procamp(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_procamp() argument
60 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_procamp()
61 const int i = head->base.index; in head907d_procamp()
78 head907d_dither(struct nv50_head *head, struct nv50_head_atom *asyh) in head907d_dither() argument
80 struct nvif_push *push = nv50_disp(head->base.base.dev)->core->chan.push; in head907d_dither()
81 const int i = head->base.index; in head907d_dither()
[all …]
/drivers/net/ethernet/intel/fm10k/
Dfm10k_mbx.c16 fifo->head = 0; in fm10k_fifo_init()
28 return fifo->tail - fifo->head; in fm10k_fifo_used()
39 return fifo->size + fifo->head - fifo->tail; in fm10k_fifo_unused()
50 return fifo->head == fifo->tail; in fm10k_fifo_empty()
62 return (fifo->head + offset) & (fifo->size - 1); in fm10k_fifo_head_offset()
85 u32 *head = fifo->buffer + fm10k_fifo_head_offset(fifo, 0); in fm10k_fifo_head_len() local
92 return FM10K_TLV_DWORD_LEN(*head); in fm10k_fifo_head_len()
106 fifo->head += len; in fm10k_fifo_head_drop()
120 fifo->head = fifo->tail; in fm10k_fifo_drop_all()
132 static u16 fm10k_mbx_index_len(struct fm10k_mbx_info *mbx, u16 head, u16 tail) in fm10k_mbx_index_len() argument
[all …]
/drivers/infiniband/sw/rdmavt/
Dcq.c32 u32 head; in rvt_cq_enter() local
41 head = RDMA_READ_UAPI_ATOMIC(u_wc->head); in rvt_cq_enter()
46 head = k_wc->head; in rvt_cq_enter()
54 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter()
55 head = cq->ibcq.cqe; in rvt_cq_enter()
58 next = head + 1; in rvt_cq_enter()
78 trace_rvt_cq_enter(cq, entry, head); in rvt_cq_enter()
80 uqueue[head].wr_id = entry->wr_id; in rvt_cq_enter()
81 uqueue[head].status = entry->status; in rvt_cq_enter()
82 uqueue[head].opcode = entry->opcode; in rvt_cq_enter()
[all …]
/drivers/gpu/drm/
Ddrm_agpsupport.c220 list_add(&entry->head, &dev->agp->memory); in drm_legacy_agp_alloc()
252 list_for_each_entry(entry, &dev->agp->memory, head) { in drm_legacy_agp_lookup_entry()
355 list_del(&entry->head); in drm_legacy_agp_free()
387 struct drm_agp_head *head = NULL; in drm_legacy_agp_init() local
389 head = kzalloc(sizeof(*head), GFP_KERNEL); in drm_legacy_agp_init()
390 if (!head) in drm_legacy_agp_init()
392 head->bridge = agp_find_bridge(pdev); in drm_legacy_agp_init()
393 if (!head->bridge) { in drm_legacy_agp_init()
394 head->bridge = agp_backend_acquire(pdev); in drm_legacy_agp_init()
395 if (!head->bridge) { in drm_legacy_agp_init()
[all …]
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ras.c151 .head = obj->head, in amdgpu_ras_debugfs_read()
246 data->head.block = block_id; in amdgpu_ras_debugfs_ctrl_parse_data()
249 data->head.type = AMDGPU_RAS_ERROR__MULTI_UNCORRECTABLE; in amdgpu_ras_debugfs_ctrl_parse_data()
251 data->head.type = AMDGPU_RAS_ERROR__SINGLE_CORRECTABLE; in amdgpu_ras_debugfs_ctrl_parse_data()
263 data->head.sub_block_index = sub_block; in amdgpu_ras_debugfs_ctrl_parse_data()
383 if (!amdgpu_ras_is_supported(adev, data.head.block)) in amdgpu_ras_debugfs_ctrl_write()
388 ret = amdgpu_ras_feature_enable(adev, &data.head, 0); in amdgpu_ras_debugfs_ctrl_write()
391 ret = amdgpu_ras_feature_enable(adev, &data.head, 1); in amdgpu_ras_debugfs_ctrl_write()
404 if ((data.head.block == AMDGPU_RAS_BLOCK__UMC) && in amdgpu_ras_debugfs_ctrl_write()
503 .head = obj->head, in amdgpu_ras_sysfs_read()
[all …]
/drivers/usb/core/
Dphy.c81 struct list_head *head; in usb_phy_roothub_init() local
87 head = &phy_roothub->list; in usb_phy_roothub_init()
89 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_init()
98 list_for_each_entry_continue_reverse(roothub_entry, head, list) in usb_phy_roothub_init()
108 struct list_head *head; in usb_phy_roothub_exit() local
114 head = &phy_roothub->list; in usb_phy_roothub_exit()
116 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_exit()
130 struct list_head *head; in usb_phy_roothub_set_mode() local
136 head = &phy_roothub->list; in usb_phy_roothub_set_mode()
138 list_for_each_entry(roothub_entry, head, list) { in usb_phy_roothub_set_mode()
[all …]

12345678910>>...53