Home
last modified time | relevance | path

Searched refs:chan (Results 1 – 25 of 1351) sorted by relevance

12345678910>>...55

/drivers/dma/
Dfsldma.c38 #define chan_dbg(chan, fmt, arg...) \ argument
39 dev_dbg(chan->dev, "%s: " fmt, chan->name, ##arg)
40 #define chan_err(chan, fmt, arg...) \ argument
41 dev_err(chan->dev, "%s: " fmt, chan->name, ##arg)
49 static void set_sr(struct fsldma_chan *chan, u32 val) in set_sr() argument
51 FSL_DMA_OUT(chan, &chan->regs->sr, val, 32); in set_sr()
54 static u32 get_sr(struct fsldma_chan *chan) in get_sr() argument
56 return FSL_DMA_IN(chan, &chan->regs->sr, 32); in get_sr()
59 static void set_mr(struct fsldma_chan *chan, u32 val) in set_mr() argument
61 FSL_DMA_OUT(chan, &chan->regs->mr, val, 32); in set_mr()
[all …]
Dstm32-dma.c216 struct stm32_dma_chan chan[STM32_DMA_MAX_CHANNELS]; member
219 static struct stm32_dma_device *stm32_dma_get_dev(struct stm32_dma_chan *chan) in stm32_dma_get_dev() argument
221 return container_of(chan->vchan.chan.device, struct stm32_dma_device, in stm32_dma_get_dev()
227 return container_of(c, struct stm32_dma_chan, vchan.chan); in to_stm32_dma_chan()
235 static struct device *chan2dev(struct stm32_dma_chan *chan) in chan2dev() argument
237 return &chan->vchan.chan.dev->device; in chan2dev()
250 static int stm32_dma_get_width(struct stm32_dma_chan *chan, in stm32_dma_get_width() argument
261 dev_err(chan2dev(chan), "Dma bus width not supported\n"); in stm32_dma_get_width()
344 static int stm32_dma_get_burst(struct stm32_dma_chan *chan, u32 maxburst) in stm32_dma_get_burst() argument
357 dev_err(chan2dev(chan), "Dma burst size not supported\n"); in stm32_dma_get_burst()
[all …]
Ddmaengine.c86 struct dma_chan *chan; in dmaengine_dbg_summary_show() local
88 list_for_each_entry(chan, &dma_dev->channels, device_node) { in dmaengine_dbg_summary_show()
89 if (chan->client_count) { in dmaengine_dbg_summary_show()
90 seq_printf(s, " %-13s| %s", dma_chan_name(chan), in dmaengine_dbg_summary_show()
91 chan->dbg_client_name ?: "in-use"); in dmaengine_dbg_summary_show()
93 if (chan->router) in dmaengine_dbg_summary_show()
95 dev_name(chan->router->dev)); in dmaengine_dbg_summary_show()
159 return chan_dev->chan; in dev_to_dma_chan()
165 struct dma_chan *chan; in memcpy_count_show() local
171 chan = dev_to_dma_chan(dev); in memcpy_count_show()
[all …]
Dnbpfaxi.c155 struct nbpf_channel *chan; member
237 struct nbpf_channel chan[]; member
305 static inline u32 nbpf_chan_read(struct nbpf_channel *chan, in nbpf_chan_read() argument
308 u32 data = ioread32(chan->base + offset); in nbpf_chan_read()
309 dev_dbg(chan->dma_chan.device->dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_chan_read()
310 __func__, chan->base, offset, data); in nbpf_chan_read()
314 static inline void nbpf_chan_write(struct nbpf_channel *chan, in nbpf_chan_write() argument
317 iowrite32(data, chan->base + offset); in nbpf_chan_write()
318 dev_dbg(chan->dma_chan.device->dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_chan_write()
319 __func__, chan->base, offset, data); in nbpf_chan_write()
[all …]
/drivers/gpu/drm/nouveau/
Dnouveau_chan.c51 struct nouveau_channel *chan = container_of(ntfy, typeof(*chan), kill); in nouveau_channel_killed() local
52 struct nouveau_cli *cli = (void *)chan->user.client; in nouveau_channel_killed()
53 NV_PRINTK(warn, cli, "channel %d killed!\n", chan->chid); in nouveau_channel_killed()
54 atomic_set(&chan->killed, 1); in nouveau_channel_killed()
55 if (chan->fence) in nouveau_channel_killed()
56 nouveau_fence_context_kill(chan->fence, -ENODEV); in nouveau_channel_killed()
61 nouveau_channel_idle(struct nouveau_channel *chan) in nouveau_channel_idle() argument
63 if (likely(chan && chan->fence && !atomic_read(&chan->killed))) { in nouveau_channel_idle()
64 struct nouveau_cli *cli = (void *)chan->user.client; in nouveau_channel_idle()
68 ret = nouveau_fence_new(chan, false, &fence); in nouveau_channel_idle()
[all …]
Dnouveau_dma.c41 READ_GET(struct nouveau_channel *chan, uint64_t *prev_get, int *timeout) in READ_GET() argument
45 val = nvif_rd32(&chan->user, chan->user_get); in READ_GET()
46 if (chan->user_get_hi) in READ_GET()
47 val |= (uint64_t)nvif_rd32(&chan->user, chan->user_get_hi) << 32; in READ_GET()
64 if (val < chan->push.addr || in READ_GET()
65 val > chan->push.addr + (chan->dma.max << 2)) in READ_GET()
68 return (val - chan->push.addr) >> 2; in READ_GET()
72 nv50_dma_push(struct nouveau_channel *chan, u64 offset, int length) in nv50_dma_push() argument
74 struct nvif_user *user = &chan->drm->client.device.user; in nv50_dma_push()
75 struct nouveau_bo *pb = chan->push.buffer; in nv50_dma_push()
[all …]
/drivers/dma/xilinx/
Dxilinx_dma.c447 void (*start_transfer)(struct xilinx_dma_chan *chan);
448 int (*stop_transfer)(struct xilinx_dma_chan *chan);
501 struct xilinx_dma_chan *chan[XILINX_MCDMA_MAX_CHANS_PER_DEVICE]; member
517 #define to_xilinx_chan(chan) \ argument
518 container_of(chan, struct xilinx_dma_chan, common)
521 #define xilinx_dma_poll_timeout(chan, reg, val, cond, delay_us, timeout_us) \ argument
522 readl_poll_timeout_atomic(chan->xdev->regs + chan->ctrl_offset + reg, \
526 static inline u32 dma_read(struct xilinx_dma_chan *chan, u32 reg) in dma_read() argument
528 return ioread32(chan->xdev->regs + reg); in dma_read()
531 static inline void dma_write(struct xilinx_dma_chan *chan, u32 reg, u32 value) in dma_write() argument
[all …]
Dzynqmp_dma.c143 #define ZYNQMP_DMA_DESC_SIZE(chan) (chan->desc_size) argument
145 #define to_chan(chan) container_of(chan, struct zynqmp_dma_chan, \ argument
253 struct zynqmp_dma_chan *chan; member
258 static inline void zynqmp_dma_writeq(struct zynqmp_dma_chan *chan, u32 reg, in zynqmp_dma_writeq() argument
261 lo_hi_writeq(value, chan->regs + reg); in zynqmp_dma_writeq()
269 static void zynqmp_dma_update_desc_to_ctrlr(struct zynqmp_dma_chan *chan, in zynqmp_dma_update_desc_to_ctrlr() argument
275 zynqmp_dma_writeq(chan, ZYNQMP_DMA_SRC_START_LSB, addr); in zynqmp_dma_update_desc_to_ctrlr()
277 zynqmp_dma_writeq(chan, ZYNQMP_DMA_DST_START_LSB, addr); in zynqmp_dma_update_desc_to_ctrlr()
285 static void zynqmp_dma_desc_config_eod(struct zynqmp_dma_chan *chan, in zynqmp_dma_desc_config_eod() argument
304 static void zynqmp_dma_config_sg_ll_desc(struct zynqmp_dma_chan *chan, in zynqmp_dma_config_sg_ll_desc() argument
[all …]
Dxilinx_dpdma.c199 struct xilinx_dpdma_chan *chan; member
247 container_of(_chan, struct xilinx_dpdma_chan, vchan.chan)
266 struct xilinx_dpdma_chan *chan[XILINX_DPDMA_NUM_CHAN]; member
300 static void xilinx_dpdma_debugfs_desc_done_irq(struct xilinx_dpdma_chan *chan) in xilinx_dpdma_debugfs_desc_done_irq() argument
302 if (chan->id == dpdma_debugfs.chan_id) in xilinx_dpdma_debugfs_desc_done_irq()
470 static void xilinx_dpdma_debugfs_desc_done_irq(struct xilinx_dpdma_chan *chan) in xilinx_dpdma_debugfs_desc_done_irq() argument
565 xilinx_dpdma_chan_alloc_sw_desc(struct xilinx_dpdma_chan *chan) in xilinx_dpdma_chan_alloc_sw_desc() argument
570 sw_desc = dma_pool_zalloc(chan->desc_pool, GFP_ATOMIC, &dma_addr); in xilinx_dpdma_chan_alloc_sw_desc()
587 xilinx_dpdma_chan_free_sw_desc(struct xilinx_dpdma_chan *chan, in xilinx_dpdma_chan_free_sw_desc() argument
590 dma_pool_free(chan->desc_pool, sw_desc, sw_desc->dma_addr); in xilinx_dpdma_chan_free_sw_desc()
[all …]
/drivers/video/fbdev/savage/
Dsavagefb-i2c.c47 struct savagefb_i2c_chan *chan = data; in savage4_gpio_setscl() local
50 r = readl(chan->ioaddr + chan->reg); in savage4_gpio_setscl()
55 writel(r, chan->ioaddr + chan->reg); in savage4_gpio_setscl()
56 readl(chan->ioaddr + chan->reg); /* flush posted write */ in savage4_gpio_setscl()
61 struct savagefb_i2c_chan *chan = data; in savage4_gpio_setsda() local
64 r = readl(chan->ioaddr + chan->reg); in savage4_gpio_setsda()
69 writel(r, chan->ioaddr + chan->reg); in savage4_gpio_setsda()
70 readl(chan->ioaddr + chan->reg); /* flush posted write */ in savage4_gpio_setsda()
75 struct savagefb_i2c_chan *chan = data; in savage4_gpio_getscl() local
77 return (0 != (readl(chan->ioaddr + chan->reg) & SAVAGE4_I2C_SCL_IN)); in savage4_gpio_getscl()
[all …]
/drivers/mailbox/
Dmailbox.c26 static int add_to_rbuf(struct mbox_chan *chan, void *mssg) in add_to_rbuf() argument
31 spin_lock_irqsave(&chan->lock, flags); in add_to_rbuf()
34 if (chan->msg_count == MBOX_TX_QUEUE_LEN) { in add_to_rbuf()
35 spin_unlock_irqrestore(&chan->lock, flags); in add_to_rbuf()
39 idx = chan->msg_free; in add_to_rbuf()
40 chan->msg_data[idx] = mssg; in add_to_rbuf()
41 chan->msg_count++; in add_to_rbuf()
44 chan->msg_free = 0; in add_to_rbuf()
46 chan->msg_free++; in add_to_rbuf()
48 spin_unlock_irqrestore(&chan->lock, flags); in add_to_rbuf()
[all …]
/drivers/video/fbdev/i810/
Di810-i2c.c44 struct i810fb_i2c_chan *chan = data; in i810i2c_setscl() local
45 struct i810fb_par *par = chan->par; in i810i2c_setscl()
49 i810_writel(mmio, chan->ddc_base, SCL_DIR_MASK | SCL_VAL_MASK); in i810i2c_setscl()
51 i810_writel(mmio, chan->ddc_base, SCL_DIR | SCL_DIR_MASK | SCL_VAL_MASK); in i810i2c_setscl()
52 i810_readl(mmio, chan->ddc_base); /* flush posted write */ in i810i2c_setscl()
57 struct i810fb_i2c_chan *chan = data; in i810i2c_setsda() local
58 struct i810fb_par *par = chan->par; in i810i2c_setsda()
62 i810_writel(mmio, chan->ddc_base, SDA_DIR_MASK | SDA_VAL_MASK); in i810i2c_setsda()
64 i810_writel(mmio, chan->ddc_base, SDA_DIR | SDA_DIR_MASK | SDA_VAL_MASK); in i810i2c_setsda()
65 i810_readl(mmio, chan->ddc_base); /* flush posted write */ in i810i2c_setsda()
[all …]
/drivers/dma/sf-pdma/
Dsf-pdma.c44 return container_of(dchan, struct sf_pdma_chan, vchan.chan); in to_sf_pdma_chan()
52 static struct sf_pdma_desc *sf_pdma_alloc_desc(struct sf_pdma_chan *chan) in sf_pdma_alloc_desc() argument
60 desc->chan = chan; in sf_pdma_alloc_desc()
74 static void sf_pdma_disclaim_chan(struct sf_pdma_chan *chan) in sf_pdma_disclaim_chan() argument
76 struct pdma_regs *regs = &chan->regs; in sf_pdma_disclaim_chan()
85 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_prep_dma_memcpy() local
89 if (chan && (!len || !dest || !src)) { in sf_pdma_prep_dma_memcpy()
90 dev_err(chan->pdma->dma_dev.dev, in sf_pdma_prep_dma_memcpy()
95 desc = sf_pdma_alloc_desc(chan); in sf_pdma_prep_dma_memcpy()
101 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in sf_pdma_prep_dma_memcpy()
[all …]
/drivers/gpu/drm/nouveau/nvkm/engine/gr/
Dnv25.c25 struct nv20_gr_chan *chan; in nv25_gr_chan_new() local
28 if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL))) in nv25_gr_chan_new()
30 nvkm_object_ctor(&nv25_gr_chan, oclass, &chan->object); in nv25_gr_chan_new()
31 chan->gr = gr; in nv25_gr_chan_new()
32 chan->chid = fifoch->chid; in nv25_gr_chan_new()
33 *pobject = &chan->object; in nv25_gr_chan_new()
37 &chan->inst); in nv25_gr_chan_new()
41 nvkm_kmap(chan->inst); in nv25_gr_chan_new()
42 nvkm_wo32(chan->inst, 0x0028, 0x00000001 | (chan->chid << 24)); in nv25_gr_chan_new()
43 nvkm_wo32(chan->inst, 0x035c, 0xffff0000); in nv25_gr_chan_new()
[all …]
Dnv35.c25 struct nv20_gr_chan *chan; in nv35_gr_chan_new() local
28 if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL))) in nv35_gr_chan_new()
30 nvkm_object_ctor(&nv35_gr_chan, oclass, &chan->object); in nv35_gr_chan_new()
31 chan->gr = gr; in nv35_gr_chan_new()
32 chan->chid = fifoch->chid; in nv35_gr_chan_new()
33 *pobject = &chan->object; in nv35_gr_chan_new()
37 &chan->inst); in nv35_gr_chan_new()
41 nvkm_kmap(chan->inst); in nv35_gr_chan_new()
42 nvkm_wo32(chan->inst, 0x0028, 0x00000001 | (chan->chid << 24)); in nv35_gr_chan_new()
43 nvkm_wo32(chan->inst, 0x040c, 0x00000101); in nv35_gr_chan_new()
[all …]
Dnv34.c25 struct nv20_gr_chan *chan; in nv34_gr_chan_new() local
28 if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL))) in nv34_gr_chan_new()
30 nvkm_object_ctor(&nv34_gr_chan, oclass, &chan->object); in nv34_gr_chan_new()
31 chan->gr = gr; in nv34_gr_chan_new()
32 chan->chid = fifoch->chid; in nv34_gr_chan_new()
33 *pobject = &chan->object; in nv34_gr_chan_new()
37 &chan->inst); in nv34_gr_chan_new()
41 nvkm_kmap(chan->inst); in nv34_gr_chan_new()
42 nvkm_wo32(chan->inst, 0x0028, 0x00000001 | (chan->chid << 24)); in nv34_gr_chan_new()
43 nvkm_wo32(chan->inst, 0x040c, 0x01000101); in nv34_gr_chan_new()
[all …]
Dnv2a.c25 struct nv20_gr_chan *chan; in nv2a_gr_chan_new() local
28 if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL))) in nv2a_gr_chan_new()
30 nvkm_object_ctor(&nv2a_gr_chan, oclass, &chan->object); in nv2a_gr_chan_new()
31 chan->gr = gr; in nv2a_gr_chan_new()
32 chan->chid = fifoch->chid; in nv2a_gr_chan_new()
33 *pobject = &chan->object; in nv2a_gr_chan_new()
37 &chan->inst); in nv2a_gr_chan_new()
41 nvkm_kmap(chan->inst); in nv2a_gr_chan_new()
42 nvkm_wo32(chan->inst, 0x0000, 0x00000001 | (chan->chid << 24)); in nv2a_gr_chan_new()
43 nvkm_wo32(chan->inst, 0x033c, 0xffff0000); in nv2a_gr_chan_new()
[all …]
/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dchan.c34 struct nvkm_fifo_chan *chan; member
44 struct nvkm_fifo_chan *chan = object->chan; in nvkm_fifo_chan_child_fini() local
45 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index]; in nvkm_fifo_chan_child_fini()
52 if (chan->func->engine_fini) { in nvkm_fifo_chan_child_fini()
53 ret = chan->func->engine_fini(chan, engine, suspend); in nvkm_fifo_chan_child_fini()
55 nvif_error(&chan->object, in nvkm_fifo_chan_child_fini()
67 nvif_trace(&chan->object, "detached %s\n", name); in nvkm_fifo_chan_child_fini()
77 struct nvkm_fifo_chan *chan = object->chan; in nvkm_fifo_chan_child_init() local
78 struct nvkm_fifo_engn *engn = &chan->engn[engine->subdev.index]; in nvkm_fifo_chan_child_init()
91 if (chan->func->engine_init) { in nvkm_fifo_chan_child_init()
[all …]
/drivers/gpu/drm/gma500/
Doaktrail_lvds_i2c.c63 #define LPC_READ_REG(chan, r) inl((chan)->reg + (r)) argument
64 #define LPC_WRITE_REG(chan, r, val) outl((val), (chan)->reg + (r)) argument
68 struct psb_intel_i2c_chan *chan = data; in get_clock() local
71 val = LPC_READ_REG(chan, RGIO); in get_clock()
73 LPC_WRITE_REG(chan, RGIO, val); in get_clock()
74 tmp = LPC_READ_REG(chan, RGLVL); in get_clock()
75 val = (LPC_READ_REG(chan, RGLVL) & GPIO_CLOCK) ? 1 : 0; in get_clock()
82 struct psb_intel_i2c_chan *chan = data; in get_data() local
85 val = LPC_READ_REG(chan, RGIO); in get_data()
87 LPC_WRITE_REG(chan, RGIO, val); in get_data()
[all …]
Dintel_i2c.c25 struct psb_intel_i2c_chan *chan = data; in get_clock() local
26 struct drm_device *dev = chan->drm_dev; in get_clock()
29 val = REG_READ(chan->reg); in get_clock()
35 struct psb_intel_i2c_chan *chan = data; in get_data() local
36 struct drm_device *dev = chan->drm_dev; in get_data()
39 val = REG_READ(chan->reg); in get_data()
45 struct psb_intel_i2c_chan *chan = data; in set_clock() local
46 struct drm_device *dev = chan->drm_dev; in set_clock()
51 REG_READ(chan->reg) & (GPIO_DATA_PULLUP_DISABLE | in set_clock()
59 REG_WRITE(chan->reg, reserved | clock_bits); in set_clock()
[all …]
/drivers/dma/dw-axi-dmac/
Ddw-axi-dmac-platform.c56 axi_chan_iowrite32(struct axi_dma_chan *chan, u32 reg, u32 val) in axi_chan_iowrite32() argument
58 iowrite32(val, chan->chan_regs + reg); in axi_chan_iowrite32()
61 static inline u32 axi_chan_ioread32(struct axi_dma_chan *chan, u32 reg) in axi_chan_ioread32() argument
63 return ioread32(chan->chan_regs + reg); in axi_chan_ioread32()
67 axi_chan_iowrite64(struct axi_dma_chan *chan, u32 reg, u64 val) in axi_chan_iowrite64() argument
73 iowrite32(lower_32_bits(val), chan->chan_regs + reg); in axi_chan_iowrite64()
74 iowrite32(upper_32_bits(val), chan->chan_regs + reg + 4); in axi_chan_iowrite64()
113 static inline void axi_chan_irq_disable(struct axi_dma_chan *chan, u32 irq_mask) in axi_chan_irq_disable() argument
118 axi_chan_iowrite32(chan, CH_INTSTATUS_ENA, DWAXIDMAC_IRQ_NONE); in axi_chan_irq_disable()
120 val = axi_chan_ioread32(chan, CH_INTSTATUS_ENA); in axi_chan_irq_disable()
[all …]
/drivers/video/fbdev/nvidia/
Dnv_i2c.c30 struct nvidia_i2c_chan *chan = data; in nvidia_gpio_setscl() local
31 struct nvidia_par *par = chan->par; in nvidia_gpio_setscl()
34 val = NVReadCrtc(par, chan->ddc_base + 1) & 0xf0; in nvidia_gpio_setscl()
41 NVWriteCrtc(par, chan->ddc_base + 1, val | 0x01); in nvidia_gpio_setscl()
46 struct nvidia_i2c_chan *chan = data; in nvidia_gpio_setsda() local
47 struct nvidia_par *par = chan->par; in nvidia_gpio_setsda()
50 val = NVReadCrtc(par, chan->ddc_base + 1) & 0xf0; in nvidia_gpio_setsda()
57 NVWriteCrtc(par, chan->ddc_base + 1, val | 0x01); in nvidia_gpio_setsda()
62 struct nvidia_i2c_chan *chan = data; in nvidia_gpio_getscl() local
63 struct nvidia_par *par = chan->par; in nvidia_gpio_getscl()
[all …]
/drivers/video/fbdev/riva/
Drivafb-i2c.c29 struct riva_i2c_chan *chan = data; in riva_gpio_setscl() local
30 struct riva_par *par = chan->par; in riva_gpio_setscl()
33 VGA_WR08(par->riva.PCIO, 0x3d4, chan->ddc_base + 1); in riva_gpio_setscl()
41 VGA_WR08(par->riva.PCIO, 0x3d4, chan->ddc_base + 1); in riva_gpio_setscl()
47 struct riva_i2c_chan *chan = data; in riva_gpio_setsda() local
48 struct riva_par *par = chan->par; in riva_gpio_setsda()
51 VGA_WR08(par->riva.PCIO, 0x3d4, chan->ddc_base + 1); in riva_gpio_setsda()
59 VGA_WR08(par->riva.PCIO, 0x3d4, chan->ddc_base + 1); in riva_gpio_setsda()
65 struct riva_i2c_chan *chan = data; in riva_gpio_getscl() local
66 struct riva_par *par = chan->par; in riva_gpio_getscl()
[all …]
/drivers/staging/media/tegra-video/
Dvi.c107 struct tegra_vi_channel *chan = vb2_get_drv_priv(vq); in tegra_channel_queue_setup() local
110 return sizes[0] < chan->format.sizeimage ? -EINVAL : 0; in tegra_channel_queue_setup()
113 sizes[0] = chan->format.sizeimage; in tegra_channel_queue_setup()
114 alloc_devs[0] = chan->vi->dev; in tegra_channel_queue_setup()
121 struct tegra_vi_channel *chan = vb2_get_drv_priv(vb->vb2_queue); in tegra_channel_buffer_prepare() local
124 unsigned long size = chan->format.sizeimage; in tegra_channel_buffer_prepare()
127 v4l2_err(chan->video.v4l2_dev, in tegra_channel_buffer_prepare()
134 buf->chan = chan; in tegra_channel_buffer_prepare()
142 struct tegra_vi_channel *chan = vb2_get_drv_priv(vb->vb2_queue); in tegra_channel_buffer_queue() local
147 spin_lock(&chan->start_lock); in tegra_channel_buffer_queue()
[all …]
/drivers/net/ethernet/ti/
Ddavinci_cpdma.c138 struct cpdma_chan *chan; member
161 #define tx_chan_num(chan) (chan) argument
162 #define rx_chan_num(chan) ((chan) + CPDMA_MAX_CHANNELS) argument
163 #define is_rx_chan(chan) ((chan)->chan_num >= CPDMA_MAX_CHANNELS) argument
164 #define is_tx_chan(chan) (!is_rx_chan(chan)) argument
166 #define chan_linear(chan) __chan_linear((chan)->chan_num) argument
174 #define chan_read(chan, fld) readl((chan)->fld) argument
177 #define chan_write(chan, fld, v) writel(v, (chan)->fld) argument
180 #define cpdma_desc_to_port(chan, mode, directed) \ argument
182 if (!is_rx_chan(chan) && ((directed == 1) || \
[all …]

12345678910>>...55