/drivers/staging/comedi/drivers/ |
D | rtd520.c | 354 int xfer_count; /* # to transfer data. 0->1/2FIFO */ member 645 } else if (devpriv->xfer_count > 0) { in rtd_interrupt() 648 if (ai_read_n(dev, s, devpriv->xfer_count) < 0) in rtd_interrupt() 873 devpriv->xfer_count = cmd->chanlist_len; in rtd_ai_cmd() 877 devpriv->xfer_count = in rtd_ai_cmd() 880 if (devpriv->xfer_count < cmd->chanlist_len) { in rtd_ai_cmd() 882 devpriv->xfer_count = cmd->chanlist_len; in rtd_ai_cmd() 884 devpriv->xfer_count = in rtd_ai_cmd() 885 DIV_ROUND_UP(devpriv->xfer_count, in rtd_ai_cmd() 887 devpriv->xfer_count *= cmd->chanlist_len; in rtd_ai_cmd() [all …]
|
/drivers/input/touchscreen/ |
D | hideep.c | 412 const __be32 *ucode, size_t xfer_count) in hideep_program_page() argument 442 ucode, xfer_count); in hideep_program_page() 444 val = be32_to_cpu(ucode[xfer_count - 1]); in hideep_program_page() 466 size_t xfer_count; in hideep_program_nvm() local 476 xfer_count = xfer_len / sizeof(*ucode); in hideep_program_nvm() 479 current_ucode, xfer_count); in hideep_program_nvm() 490 ucode, xfer_count); in hideep_program_nvm() 501 ucode += xfer_count; in hideep_program_nvm() 515 size_t xfer_count; in hideep_verify_nvm() local 522 xfer_count = xfer_len / sizeof(*ucode); in hideep_verify_nvm() [all …]
|
D | raydium_i2c_ts.c | 150 struct i2c_msg *xfer, size_t xfer_count) in raydium_i2c_xfer() argument 158 xfer_count -= xfer_start_idx; in raydium_i2c_xfer() 160 ret = i2c_transfer(client->adapter, &xfer[xfer_start_idx], xfer_count); in raydium_i2c_xfer() 161 if (likely(ret == xfer_count)) in raydium_i2c_xfer()
|
/drivers/dma/ |
D | txx9dmac.c | 717 size_t xfer_count; in txx9dmac_prep_dma_memcpy() local 730 for (offset = 0; offset < len; offset += xfer_count) { in txx9dmac_prep_dma_memcpy() 731 xfer_count = min_t(size_t, len - offset, TXX9_DMA_MAX_COUNT); in txx9dmac_prep_dma_memcpy() 737 if (xfer_count > 0x100 && in txx9dmac_prep_dma_memcpy() 738 (xfer_count & 0xff) >= 0xfa && in txx9dmac_prep_dma_memcpy() 739 (xfer_count & 0xff) <= 0xff) in txx9dmac_prep_dma_memcpy() 740 xfer_count -= 0x20; in txx9dmac_prep_dma_memcpy() 742 if (xfer_count > 0x80 && in txx9dmac_prep_dma_memcpy() 743 (xfer_count & 0x7f) >= 0x7e && in txx9dmac_prep_dma_memcpy() 744 (xfer_count & 0x7f) <= 0x7f) in txx9dmac_prep_dma_memcpy() [all …]
|
D | at_hdmac.c | 668 size_t xfer_count; in atc_prep_dma_interleaved() local 709 xfer_count = len >> dwidth; in atc_prep_dma_interleaved() 710 if (xfer_count > ATC_BTSIZE_MAX) { in atc_prep_dma_interleaved() 735 desc->lli.ctrla = ctrla | xfer_count; in atc_prep_dma_interleaved() 769 size_t xfer_count; in atc_prep_dma_memcpy() local 798 for (offset = 0; offset < len; offset += xfer_count << src_width) { in atc_prep_dma_memcpy() 799 xfer_count = min_t(size_t, (len - offset) >> src_width, in atc_prep_dma_memcpy() 808 desc->lli.ctrla = ctrla | xfer_count; in atc_prep_dma_memcpy() 812 desc->len = xfer_count << src_width; in atc_prep_dma_memcpy() 840 size_t xfer_count; in atc_create_memset_desc() local [all …]
|
D | stm32-dma.c | 994 size_t xfer_count, offset; in stm32_dma_prep_dma_memcpy() local 1005 for (offset = 0, i = 0; offset < len; offset += xfer_count, i++) { in stm32_dma_prep_dma_memcpy() 1006 xfer_count = min_t(size_t, len - offset, in stm32_dma_prep_dma_memcpy() 1029 desc->sg_req[i].chan_reg.dma_sndtr = xfer_count; in stm32_dma_prep_dma_memcpy() 1030 desc->sg_req[i].len = xfer_count; in stm32_dma_prep_dma_memcpy()
|
D | stm32-mdma.c | 920 size_t xfer_count, offset; in stm32_mdma_prep_dma_memcpy() local 1038 i++, offset += xfer_count) { in stm32_mdma_prep_dma_memcpy() 1039 xfer_count = min_t(size_t, len - offset, in stm32_mdma_prep_dma_memcpy() 1076 xfer_count, ctcr, ctbr, in stm32_mdma_prep_dma_memcpy()
|
/drivers/net/ethernet/8390/ |
D | ne.c | 616 int xfer_count = count; in ne_block_input() local 644 xfer_count++; in ne_block_input() 667 if (((ring_offset + xfer_count) & 0xff) == low) in ne_block_input() 673 ring_offset + xfer_count, addr); in ne_block_input()
|
D | pcnet_cs.c | 1160 int xfer_count = count; in dma_block_input() local 1183 xfer_count++; in dma_block_input() 1198 if (((ring_offset + xfer_count) & 0xff) == (addr & 0xff)) in dma_block_input() 1204 ring_offset + xfer_count, addr); in dma_block_input()
|
D | axnet_cs.c | 650 int xfer_count = count; in block_input() local 662 xfer_count++; in block_input()
|
/drivers/staging/ralink-gdma/ |
D | ralink-gdma.c | 568 size_t xfer_count; in gdma_dma_prep_dma_memcpy() local 575 xfer_count = GDMA_REG_CTRL0_TX_MASK; in gdma_dma_prep_dma_memcpy() 576 num_periods = DIV_ROUND_UP(len, xfer_count); in gdma_dma_prep_dma_memcpy() 588 if (len > xfer_count) in gdma_dma_prep_dma_memcpy() 589 desc->sg[i].len = xfer_count; in gdma_dma_prep_dma_memcpy()
|
/drivers/dma/dw/ |
D | core.c | 548 size_t xfer_count; in dwc_prep_dma_memcpy() local 578 for (offset = 0; offset < len; offset += xfer_count) { in dwc_prep_dma_memcpy() 583 ctlhi = dw->bytes2block(dwc, len - offset, src_width, &xfer_count); in dwc_prep_dma_memcpy() 589 desc->len = xfer_count; in dwc_prep_dma_memcpy()
|
/drivers/usb/dwc2/ |
D | hcd.h | 140 u32 xfer_count; member
|
D | hcd.c | 1140 remaining_count = chan->xfer_len - chan->xfer_count; in dwc2_hc_write_packet() 1161 chan->xfer_count += byte_count; in dwc2_hc_write_packet() 1574 if (chan->xfer_count < chan->xfer_len) { in dwc2_hc_continue_transfer() 2651 chan->xfer_count = 0; in dwc2_assign_and_init_hc() 3871 dev_dbg(hsotg->dev, " xfer_count: %d\n", chan->xfer_count); in dwc2_hcd_dump_state()
|
D | hcd_intr.c | 222 chan->xfer_count += bcnt; in dwc2_rx_fifo_level_intr()
|
/drivers/scsi/smartpqi/ |
D | smartpqi_init.c | 2668 int xfer_count; in pqi_process_raid_io_error() local 2683 xfer_count = in pqi_process_raid_io_error() 2685 residual_count = scsi_bufflen(scmd) - xfer_count; in pqi_process_raid_io_error() 2687 if (xfer_count < scmd->underflow) in pqi_process_raid_io_error() 2769 int xfer_count; in pqi_process_aio_io_error() local 2792 xfer_count = scsi_bufflen(scmd) - residual_count; in pqi_process_aio_io_error() 2793 if (xfer_count < scmd->underflow) in pqi_process_aio_io_error()
|