Lines Matching +full:dma +full:- +full:maxburst
4 * Copyright (C) 2017, STMicroelectronics - All Rights Reserved
126 * struct stm32_spi - private data of the SPI controller
136 * @cur_midi: master inter-data idleness in ns
142 * @cur_usedma: boolean to know if dma is used in current transfer
147 * @dma_tx: dma channel for TX transfer
148 * @dma_rx: dma channel for RX transfer
182 writel_relaxed(readl_relaxed(spi->base + offset) | bits, in stm32_spi_set_bits()
183 spi->base + offset); in stm32_spi_set_bits()
189 writel_relaxed(readl_relaxed(spi->base + offset) & ~bits, in stm32_spi_clr_bits()
190 spi->base + offset); in stm32_spi_clr_bits()
194 * stm32_spi_get_fifo_size - Return fifo size
202 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_get_fifo_size()
206 while (readl_relaxed(spi->base + STM32_SPI_SR) & SPI_SR_TXP) in stm32_spi_get_fifo_size()
207 writeb_relaxed(++count, spi->base + STM32_SPI_TXDR); in stm32_spi_get_fifo_size()
211 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_get_fifo_size()
213 dev_dbg(spi->dev, "%d x 8-bit fifo size\n", count); in stm32_spi_get_fifo_size()
219 * stm32_spi_get_bpw_mask - Return bits per word mask
227 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_get_bpw_mask()
231 * maximum data size of periperal instances is limited to 16-bit in stm32_spi_get_bpw_mask()
235 cfg1 = readl_relaxed(spi->base + STM32_SPI_CFG1); in stm32_spi_get_bpw_mask()
239 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_get_bpw_mask()
241 dev_dbg(spi->dev, "%d-bit maximum data frame\n", max_bpw); in stm32_spi_get_bpw_mask()
247 * stm32_spi_prepare_mbr - Determine SPI_CFG1.MBR value
251 * Return SPI_CFG1.MBR value in case of success or -EINVAL
257 /* Ensure spi->clk_rate is even */ in stm32_spi_prepare_mbr()
258 div = DIV_ROUND_UP(spi->clk_rate & ~0x1, speed_hz); in stm32_spi_prepare_mbr()
261 * SPI framework set xfer->speed_hz to master->max_speed_hz if in stm32_spi_prepare_mbr()
262 * xfer->speed_hz is greater than master->max_speed_hz, and it returns in stm32_spi_prepare_mbr()
263 * an error when xfer->speed_hz is lower than master->min_speed_hz, so in stm32_spi_prepare_mbr()
269 return -EINVAL; in stm32_spi_prepare_mbr()
272 if (div & (div - 1)) in stm32_spi_prepare_mbr()
275 mbrdiv = fls(div) - 1; in stm32_spi_prepare_mbr()
277 spi->cur_speed = spi->clk_rate / (1 << mbrdiv); in stm32_spi_prepare_mbr()
279 return mbrdiv - 1; in stm32_spi_prepare_mbr()
283 * stm32_spi_prepare_fthlv - Determine FIFO threshold level
291 half_fifo = (spi->fifo_size / 2); in stm32_spi_prepare_fthlv()
293 if (spi->cur_bpw <= 8) in stm32_spi_prepare_fthlv()
295 else if (spi->cur_bpw <= 16) in stm32_spi_prepare_fthlv()
301 if (spi->cur_bpw > 8) in stm32_spi_prepare_fthlv()
302 fthlv -= (fthlv % 2); /* multiple of 2 */ in stm32_spi_prepare_fthlv()
304 fthlv -= (fthlv % 4); /* multiple of 4 */ in stm32_spi_prepare_fthlv()
310 * stm32_spi_write_txfifo - Write bytes in Transmit Data Register
318 while ((spi->tx_len > 0) && in stm32_spi_write_txfifo()
319 (readl_relaxed(spi->base + STM32_SPI_SR) & SPI_SR_TXP)) { in stm32_spi_write_txfifo()
320 u32 offs = spi->cur_xferlen - spi->tx_len; in stm32_spi_write_txfifo()
322 if (spi->tx_len >= sizeof(u32)) { in stm32_spi_write_txfifo()
323 const u32 *tx_buf32 = (const u32 *)(spi->tx_buf + offs); in stm32_spi_write_txfifo()
325 writel_relaxed(*tx_buf32, spi->base + STM32_SPI_TXDR); in stm32_spi_write_txfifo()
326 spi->tx_len -= sizeof(u32); in stm32_spi_write_txfifo()
327 } else if (spi->tx_len >= sizeof(u16)) { in stm32_spi_write_txfifo()
328 const u16 *tx_buf16 = (const u16 *)(spi->tx_buf + offs); in stm32_spi_write_txfifo()
330 writew_relaxed(*tx_buf16, spi->base + STM32_SPI_TXDR); in stm32_spi_write_txfifo()
331 spi->tx_len -= sizeof(u16); in stm32_spi_write_txfifo()
333 const u8 *tx_buf8 = (const u8 *)(spi->tx_buf + offs); in stm32_spi_write_txfifo()
335 writeb_relaxed(*tx_buf8, spi->base + STM32_SPI_TXDR); in stm32_spi_write_txfifo()
336 spi->tx_len -= sizeof(u8); in stm32_spi_write_txfifo()
340 dev_dbg(spi->dev, "%s: %d bytes left\n", __func__, spi->tx_len); in stm32_spi_write_txfifo()
344 * stm32_spi_read_rxfifo - Read bytes in Receive Data Register
352 u32 sr = readl_relaxed(spi->base + STM32_SPI_SR); in stm32_spi_read_rxfifo()
355 while ((spi->rx_len > 0) && in stm32_spi_read_rxfifo()
358 u32 offs = spi->cur_xferlen - spi->rx_len; in stm32_spi_read_rxfifo()
360 if ((spi->rx_len >= sizeof(u32)) || in stm32_spi_read_rxfifo()
362 u32 *rx_buf32 = (u32 *)(spi->rx_buf + offs); in stm32_spi_read_rxfifo()
364 *rx_buf32 = readl_relaxed(spi->base + STM32_SPI_RXDR); in stm32_spi_read_rxfifo()
365 spi->rx_len -= sizeof(u32); in stm32_spi_read_rxfifo()
366 } else if ((spi->rx_len >= sizeof(u16)) || in stm32_spi_read_rxfifo()
367 (flush && (rxplvl >= 2 || spi->cur_bpw > 8))) { in stm32_spi_read_rxfifo()
368 u16 *rx_buf16 = (u16 *)(spi->rx_buf + offs); in stm32_spi_read_rxfifo()
370 *rx_buf16 = readw_relaxed(spi->base + STM32_SPI_RXDR); in stm32_spi_read_rxfifo()
371 spi->rx_len -= sizeof(u16); in stm32_spi_read_rxfifo()
373 u8 *rx_buf8 = (u8 *)(spi->rx_buf + offs); in stm32_spi_read_rxfifo()
375 *rx_buf8 = readb_relaxed(spi->base + STM32_SPI_RXDR); in stm32_spi_read_rxfifo()
376 spi->rx_len -= sizeof(u8); in stm32_spi_read_rxfifo()
379 sr = readl_relaxed(spi->base + STM32_SPI_SR); in stm32_spi_read_rxfifo()
383 dev_dbg(spi->dev, "%s%s: %d bytes left\n", __func__, in stm32_spi_read_rxfifo()
384 flush ? "(flush)" : "", spi->rx_len); in stm32_spi_read_rxfifo()
388 * stm32_spi_enable - Enable SPI controller
396 dev_dbg(spi->dev, "enable controller\n"); in stm32_spi_enable()
402 * stm32_spi_disable - Disable SPI controller
405 * RX-Fifo is flushed when SPI controller is disabled. To prevent any data
407 * RX-Fifo.
414 dev_dbg(spi->dev, "disable controller\n"); in stm32_spi_disable()
416 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_disable()
418 cr1 = readl_relaxed(spi->base + STM32_SPI_CR1); in stm32_spi_disable()
421 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_disable()
426 if (readl_relaxed_poll_timeout_atomic(spi->base + STM32_SPI_SR, in stm32_spi_disable()
431 spi->base + STM32_SPI_CR1); in stm32_spi_disable()
433 spi->base + STM32_SPI_SR, in stm32_spi_disable()
436 dev_warn(spi->dev, in stm32_spi_disable()
441 if (!spi->cur_usedma && spi->rx_buf && (spi->rx_len > 0)) in stm32_spi_disable()
444 if (spi->cur_usedma && spi->tx_buf) in stm32_spi_disable()
445 dmaengine_terminate_all(spi->dma_tx); in stm32_spi_disable()
446 if (spi->cur_usedma && spi->rx_buf) in stm32_spi_disable()
447 dmaengine_terminate_all(spi->dma_rx); in stm32_spi_disable()
455 writel_relaxed(0, spi->base + STM32_SPI_IER); in stm32_spi_disable()
456 writel_relaxed(SPI_IFCR_ALL, spi->base + STM32_SPI_IFCR); in stm32_spi_disable()
458 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_disable()
462 * stm32_spi_can_dma - Determine if the transfer is eligible for DMA use
464 * If the current transfer size is greater than fifo size, use DMA.
472 dev_dbg(spi->dev, "%s: %s\n", __func__, in stm32_spi_can_dma()
473 (transfer->len > spi->fifo_size) ? "true" : "false"); in stm32_spi_can_dma()
475 return (transfer->len > spi->fifo_size); in stm32_spi_can_dma()
479 * stm32_spi_irq - Interrupt handler for SPI controller events
491 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_irq()
493 sr = readl_relaxed(spi->base + STM32_SPI_SR); in stm32_spi_irq()
494 ier = readl_relaxed(spi->base + STM32_SPI_IER); in stm32_spi_irq()
501 * Full-Duplex, need to poll RXP event to know if there are remaining in stm32_spi_irq()
504 if (spi->rx_buf && !spi->cur_usedma) in stm32_spi_irq()
508 dev_dbg(spi->dev, "spurious IT (sr=0x%08x, ier=0x%08x)\n", in stm32_spi_irq()
510 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_irq()
515 dev_warn(spi->dev, "Communication suspended\n"); in stm32_spi_irq()
516 if (!spi->cur_usedma && (spi->rx_buf && (spi->rx_len > 0))) in stm32_spi_irq()
519 * If communication is suspended while using DMA, it means in stm32_spi_irq()
522 if (spi->cur_usedma) in stm32_spi_irq()
527 dev_warn(spi->dev, "Mode fault: transfer aborted\n"); in stm32_spi_irq()
532 dev_warn(spi->dev, "Overrun: received value discarded\n"); in stm32_spi_irq()
533 if (!spi->cur_usedma && (spi->rx_buf && (spi->rx_len > 0))) in stm32_spi_irq()
536 * If overrun is detected while using DMA, it means that in stm32_spi_irq()
539 if (spi->cur_usedma) in stm32_spi_irq()
544 if (!spi->cur_usedma && (spi->rx_buf && (spi->rx_len > 0))) in stm32_spi_irq()
550 if (!spi->cur_usedma && (spi->tx_buf && (spi->tx_len > 0))) in stm32_spi_irq()
554 if (!spi->cur_usedma && (spi->rx_buf && (spi->rx_len > 0))) in stm32_spi_irq()
557 writel_relaxed(mask, spi->base + STM32_SPI_IFCR); in stm32_spi_irq()
559 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_irq()
570 * stm32_spi_setup - setup device chip select
576 if (!gpio_is_valid(spi_dev->cs_gpio)) { in stm32_spi_setup()
577 dev_err(&spi_dev->dev, "%d is not a valid gpio\n", in stm32_spi_setup()
578 spi_dev->cs_gpio); in stm32_spi_setup()
579 return -EINVAL; in stm32_spi_setup()
582 dev_dbg(&spi_dev->dev, "%s: set gpio%d output %s\n", __func__, in stm32_spi_setup()
583 spi_dev->cs_gpio, in stm32_spi_setup()
584 (spi_dev->mode & SPI_CS_HIGH) ? "low" : "high"); in stm32_spi_setup()
586 ret = gpio_direction_output(spi_dev->cs_gpio, in stm32_spi_setup()
587 !(spi_dev->mode & SPI_CS_HIGH)); in stm32_spi_setup()
593 * stm32_spi_prepare_msg - set up the controller to transfer a single message
599 struct spi_device *spi_dev = msg->spi; in stm32_spi_prepare_msg()
600 struct device_node *np = spi_dev->dev.of_node; in stm32_spi_prepare_msg()
605 spi->cur_midi = 0; in stm32_spi_prepare_msg()
606 if (np && !of_property_read_u32(np, "st,spi-midi-ns", &spi->cur_midi)) in stm32_spi_prepare_msg()
607 dev_dbg(spi->dev, "%dns inter-data idleness\n", spi->cur_midi); in stm32_spi_prepare_msg()
609 if (spi_dev->mode & SPI_CPOL) in stm32_spi_prepare_msg()
614 if (spi_dev->mode & SPI_CPHA) in stm32_spi_prepare_msg()
619 if (spi_dev->mode & SPI_LSB_FIRST) in stm32_spi_prepare_msg()
624 dev_dbg(spi->dev, "cpol=%d cpha=%d lsb_first=%d cs_high=%d\n", in stm32_spi_prepare_msg()
625 spi_dev->mode & SPI_CPOL, in stm32_spi_prepare_msg()
626 spi_dev->mode & SPI_CPHA, in stm32_spi_prepare_msg()
627 spi_dev->mode & SPI_LSB_FIRST, in stm32_spi_prepare_msg()
628 spi_dev->mode & SPI_CS_HIGH); in stm32_spi_prepare_msg()
630 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_prepare_msg()
634 (readl_relaxed(spi->base + STM32_SPI_CFG2) & in stm32_spi_prepare_msg()
636 spi->base + STM32_SPI_CFG2); in stm32_spi_prepare_msg()
638 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_prepare_msg()
644 * stm32_spi_dma_cb - dma callback
646 * DMA callback is called when the transfer is complete or when an error
655 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_dma_cb()
657 sr = readl_relaxed(spi->base + STM32_SPI_SR); in stm32_spi_dma_cb()
659 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_dma_cb()
662 dev_warn(spi->dev, "DMA error (sr=0x%08x)\n", sr); in stm32_spi_dma_cb()
668 * stm32_spi_dma_config - configure dma slave channel depending on current
676 u32 maxburst; in stm32_spi_dma_config() local
678 if (spi->cur_bpw <= 8) in stm32_spi_dma_config()
680 else if (spi->cur_bpw <= 16) in stm32_spi_dma_config()
685 /* Valid for DMA Half or Full Fifo threshold */ in stm32_spi_dma_config()
686 if (spi->cur_fthlv == 2) in stm32_spi_dma_config()
687 maxburst = 1; in stm32_spi_dma_config()
689 maxburst = spi->cur_fthlv; in stm32_spi_dma_config()
692 dma_conf->direction = dir; in stm32_spi_dma_config()
693 if (dma_conf->direction == DMA_DEV_TO_MEM) { /* RX */ in stm32_spi_dma_config()
694 dma_conf->src_addr = spi->phys_addr + STM32_SPI_RXDR; in stm32_spi_dma_config()
695 dma_conf->src_addr_width = buswidth; in stm32_spi_dma_config()
696 dma_conf->src_maxburst = maxburst; in stm32_spi_dma_config()
698 dev_dbg(spi->dev, "Rx DMA config buswidth=%d, maxburst=%d\n", in stm32_spi_dma_config()
699 buswidth, maxburst); in stm32_spi_dma_config()
700 } else if (dma_conf->direction == DMA_MEM_TO_DEV) { /* TX */ in stm32_spi_dma_config()
701 dma_conf->dst_addr = spi->phys_addr + STM32_SPI_TXDR; in stm32_spi_dma_config()
702 dma_conf->dst_addr_width = buswidth; in stm32_spi_dma_config()
703 dma_conf->dst_maxburst = maxburst; in stm32_spi_dma_config()
705 dev_dbg(spi->dev, "Tx DMA config buswidth=%d, maxburst=%d\n", in stm32_spi_dma_config()
706 buswidth, maxburst); in stm32_spi_dma_config()
711 * stm32_spi_transfer_one_irq - transfer a single spi_transfer using
723 if (spi->tx_buf && spi->rx_buf) /* Full Duplex */ in stm32_spi_transfer_one_irq()
725 else if (spi->tx_buf) /* Half-Duplex TX dir or Simplex TX */ in stm32_spi_transfer_one_irq()
727 else if (spi->rx_buf) /* Half-Duplex RX dir or Simplex RX */ in stm32_spi_transfer_one_irq()
733 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_transfer_one_irq()
738 if (spi->tx_buf) in stm32_spi_transfer_one_irq()
743 writel_relaxed(ier, spi->base + STM32_SPI_IER); in stm32_spi_transfer_one_irq()
745 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_transfer_one_irq()
751 * stm32_spi_transfer_one_dma - transfer a single spi_transfer using DMA
764 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_transfer_one_dma()
767 if (spi->rx_buf) { in stm32_spi_transfer_one_dma()
769 dmaengine_slave_config(spi->dma_rx, &rx_dma_conf); in stm32_spi_transfer_one_dma()
771 /* Enable Rx DMA request */ in stm32_spi_transfer_one_dma()
775 spi->dma_rx, xfer->rx_sg.sgl, in stm32_spi_transfer_one_dma()
776 xfer->rx_sg.nents, in stm32_spi_transfer_one_dma()
782 if (spi->tx_buf) { in stm32_spi_transfer_one_dma()
784 dmaengine_slave_config(spi->dma_tx, &tx_dma_conf); in stm32_spi_transfer_one_dma()
787 spi->dma_tx, xfer->tx_sg.sgl, in stm32_spi_transfer_one_dma()
788 xfer->tx_sg.nents, in stm32_spi_transfer_one_dma()
793 if ((spi->tx_buf && !tx_dma_desc) || in stm32_spi_transfer_one_dma()
794 (spi->rx_buf && !rx_dma_desc)) in stm32_spi_transfer_one_dma()
798 rx_dma_desc->callback = stm32_spi_dma_cb; in stm32_spi_transfer_one_dma()
799 rx_dma_desc->callback_param = spi; in stm32_spi_transfer_one_dma()
802 dev_err(spi->dev, "Rx DMA submit failed\n"); in stm32_spi_transfer_one_dma()
805 /* Enable Rx DMA channel */ in stm32_spi_transfer_one_dma()
806 dma_async_issue_pending(spi->dma_rx); in stm32_spi_transfer_one_dma()
810 if (spi->cur_comm == SPI_SIMPLEX_TX) { in stm32_spi_transfer_one_dma()
811 tx_dma_desc->callback = stm32_spi_dma_cb; in stm32_spi_transfer_one_dma()
812 tx_dma_desc->callback_param = spi; in stm32_spi_transfer_one_dma()
816 dev_err(spi->dev, "Tx DMA submit failed\n"); in stm32_spi_transfer_one_dma()
819 /* Enable Tx DMA channel */ in stm32_spi_transfer_one_dma()
820 dma_async_issue_pending(spi->dma_tx); in stm32_spi_transfer_one_dma()
822 /* Enable Tx DMA request */ in stm32_spi_transfer_one_dma()
828 writel_relaxed(ier, spi->base + STM32_SPI_IER); in stm32_spi_transfer_one_dma()
834 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_transfer_one_dma()
839 if (spi->rx_buf) in stm32_spi_transfer_one_dma()
840 dmaengine_terminate_all(spi->dma_rx); in stm32_spi_transfer_one_dma()
845 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_transfer_one_dma()
847 dev_info(spi->dev, "DMA issue: fall back to irq transfer\n"); in stm32_spi_transfer_one_dma()
853 * stm32_spi_transfer_one_setup - common setup to transfer a single
854 * spi_transfer either using DMA or
866 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_transfer_one_setup()
868 if (spi->cur_bpw != transfer->bits_per_word) { in stm32_spi_transfer_one_setup()
871 spi->cur_bpw = transfer->bits_per_word; in stm32_spi_transfer_one_setup()
872 bpw = spi->cur_bpw - 1; in stm32_spi_transfer_one_setup()
877 spi->cur_fthlv = stm32_spi_prepare_fthlv(spi); in stm32_spi_transfer_one_setup()
878 fthlv = spi->cur_fthlv - 1; in stm32_spi_transfer_one_setup()
884 if (spi->cur_speed != transfer->speed_hz) { in stm32_spi_transfer_one_setup()
887 /* Update spi->cur_speed with real clock speed */ in stm32_spi_transfer_one_setup()
888 mbr = stm32_spi_prepare_mbr(spi, transfer->speed_hz); in stm32_spi_transfer_one_setup()
894 transfer->speed_hz = spi->cur_speed; in stm32_spi_transfer_one_setup()
901 writel_relaxed((readl_relaxed(spi->base + STM32_SPI_CFG1) & in stm32_spi_transfer_one_setup()
903 spi->base + STM32_SPI_CFG1); in stm32_spi_transfer_one_setup()
906 if (spi_dev->mode & SPI_3WIRE) { /* MISO/MOSI signals shared */ in stm32_spi_transfer_one_setup()
908 * SPI_3WIRE and xfer->tx_buf != NULL and xfer->rx_buf != NULL in stm32_spi_transfer_one_setup()
914 if (!transfer->tx_buf) in stm32_spi_transfer_one_setup()
916 else if (!transfer->rx_buf) in stm32_spi_transfer_one_setup()
919 if (!transfer->tx_buf) in stm32_spi_transfer_one_setup()
921 else if (!transfer->rx_buf) in stm32_spi_transfer_one_setup()
924 if (spi->cur_comm != mode) { in stm32_spi_transfer_one_setup()
925 spi->cur_comm = mode; in stm32_spi_transfer_one_setup()
932 if ((transfer->len > 1) && (spi->cur_midi > 0)) { in stm32_spi_transfer_one_setup()
933 u32 sck_period_ns = DIV_ROUND_UP(SPI_1HZ_NS, spi->cur_speed); in stm32_spi_transfer_one_setup()
934 u32 midi = min((u32)DIV_ROUND_UP(spi->cur_midi, sck_period_ns), in stm32_spi_transfer_one_setup()
937 dev_dbg(spi->dev, "period=%dns, midi=%d(=%dns)\n", in stm32_spi_transfer_one_setup()
944 writel_relaxed((readl_relaxed(spi->base + STM32_SPI_CFG2) & in stm32_spi_transfer_one_setup()
946 spi->base + STM32_SPI_CFG2); in stm32_spi_transfer_one_setup()
948 if (spi->cur_bpw <= 8) in stm32_spi_transfer_one_setup()
949 nb_words = transfer->len; in stm32_spi_transfer_one_setup()
950 else if (spi->cur_bpw <= 16) in stm32_spi_transfer_one_setup()
951 nb_words = DIV_ROUND_UP(transfer->len * 8, 16); in stm32_spi_transfer_one_setup()
953 nb_words = DIV_ROUND_UP(transfer->len * 8, 32); in stm32_spi_transfer_one_setup()
957 writel_relaxed(nb_words, spi->base + STM32_SPI_CR2); in stm32_spi_transfer_one_setup()
959 ret = -EMSGSIZE; in stm32_spi_transfer_one_setup()
963 spi->cur_xferlen = transfer->len; in stm32_spi_transfer_one_setup()
965 dev_dbg(spi->dev, "transfer communication mode set to %d\n", in stm32_spi_transfer_one_setup()
966 spi->cur_comm); in stm32_spi_transfer_one_setup()
967 dev_dbg(spi->dev, in stm32_spi_transfer_one_setup()
968 "data frame of %d-bit, data packet of %d data frames\n", in stm32_spi_transfer_one_setup()
969 spi->cur_bpw, spi->cur_fthlv); in stm32_spi_transfer_one_setup()
970 dev_dbg(spi->dev, "speed set to %dHz\n", spi->cur_speed); in stm32_spi_transfer_one_setup()
971 dev_dbg(spi->dev, "transfer of %d bytes (%d data frames)\n", in stm32_spi_transfer_one_setup()
972 spi->cur_xferlen, nb_words); in stm32_spi_transfer_one_setup()
973 dev_dbg(spi->dev, "dma %s\n", in stm32_spi_transfer_one_setup()
974 (spi->cur_usedma) ? "enabled" : "disabled"); in stm32_spi_transfer_one_setup()
977 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_transfer_one_setup()
983 * stm32_spi_transfer_one - transfer a single spi_transfer
995 spi->tx_buf = transfer->tx_buf; in stm32_spi_transfer_one()
996 spi->rx_buf = transfer->rx_buf; in stm32_spi_transfer_one()
997 spi->tx_len = spi->tx_buf ? transfer->len : 0; in stm32_spi_transfer_one()
998 spi->rx_len = spi->rx_buf ? transfer->len : 0; in stm32_spi_transfer_one()
1000 spi->cur_usedma = (master->can_dma && in stm32_spi_transfer_one()
1005 dev_err(spi->dev, "SPI transfer setup failed\n"); in stm32_spi_transfer_one()
1009 if (spi->cur_usedma) in stm32_spi_transfer_one()
1016 * stm32_spi_unprepare_msg - relax the hardware
1034 * stm32_spi_config - Configure SPI controller as SPI master
1040 spin_lock_irqsave(&spi->lock, flags); in stm32_spi_config()
1046 * - SS input value high in stm32_spi_config()
1047 * - transmitter half duplex direction in stm32_spi_config()
1048 * - automatic communication suspend when RX-Fifo is full in stm32_spi_config()
1055 * - Set the master mode (default Motorola mode) in stm32_spi_config()
1056 * - Consider 1 master/n slaves configuration and in stm32_spi_config()
1058 * - keep control of all associated GPIOs in stm32_spi_config()
1064 spin_unlock_irqrestore(&spi->lock, flags); in stm32_spi_config()
1070 { .compatible = "st,stm32h7-spi", },
1082 master = spi_alloc_master(&pdev->dev, sizeof(struct stm32_spi)); in stm32_spi_probe()
1084 dev_err(&pdev->dev, "spi master allocation failed\n"); in stm32_spi_probe()
1085 return -ENOMEM; in stm32_spi_probe()
1090 spi->dev = &pdev->dev; in stm32_spi_probe()
1091 spi->master = master; in stm32_spi_probe()
1092 spin_lock_init(&spi->lock); in stm32_spi_probe()
1095 spi->base = devm_ioremap_resource(&pdev->dev, res); in stm32_spi_probe()
1096 if (IS_ERR(spi->base)) { in stm32_spi_probe()
1097 ret = PTR_ERR(spi->base); in stm32_spi_probe()
1100 spi->phys_addr = (dma_addr_t)res->start; in stm32_spi_probe()
1102 spi->irq = platform_get_irq(pdev, 0); in stm32_spi_probe()
1103 if (spi->irq <= 0) { in stm32_spi_probe()
1104 dev_err(&pdev->dev, "no irq: %d\n", spi->irq); in stm32_spi_probe()
1105 ret = -ENOENT; in stm32_spi_probe()
1108 ret = devm_request_threaded_irq(&pdev->dev, spi->irq, NULL, in stm32_spi_probe()
1110 pdev->name, master); in stm32_spi_probe()
1112 dev_err(&pdev->dev, "irq%d request failed: %d\n", spi->irq, in stm32_spi_probe()
1117 spi->clk = devm_clk_get(&pdev->dev, 0); in stm32_spi_probe()
1118 if (IS_ERR(spi->clk)) { in stm32_spi_probe()
1119 ret = PTR_ERR(spi->clk); in stm32_spi_probe()
1120 dev_err(&pdev->dev, "clk get failed: %d\n", ret); in stm32_spi_probe()
1124 ret = clk_prepare_enable(spi->clk); in stm32_spi_probe()
1126 dev_err(&pdev->dev, "clk enable failed: %d\n", ret); in stm32_spi_probe()
1129 spi->clk_rate = clk_get_rate(spi->clk); in stm32_spi_probe()
1130 if (!spi->clk_rate) { in stm32_spi_probe()
1131 dev_err(&pdev->dev, "clk rate = 0\n"); in stm32_spi_probe()
1132 ret = -EINVAL; in stm32_spi_probe()
1136 spi->rst = devm_reset_control_get_exclusive(&pdev->dev, NULL); in stm32_spi_probe()
1137 if (!IS_ERR(spi->rst)) { in stm32_spi_probe()
1138 reset_control_assert(spi->rst); in stm32_spi_probe()
1140 reset_control_deassert(spi->rst); in stm32_spi_probe()
1143 spi->fifo_size = stm32_spi_get_fifo_size(spi); in stm32_spi_probe()
1147 dev_err(&pdev->dev, "controller configuration failed: %d\n", in stm32_spi_probe()
1152 master->dev.of_node = pdev->dev.of_node; in stm32_spi_probe()
1153 master->auto_runtime_pm = true; in stm32_spi_probe()
1154 master->bus_num = pdev->id; in stm32_spi_probe()
1155 master->mode_bits = SPI_MODE_3 | SPI_CS_HIGH | SPI_LSB_FIRST | in stm32_spi_probe()
1157 master->bits_per_word_mask = stm32_spi_get_bpw_mask(spi); in stm32_spi_probe()
1158 master->max_speed_hz = spi->clk_rate / SPI_MBR_DIV_MIN; in stm32_spi_probe()
1159 master->min_speed_hz = spi->clk_rate / SPI_MBR_DIV_MAX; in stm32_spi_probe()
1160 master->setup = stm32_spi_setup; in stm32_spi_probe()
1161 master->prepare_message = stm32_spi_prepare_msg; in stm32_spi_probe()
1162 master->transfer_one = stm32_spi_transfer_one; in stm32_spi_probe()
1163 master->unprepare_message = stm32_spi_unprepare_msg; in stm32_spi_probe()
1165 spi->dma_tx = dma_request_slave_channel(spi->dev, "tx"); in stm32_spi_probe()
1166 if (!spi->dma_tx) in stm32_spi_probe()
1167 dev_warn(&pdev->dev, "failed to request tx dma channel\n"); in stm32_spi_probe()
1169 master->dma_tx = spi->dma_tx; in stm32_spi_probe()
1171 spi->dma_rx = dma_request_slave_channel(spi->dev, "rx"); in stm32_spi_probe()
1172 if (!spi->dma_rx) in stm32_spi_probe()
1173 dev_warn(&pdev->dev, "failed to request rx dma channel\n"); in stm32_spi_probe()
1175 master->dma_rx = spi->dma_rx; in stm32_spi_probe()
1177 if (spi->dma_tx || spi->dma_rx) in stm32_spi_probe()
1178 master->can_dma = stm32_spi_can_dma; in stm32_spi_probe()
1180 pm_runtime_set_active(&pdev->dev); in stm32_spi_probe()
1181 pm_runtime_enable(&pdev->dev); in stm32_spi_probe()
1183 ret = devm_spi_register_master(&pdev->dev, master); in stm32_spi_probe()
1185 dev_err(&pdev->dev, "spi master registration failed: %d\n", in stm32_spi_probe()
1190 if (!master->cs_gpios) { in stm32_spi_probe()
1191 dev_err(&pdev->dev, "no CS gpios available\n"); in stm32_spi_probe()
1192 ret = -EINVAL; in stm32_spi_probe()
1196 for (i = 0; i < master->num_chipselect; i++) { in stm32_spi_probe()
1197 if (!gpio_is_valid(master->cs_gpios[i])) { in stm32_spi_probe()
1198 dev_err(&pdev->dev, "%i is not a valid gpio\n", in stm32_spi_probe()
1199 master->cs_gpios[i]); in stm32_spi_probe()
1200 ret = -EINVAL; in stm32_spi_probe()
1204 ret = devm_gpio_request(&pdev->dev, master->cs_gpios[i], in stm32_spi_probe()
1207 dev_err(&pdev->dev, "can't get CS gpio %i\n", in stm32_spi_probe()
1208 master->cs_gpios[i]); in stm32_spi_probe()
1213 dev_info(&pdev->dev, "driver initialized\n"); in stm32_spi_probe()
1218 if (spi->dma_tx) in stm32_spi_probe()
1219 dma_release_channel(spi->dma_tx); in stm32_spi_probe()
1220 if (spi->dma_rx) in stm32_spi_probe()
1221 dma_release_channel(spi->dma_rx); in stm32_spi_probe()
1223 pm_runtime_disable(&pdev->dev); in stm32_spi_probe()
1225 clk_disable_unprepare(spi->clk); in stm32_spi_probe()
1239 if (master->dma_tx) in stm32_spi_remove()
1240 dma_release_channel(master->dma_tx); in stm32_spi_remove()
1241 if (master->dma_rx) in stm32_spi_remove()
1242 dma_release_channel(master->dma_rx); in stm32_spi_remove()
1244 clk_disable_unprepare(spi->clk); in stm32_spi_remove()
1246 pm_runtime_disable(&pdev->dev); in stm32_spi_remove()
1257 clk_disable_unprepare(spi->clk); in stm32_spi_runtime_suspend()
1267 return clk_prepare_enable(spi->clk); in stm32_spi_runtime_resume()
1296 clk_disable_unprepare(spi->clk); in stm32_spi_resume()