Lines Matching refs:op
136 static void mtk_nor_set_addr(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_set_addr() argument
138 u32 addr = op->addr.val; in mtk_nor_set_addr()
145 if (op->addr.nbytes == 4) { in mtk_nor_set_addr()
153 static bool need_bounce(struct mtk_nor *sp, const struct spi_mem_op *op) in need_bounce() argument
155 return ((uintptr_t)op->data.buf.in & MTK_NOR_DMA_ALIGN_MASK); in need_bounce()
158 static bool mtk_nor_match_read(const struct spi_mem_op *op) in mtk_nor_match_read() argument
162 if (op->dummy.buswidth) in mtk_nor_match_read()
163 dummy = op->dummy.nbytes * BITS_PER_BYTE / op->dummy.buswidth; in mtk_nor_match_read()
165 if ((op->data.buswidth == 2) || (op->data.buswidth == 4)) { in mtk_nor_match_read()
166 if (op->addr.buswidth == 1) in mtk_nor_match_read()
168 else if (op->addr.buswidth == 2) in mtk_nor_match_read()
170 else if (op->addr.buswidth == 4) in mtk_nor_match_read()
172 } else if ((op->addr.buswidth == 1) && (op->data.buswidth == 1)) { in mtk_nor_match_read()
173 if (op->cmd.opcode == 0x03) in mtk_nor_match_read()
175 else if (op->cmd.opcode == 0x0b) in mtk_nor_match_read()
181 static bool mtk_nor_match_prg(const struct spi_mem_op *op) in mtk_nor_match_prg() argument
186 if ((op->cmd.buswidth > 1) || (op->addr.buswidth > 1) || in mtk_nor_match_prg()
187 (op->dummy.buswidth > 1) || (op->data.buswidth > 1)) in mtk_nor_match_prg()
190 tx_len = op->cmd.nbytes + op->addr.nbytes; in mtk_nor_match_prg()
192 if (op->data.dir == SPI_MEM_DATA_OUT) { in mtk_nor_match_prg()
194 tx_len += op->dummy.nbytes; in mtk_nor_match_prg()
202 if ((!op->addr.nbytes) && in mtk_nor_match_prg()
203 (tx_len + op->data.nbytes > MTK_NOR_REG_PRGDATA_MAX + 1)) in mtk_nor_match_prg()
205 } else if (op->data.dir == SPI_MEM_DATA_IN) { in mtk_nor_match_prg()
209 rx_len = op->data.nbytes; in mtk_nor_match_prg()
210 prg_left = MTK_NOR_PRG_CNT_MAX / 8 - tx_len - op->dummy.nbytes; in mtk_nor_match_prg()
214 if (!op->addr.nbytes) in mtk_nor_match_prg()
219 prg_len = tx_len + op->dummy.nbytes + rx_len; in mtk_nor_match_prg()
223 prg_len = tx_len + op->dummy.nbytes; in mtk_nor_match_prg()
230 static void mtk_nor_adj_prg_size(struct spi_mem_op *op) in mtk_nor_adj_prg_size() argument
234 tx_len = op->cmd.nbytes + op->addr.nbytes; in mtk_nor_adj_prg_size()
235 if (op->data.dir == SPI_MEM_DATA_OUT) { in mtk_nor_adj_prg_size()
236 tx_len += op->dummy.nbytes; in mtk_nor_adj_prg_size()
238 if (op->data.nbytes > tx_left) in mtk_nor_adj_prg_size()
239 op->data.nbytes = tx_left; in mtk_nor_adj_prg_size()
240 } else if (op->data.dir == SPI_MEM_DATA_IN) { in mtk_nor_adj_prg_size()
241 prg_left = MTK_NOR_PRG_CNT_MAX / 8 - tx_len - op->dummy.nbytes; in mtk_nor_adj_prg_size()
244 if (op->data.nbytes > prg_left) in mtk_nor_adj_prg_size()
245 op->data.nbytes = prg_left; in mtk_nor_adj_prg_size()
249 static int mtk_nor_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op) in mtk_nor_adjust_op_size() argument
253 if (!op->data.nbytes) in mtk_nor_adjust_op_size()
256 if ((op->addr.nbytes == 3) || (op->addr.nbytes == 4)) { in mtk_nor_adjust_op_size()
257 if ((op->data.dir == SPI_MEM_DATA_IN) && in mtk_nor_adjust_op_size()
258 mtk_nor_match_read(op)) { in mtk_nor_adjust_op_size()
260 if (op->data.nbytes > 0x400000) in mtk_nor_adjust_op_size()
261 op->data.nbytes = 0x400000; in mtk_nor_adjust_op_size()
263 if ((op->addr.val & MTK_NOR_DMA_ALIGN_MASK) || in mtk_nor_adjust_op_size()
264 (op->data.nbytes < MTK_NOR_DMA_ALIGN)) in mtk_nor_adjust_op_size()
265 op->data.nbytes = 1; in mtk_nor_adjust_op_size()
266 else if (!need_bounce(sp, op)) in mtk_nor_adjust_op_size()
267 op->data.nbytes &= ~MTK_NOR_DMA_ALIGN_MASK; in mtk_nor_adjust_op_size()
268 else if (op->data.nbytes > MTK_NOR_BOUNCE_BUF_SIZE) in mtk_nor_adjust_op_size()
269 op->data.nbytes = MTK_NOR_BOUNCE_BUF_SIZE; in mtk_nor_adjust_op_size()
271 } else if (op->data.dir == SPI_MEM_DATA_OUT) { in mtk_nor_adjust_op_size()
272 if (op->data.nbytes >= MTK_NOR_PP_SIZE) in mtk_nor_adjust_op_size()
273 op->data.nbytes = MTK_NOR_PP_SIZE; in mtk_nor_adjust_op_size()
275 op->data.nbytes = 1; in mtk_nor_adjust_op_size()
280 mtk_nor_adj_prg_size(op); in mtk_nor_adjust_op_size()
285 const struct spi_mem_op *op) in mtk_nor_supports_op() argument
287 if (!spi_mem_default_supports_op(mem, op)) in mtk_nor_supports_op()
290 if (op->cmd.buswidth != 1) in mtk_nor_supports_op()
293 if ((op->addr.nbytes == 3) || (op->addr.nbytes == 4)) { in mtk_nor_supports_op()
294 switch(op->data.dir) { in mtk_nor_supports_op()
296 if (mtk_nor_match_read(op)) in mtk_nor_supports_op()
300 if ((op->addr.buswidth == 1) && in mtk_nor_supports_op()
301 (op->dummy.nbytes == 0) && in mtk_nor_supports_op()
302 (op->data.buswidth == 1)) in mtk_nor_supports_op()
310 return mtk_nor_match_prg(op); in mtk_nor_supports_op()
313 static void mtk_nor_setup_bus(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_setup_bus() argument
317 if (op->addr.nbytes == 4) in mtk_nor_setup_bus()
320 if (op->data.buswidth == 4) { in mtk_nor_setup_bus()
322 writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA(4)); in mtk_nor_setup_bus()
323 if (op->addr.buswidth == 4) in mtk_nor_setup_bus()
325 } else if (op->data.buswidth == 2) { in mtk_nor_setup_bus()
327 writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA(3)); in mtk_nor_setup_bus()
328 if (op->addr.buswidth == 2) in mtk_nor_setup_bus()
331 if (op->cmd.opcode == 0x0b) in mtk_nor_setup_bus()
382 static int mtk_nor_read_bounce(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_read_bounce() argument
387 if (op->data.nbytes & MTK_NOR_DMA_ALIGN_MASK) in mtk_nor_read_bounce()
388 rdlen = (op->data.nbytes + MTK_NOR_DMA_ALIGN) & ~MTK_NOR_DMA_ALIGN_MASK; in mtk_nor_read_bounce()
390 rdlen = op->data.nbytes; in mtk_nor_read_bounce()
392 ret = mtk_nor_dma_exec(sp, op->addr.val, rdlen, sp->buffer_dma); in mtk_nor_read_bounce()
395 memcpy(op->data.buf.in, sp->buffer, op->data.nbytes); in mtk_nor_read_bounce()
400 static int mtk_nor_read_dma(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_read_dma() argument
405 if (need_bounce(sp, op)) in mtk_nor_read_dma()
406 return mtk_nor_read_bounce(sp, op); in mtk_nor_read_dma()
408 dma_addr = dma_map_single(sp->dev, op->data.buf.in, in mtk_nor_read_dma()
409 op->data.nbytes, DMA_FROM_DEVICE); in mtk_nor_read_dma()
414 ret = mtk_nor_dma_exec(sp, op->addr.val, op->data.nbytes, dma_addr); in mtk_nor_read_dma()
416 dma_unmap_single(sp->dev, dma_addr, op->data.nbytes, DMA_FROM_DEVICE); in mtk_nor_read_dma()
421 static int mtk_nor_read_pio(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_read_pio() argument
423 u8 *buf = op->data.buf.in; in mtk_nor_read_pio()
465 static int mtk_nor_pp_buffered(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_pp_buffered() argument
467 const u8 *buf = op->data.buf.out; in mtk_nor_pp_buffered()
475 for (i = 0; i < op->data.nbytes; i += 4) { in mtk_nor_pp_buffered()
481 (op->data.nbytes + 5) * BITS_PER_BYTE); in mtk_nor_pp_buffered()
485 const struct spi_mem_op *op) in mtk_nor_pp_unbuffered() argument
487 const u8 *buf = op->data.buf.out; in mtk_nor_pp_unbuffered()
497 static int mtk_nor_spi_mem_prg(struct mtk_nor *sp, const struct spi_mem_op *op) in mtk_nor_spi_mem_prg() argument
506 tx_len = op->cmd.nbytes + op->addr.nbytes; in mtk_nor_spi_mem_prg()
509 if (op->data.dir == SPI_MEM_DATA_OUT) in mtk_nor_spi_mem_prg()
510 tx_len += op->dummy.nbytes + op->data.nbytes; in mtk_nor_spi_mem_prg()
511 else if (op->data.dir == SPI_MEM_DATA_IN) in mtk_nor_spi_mem_prg()
512 rx_len = op->data.nbytes; in mtk_nor_spi_mem_prg()
514 prg_len = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes + in mtk_nor_spi_mem_prg()
515 op->data.nbytes; in mtk_nor_spi_mem_prg()
526 for (i = op->cmd.nbytes; i > 0; i--, reg_offset--) { in mtk_nor_spi_mem_prg()
528 bufbyte = (op->cmd.opcode >> ((i - 1) * BITS_PER_BYTE)) & 0xff; in mtk_nor_spi_mem_prg()
532 for (i = op->addr.nbytes; i > 0; i--, reg_offset--) { in mtk_nor_spi_mem_prg()
534 bufbyte = (op->addr.val >> ((i - 1) * BITS_PER_BYTE)) & 0xff; in mtk_nor_spi_mem_prg()
538 if (op->data.dir == SPI_MEM_DATA_OUT) { in mtk_nor_spi_mem_prg()
539 for (i = 0; i < op->dummy.nbytes; i++, reg_offset--) { in mtk_nor_spi_mem_prg()
544 for (i = 0; i < op->data.nbytes; i++, reg_offset--) { in mtk_nor_spi_mem_prg()
546 writeb(((const u8 *)(op->data.buf.out))[i], reg); in mtk_nor_spi_mem_prg()
564 if (op->data.dir == SPI_MEM_DATA_IN) { in mtk_nor_spi_mem_prg()
565 for (i = op->data.nbytes - 1; i >= 0; i--, reg_offset++) { in mtk_nor_spi_mem_prg()
567 ((u8 *)(op->data.buf.in))[i] = readb(reg); in mtk_nor_spi_mem_prg()
574 static int mtk_nor_exec_op(struct spi_mem *mem, const struct spi_mem_op *op) in mtk_nor_exec_op() argument
579 if ((op->data.nbytes == 0) || in mtk_nor_exec_op()
580 ((op->addr.nbytes != 3) && (op->addr.nbytes != 4))) in mtk_nor_exec_op()
581 return mtk_nor_spi_mem_prg(sp, op); in mtk_nor_exec_op()
583 if (op->data.dir == SPI_MEM_DATA_OUT) { in mtk_nor_exec_op()
584 mtk_nor_set_addr(sp, op); in mtk_nor_exec_op()
585 writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA0); in mtk_nor_exec_op()
586 if (op->data.nbytes == MTK_NOR_PP_SIZE) in mtk_nor_exec_op()
587 return mtk_nor_pp_buffered(sp, op); in mtk_nor_exec_op()
588 return mtk_nor_pp_unbuffered(sp, op); in mtk_nor_exec_op()
591 if ((op->data.dir == SPI_MEM_DATA_IN) && mtk_nor_match_read(op)) { in mtk_nor_exec_op()
595 mtk_nor_setup_bus(sp, op); in mtk_nor_exec_op()
596 if (op->data.nbytes == 1) { in mtk_nor_exec_op()
597 mtk_nor_set_addr(sp, op); in mtk_nor_exec_op()
598 return mtk_nor_read_pio(sp, op); in mtk_nor_exec_op()
600 return mtk_nor_read_dma(sp, op); in mtk_nor_exec_op()
604 return mtk_nor_spi_mem_prg(sp, op); in mtk_nor_exec_op()