/crypto/async_tx/ |
D | async_tx.c | 66 struct dma_async_tx_descriptor *tx) in async_tx_channel_switch() argument 74 if (txd_parent(depend_tx) && depend_tx->chan == tx->chan) { in async_tx_channel_switch() 75 txd_chain(depend_tx, tx); in async_tx_channel_switch() 100 txd_chain(intr_tx, tx); in async_tx_channel_switch() 121 tx->tx_submit(tx); in async_tx_channel_switch() 143 async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, in async_tx_submit() argument 148 tx->callback = submit->cb_fn; in async_tx_submit() 149 tx->callback_param = submit->cb_param; in async_tx_submit() 161 txd_parent(tx)); in async_tx_submit() 173 txd_chain(depend_tx, tx); in async_tx_submit() [all …]
|
D | async_raid6_recov.c | 35 struct dma_async_tx_descriptor *tx; in async_sum_product() local 50 tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef, in async_sum_product() 52 if (tx) { in async_sum_product() 53 dma_set_unmap(tx, unmap); in async_sum_product() 54 async_tx_submit(chan, tx, submit); in async_sum_product() 56 return tx; in async_sum_product() 99 struct dma_async_tx_descriptor *tx; in async_mult() local 115 tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr, in async_mult() 118 if (tx) { in async_mult() 119 dma_set_unmap(tx, unmap); in async_mult() [all …]
|
D | async_xor.c | 27 struct dma_async_tx_descriptor *tx = NULL; in do_async_xor() local 65 tx = dma->device_prep_dma_xor(chan, dma_dest, src_list, in do_async_xor() 69 if (unlikely(!tx)) in do_async_xor() 73 while (unlikely(!tx)) { in do_async_xor() 75 tx = dma->device_prep_dma_xor(chan, dma_dest, in do_async_xor() 82 dma_set_unmap(tx, unmap); in do_async_xor() 83 async_tx_submit(chan, tx, submit); in do_async_xor() 84 submit->depend_tx = tx; in do_async_xor() 96 return tx; in do_async_xor() 175 struct dma_async_tx_descriptor *tx; in async_xor() local [all …]
|
D | async_pq.c | 41 struct dma_async_tx_descriptor *tx = NULL; in do_async_gen_syndrome() local 78 tx = dma->device_prep_dma_pq(chan, dma_dest, in do_async_gen_syndrome() 83 if (likely(tx)) in do_async_gen_syndrome() 89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome() 90 async_tx_submit(chan, tx, submit); in do_async_gen_syndrome() 91 submit->depend_tx = tx; in do_async_gen_syndrome() 100 return tx; in do_async_gen_syndrome() 183 struct dma_async_tx_descriptor *tx; in async_gen_syndrome() local 228 tx = do_async_gen_syndrome(chan, coefs, j, unmap, dma_flags, submit); in async_gen_syndrome() 230 return tx; in async_gen_syndrome() [all …]
|
D | raid6test.c | 59 struct dma_async_tx_descriptor *tx = NULL; in raid6_dual_recov() local 69 tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); in raid6_dual_recov() 89 tx = async_xor(dest, blocks, 0, count, bytes, &submit); in raid6_dual_recov() 91 init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv); in raid6_dual_recov() 92 tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); in raid6_dual_recov() 98 tx = async_raid6_datap_recov(disks, bytes, faila, ptrs, &submit); in raid6_dual_recov() 102 tx = async_raid6_2data_recov(disks, bytes, faila, failb, ptrs, &submit); in raid6_dual_recov() 106 init_async_submit(&submit, ASYNC_TX_ACK, tx, callback, &cmp, addr_conv); in raid6_dual_recov() 107 tx = async_syndrome_val(ptrs, 0, disks, bytes, &result, spare, &submit); in raid6_dual_recov() 108 async_tx_issue_pending(tx); in raid6_dual_recov() [all …]
|
D | async_memcpy.c | 39 struct dma_async_tx_descriptor *tx = NULL; in async_memcpy() local 61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy() 66 if (tx) { in async_memcpy() 69 dma_set_unmap(tx, unmap); in async_memcpy() 70 async_tx_submit(chan, tx, submit); in async_memcpy() 91 return tx; in async_memcpy()
|
/crypto/ |
D | ecc.c | 1267 u64 tx[ECC_MAX_DIGITS]; in ecc_point_mult_shamir() local 1271 vli_set(tx, point->x, ndigits); in ecc_point_mult_shamir() 1273 apply_z(tx, ty, z, curve->p, ndigits); in ecc_point_mult_shamir() 1274 vli_mod_sub(tz, rx, tx, curve->p, ndigits); in ecc_point_mult_shamir() 1275 xycz_add(tx, ty, rx, ry, curve->p, ndigits); in ecc_point_mult_shamir()
|