/kernel/linux/linux-5.10/fs/ubifs/ |
D | scan.c | 30 int pad_len = 0, max_pad_len = min_t(int, UBIFS_PAD_NODE_SZ, len); in scan_padding_bytes() local 35 while (pad_len < max_pad_len && *p++ == UBIFS_PADDING_BYTE) in scan_padding_bytes() 36 pad_len += 1; in scan_padding_bytes() 38 if (!pad_len || (pad_len & 7)) in scan_padding_bytes() 41 dbg_scan("%d padding bytes", pad_len); in scan_padding_bytes() 43 return pad_len; in scan_padding_bytes() 84 int pad_len = le32_to_cpu(pad->pad_len); in ubifs_scan_a_node() local 88 if (pad_len < 0 || in ubifs_scan_a_node() 89 offs + node_len + pad_len > c->leb_size) { in ubifs_scan_a_node() 99 if ((node_len + pad_len) & 7) { in ubifs_scan_a_node() [all …]
|
D | crypto.c | 32 unsigned int pad_len = round_up(in_len, UBIFS_CIPHER_BLOCK_SIZE); in ubifs_encrypt() local 35 ubifs_assert(c, pad_len <= *out_len); in ubifs_encrypt() 39 if (pad_len != in_len) in ubifs_encrypt() 40 memset(p + in_len, 0, pad_len - in_len); in ubifs_encrypt() 42 err = fscrypt_encrypt_block_inplace(inode, virt_to_page(p), pad_len, in ubifs_encrypt() 48 *out_len = pad_len; in ubifs_encrypt()
|
D | recovery.c | 432 int empty_offs, pad_len; in clean_buf() local 438 pad_len = empty_offs - *offs; in clean_buf() 439 ubifs_pad(c, *buf, pad_len); in clean_buf() 440 *offs += pad_len; in clean_buf() 441 *buf += pad_len; in clean_buf() 442 *len -= pad_len; in clean_buf() 538 int pad_len = len - ALIGN(endpt, 8); in fix_unclean_leb() local 540 if (pad_len > 0) { in fix_unclean_leb() 541 void *buf = sleb->buf + len - pad_len; in fix_unclean_leb() 543 ubifs_pad(c, buf, pad_len); in fix_unclean_leb() [all …]
|
D | lpt_commit.c | 1035 int offs, pad_len; in get_pad_len() local 1040 pad_len = ALIGN(offs, c->min_io_size) - offs; in get_pad_len() 1041 return pad_len; in get_pad_len() 1118 int pad_len; in lpt_gc_lnum() local 1120 pad_len = get_pad_len(c, buf, len); in lpt_gc_lnum() 1121 if (pad_len) { in lpt_gc_lnum() 1122 buf += pad_len; in lpt_gc_lnum() 1123 len -= pad_len; in lpt_gc_lnum() 1613 int i, pad_len; in dbg_check_ltab_lnum() local 1615 pad_len = get_pad_len(c, p, len); in dbg_check_ltab_lnum() [all …]
|
D | tnc_commit.c | 109 int len, gap_remains, gap_pos, written, pad_len; in fill_gap() local 144 pad_len = c->ileb_len - gap_pos; in fill_gap() 147 pad_len = gap_remains; in fill_gap() 149 lnum, gap_start, gap_end, gap_end - gap_start, written, pad_len); in fill_gap() 150 ubifs_pad(c, c->ileb_buf + gap_pos, pad_len); in fill_gap() 151 *dirt += pad_len; in fill_gap()
|
/kernel/linux/linux-5.10/drivers/s390/crypto/ |
D | zcrypt_cca_key.h | 83 unsigned short pad_len; member 186 int short_len, long_len, pad_len, key_len, size; in zcrypt_type6_crt_key() local 201 pad_len = -(3*long_len + 2*short_len) & 7; in zcrypt_type6_crt_key() 202 key_len = 3*long_len + 2*short_len + pad_len + crt->inputdatalength; in zcrypt_type6_crt_key() 221 key->pvt.pad_len = pad_len; in zcrypt_type6_crt_key() 234 memset(key->key_parts + 3*long_len + 2*short_len + pad_len, in zcrypt_type6_crt_key()
|
D | zcrypt_msgtype6.c | 652 unsigned int reply_len, pad_len; in convert_type86_ica() local 694 pad_len = outputdatalength - reply_len; in convert_type86_ica() 695 if (pad_len > 0) { in convert_type86_ica() 696 if (pad_len < 10) in convert_type86_ica() 699 if (copy_to_user(outputdata, static_pad, pad_len - 1)) in convert_type86_ica() 701 if (put_user(0, outputdata + pad_len - 1)) in convert_type86_ica() 705 if (copy_to_user(outputdata + pad_len, data, reply_len)) in convert_type86_ica()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/qualcomm/ |
D | qca_uart.c | 203 u8 pad_len = 0; in qcauart_netdev_xmit() local 220 pad_len = QCAFRM_MIN_LEN - skb->len; in qcauart_netdev_xmit() 222 pos += qcafrm_create_header(pos, skb->len + pad_len); in qcauart_netdev_xmit() 227 if (pad_len) { in qcauart_netdev_xmit() 228 memset(pos, 0, pad_len); in qcauart_netdev_xmit() 229 pos += pad_len; in qcauart_netdev_xmit()
|
D | qca_spi.c | 725 u8 pad_len = 0; in qcaspi_netdev_xmit() local 728 pad_len = QCAFRM_MIN_LEN - skb->len; in qcaspi_netdev_xmit() 738 (skb_tailroom(skb) < QCAFRM_FOOTER_LEN + pad_len)) { in qcaspi_netdev_xmit() 740 QCAFRM_FOOTER_LEN + pad_len, GFP_ATOMIC); in qcaspi_netdev_xmit() 749 frame_len = skb->len + pad_len; in qcaspi_netdev_xmit() 754 if (pad_len) { in qcaspi_netdev_xmit() 755 ptmp = skb_put_zero(skb, pad_len); in qcaspi_netdev_xmit()
|
/kernel/linux/linux-5.10/include/linux/ |
D | if_rmnet.h | 10 u8 pad_len:6; member 16 u8 pad_len:6;
|
/kernel/linux/linux-5.10/drivers/net/caif/ |
D | caif_virtio.c | 221 u32 cfpkt_len, pad_len; in cfv_alloc_and_copy_skb() local 235 pad_len = (unsigned long)(frm + cfv->rx_hr) & (IP_HDR_ALIGN - 1); in cfv_alloc_and_copy_skb() 237 skb = netdev_alloc_skb(cfv->ndev, frm_len + pad_len); in cfv_alloc_and_copy_skb() 243 skb_reserve(skb, cfv->rx_hr + pad_len); in cfv_alloc_and_copy_skb() 485 u8 pad_len, hdr_ofs; in cfv_alloc_and_copy_to_shm() local 502 pad_len = hdr_ofs & (IP_HDR_ALIGN - 1); in cfv_alloc_and_copy_to_shm() 503 buf_info->size = cfv->tx_hr + skb->len + cfv->tx_tr + pad_len; in cfv_alloc_and_copy_to_shm() 511 skb_copy_bits(skb, 0, buf_info->vaddr + cfv->tx_hr + pad_len, skb->len); in cfv_alloc_and_copy_to_shm() 512 sg_init_one(sg, buf_info->vaddr + pad_len, in cfv_alloc_and_copy_to_shm()
|
/kernel/linux/linux-5.10/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha1.c | 151 unsigned int pad_len; in octeon_sha1_final() local 162 pad_len = (index < 56) ? (56 - index) : ((64+56) - index); in octeon_sha1_final() 167 __octeon_sha1_update(sctx, padding, pad_len); in octeon_sha1_final()
|
D | octeon-sha512.c | 177 unsigned int pad_len; in octeon_sha512_final() local 189 pad_len = (index < 112) ? (112 - index) : ((128+112) - index); in octeon_sha512_final() 194 __octeon_sha512_update(sctx, padding, pad_len); in octeon_sha512_final()
|
D | octeon-sha256.c | 163 unsigned int pad_len; in octeon_sha256_final() local 174 pad_len = (index < 56) ? (56 - index) : ((64+56) - index); in octeon_sha256_final() 179 __octeon_sha256_update(sctx, padding, pad_len); in octeon_sha256_final()
|
/kernel/linux/linux-5.10/drivers/crypto/bcm/ |
D | spu.c | 467 u32 pad_len = 0; in spum_gcm_ccm_pad_len() local 472 pad_len = ((data_size + m1) & ~m1) - data_size; in spum_gcm_ccm_pad_len() 474 return pad_len; in spum_gcm_ccm_pad_len() 608 hash_parms->pad_len; in spum_create_request() 621 hash_parms->pad_len); in spum_create_request() 639 auth_len -= hash_parms->pad_len; in spum_create_request() 640 cipher_len -= hash_parms->pad_len; in spum_create_request() 661 flow_log(" hash_pad_len:%u\n", hash_parms->pad_len); in spum_create_request()
|
D | cipher.c | 199 u8 tx_frag_num, unsigned int chunksize, u32 pad_len) in spu_skcipher_tx_sg_create() argument 232 if (pad_len) in spu_skcipher_tx_sg_create() 233 sg_set_buf(sg++, rctx->msg_buf.spu_req_pad, pad_len); in spu_skcipher_tx_sg_create() 315 u32 pad_len; /* total length of all padding */ in handle_skcipher_req() local 410 pad_len = stat_pad_len; in handle_skcipher_req() 411 if (pad_len) { in handle_skcipher_req() 422 packet_dump(" pad: ", rctx->msg_buf.spu_req_pad, pad_len); in handle_skcipher_req() 454 pad_len); in handle_skcipher_req() 584 unsigned int new_data_len, u32 pad_len) in spu_ahash_tx_sg_create() argument 616 if (pad_len) in spu_ahash_tx_sg_create() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
D | vnic_main.c | 260 u8 pad_len, q_idx = skb->queue_mapping; in hfi1_netdev_start_xmit() local 282 pad_len = -(skb->len + OPA_VNIC_ICRC_TAIL_LEN) & 0x7; in hfi1_netdev_start_xmit() 283 pad_len += OPA_VNIC_ICRC_TAIL_LEN; in hfi1_netdev_start_xmit() 290 pkt_len = (skb->len + pad_len) >> 2; in hfi1_netdev_start_xmit() 296 v_dbg("pbc 0x%016llX len %d pad_len %d\n", pbc, skb->len, pad_len); in hfi1_netdev_start_xmit() 297 err = dd->process_vnic_dma_send(dd, q_idx, vinfo, skb, pbc, pad_len); in hfi1_netdev_start_xmit()
|
/kernel/linux/linux-5.10/drivers/infiniband/ulp/opa_vnic/ |
D | opa_vnic_encap.c | 473 u32 pad_len; in opa_vnic_wire_length() local 476 pad_len = -(skb->len + OPA_VNIC_ICRC_TAIL_LEN) & 0x7; in opa_vnic_wire_length() 477 pad_len += OPA_VNIC_ICRC_TAIL_LEN; in opa_vnic_wire_length() 479 return (skb->len + pad_len) >> 3; in opa_vnic_wire_length()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/qualcomm/rmnet/ |
D | rmnet_map.h | 40 (Y)->data)->pad_len)
|
/kernel/linux/linux-5.10/drivers/net/wireless/ti/wlcore/ |
D | rx.h | 130 u8 pad_len; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/realtek/ |
D | atp.c | 496 static void write_packet(long ioaddr, int length, unsigned char *packet, int pad_len, int data_mode) in write_packet() argument 501 pad_len++; in write_packet() 510 } while (--length > pad_len) ; in write_packet() 526 while (--length > pad_len) in write_packet()
|
/kernel/linux/linux-5.10/crypto/ |
D | rsa-pkcs1pad.c | 179 unsigned int pad_len; in pkcs1pad_encrypt_sign_complete() local 187 pad_len = ctx->key_size - len; in pkcs1pad_encrypt_sign_complete() 190 if (likely(!pad_len)) in pkcs1pad_encrypt_sign_complete() 199 out_buf + pad_len, len); in pkcs1pad_encrypt_sign_complete()
|
/kernel/linux/linux-5.10/drivers/net/wireless/ath/ath9k/ |
D | hif_usb.c | 580 u16 pad_len; in ath9k_hif_usb_rx_stream() local 600 pad_len = 4 - (pkt_len & 0x3); in ath9k_hif_usb_rx_stream() 601 if (pad_len == 4) in ath9k_hif_usb_rx_stream() 602 pad_len = 0; in ath9k_hif_usb_rx_stream() 605 index = index + 4 + pkt_len + pad_len; in ath9k_hif_usb_rx_stream() 612 hif_dev->rx_pad_len = pad_len; in ath9k_hif_usb_rx_stream()
|
/kernel/linux/linux-5.10/drivers/scsi/bnx2i/ |
D | bnx2i_hwi.c | 1442 int pad_len; in bnx2i_process_login_resp() local 1472 pad_len = 0; in bnx2i_process_login_resp() 1474 pad_len = 4 - (pld_len % 4); in bnx2i_process_login_resp() 1476 if (pad_len) { in bnx2i_process_login_resp() 1478 for (i = 0; i < pad_len; i++) { in bnx2i_process_login_resp() 1510 int pad_len; in bnx2i_process_text_resp() local 1533 pad_len = 0; in bnx2i_process_text_resp() 1535 pad_len = 4 - (pld_len % 4); in bnx2i_process_text_resp() 1537 if (pad_len) { in bnx2i_process_text_resp() 1539 for (i = 0; i < pad_len; i++) { in bnx2i_process_text_resp()
|
/kernel/linux/linux-5.10/drivers/iommu/ |
D | dma-iommu.c | 861 size_t pad_len = (mask - iova_len + 1) & mask; in iommu_dma_map_sg() local 882 if (pad_len && pad_len < s_length - 1) { in iommu_dma_map_sg() 883 prev->length += pad_len; in iommu_dma_map_sg() 884 iova_len += pad_len; in iommu_dma_map_sg()
|