/drivers/mtd/lpddr/ |
D | lpddr2_nvm.c | 321 u_long add, current_len, tot_len, target_len, my_data; in lpddr2_nvm_write() local 332 tot_len = 0; in lpddr2_nvm_write() 334 while (tot_len < target_len) { in lpddr2_nvm_write() 336 my_data = write_buf[tot_len]; in lpddr2_nvm_write() 337 my_data += (write_buf[tot_len+1]) << 8; in lpddr2_nvm_write() 339 my_data += (write_buf[tot_len+2]) << 16; in lpddr2_nvm_write() 340 my_data += (write_buf[tot_len+3]) << 24; in lpddr2_nvm_write() 348 tot_len += pcm_data->bus_width; in lpddr2_nvm_write() 350 current_len = min(target_len - tot_len, in lpddr2_nvm_write() 353 0x00, add, current_len, write_buf + tot_len); in lpddr2_nvm_write() [all …]
|
/drivers/net/slip/ |
D | slhc.c | 405 if(ip->tot_len != cs->cs_ip.tot_len && in slhc_compress() 406 ntohs(cs->cs_ip.tot_len) == hlen) in slhc_compress() 417 deltaS == ntohs(cs->cs_ip.tot_len) - hlen){ in slhc_compress() 424 if(deltaS == ntohs(cs->cs_ip.tot_len) - hlen){ in slhc_compress() 553 i = ntohs(ip->tot_len) - hdrlen; in slhc_uncompress() 561 ntohs(ip->tot_len) - hdrlen); in slhc_uncompress() 611 ip->tot_len = htons(len); in slhc_uncompress()
|
/drivers/net/netdevsim/ |
D | psample.c | 40 int tot_len, data_len = NSIM_PSAMPLE_L4_DATA_LEN; in nsim_dev_psample_skb_build() local 49 tot_len = sizeof(struct iphdr) + sizeof(struct udphdr) + data_len; in nsim_dev_psample_skb_build() 66 iph->tot_len = htons(tot_len); in nsim_dev_psample_skb_build()
|
D | dev.c | 658 int tot_len, data_len = NSIM_TRAP_L4_DATA_LEN; in nsim_dev_trap_skb_build() local 667 tot_len = sizeof(struct iphdr) + sizeof(struct udphdr) + data_len; in nsim_dev_trap_skb_build() 684 iph->tot_len = htons(tot_len); in nsim_dev_trap_skb_build()
|
/drivers/infiniband/core/ |
D | ud_header.c | 133 { STRUCT_FIELD(ip4, tot_len), 290 iph.tot_len = header->ip4.tot_len; in ib_ud_ip4_csum() 370 header->ip4.tot_len = in ib_ud_header_init()
|
D | lag.c | 50 iph->tot_len = htons(sizeof(struct udphdr) + sizeof(struct in rdma_build_skb()
|
/drivers/scsi/fnic/ |
D | fnic_fcs.c | 1084 u32 tot_len, eth_hdr_len; in fnic_send_frame() local 1116 tot_len = skb->len; in fnic_send_frame() 1117 BUG_ON(tot_len % 4); in fnic_send_frame() 1124 pa = dma_map_single(&fnic->pdev->dev, eth_hdr, tot_len, DMA_TO_DEVICE); in fnic_send_frame() 1132 (char *)eth_hdr, tot_len)) != 0) { in fnic_send_frame() 1139 dma_unmap_single(&fnic->pdev->dev, pa, tot_len, DMA_TO_DEVICE); in fnic_send_frame() 1144 fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp), in fnic_send_frame()
|
/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_main.c | 349 skb = alloc_skb(gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 353 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 359 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
|
/drivers/crypto/bcm/ |
D | util.h | 97 u8 from_nents, u32 tot_len);
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1518 if (gl->tot_len <= RX_COPY_THRES) { in t4vf_pktgl_to_skb() 1520 skb = alloc_skb(gl->tot_len, GFP_ATOMIC); in t4vf_pktgl_to_skb() 1523 __skb_put(skb, gl->tot_len); in t4vf_pktgl_to_skb() 1524 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in t4vf_pktgl_to_skb() 1533 skb->len = gl->tot_len; in t4vf_pktgl_to_skb() 1584 skb->len = gl->tot_len - s->pktshift; in do_gro() 1802 gl.tot_len = len; in process_responses()
|
D | adapter.h | 156 unsigned int tot_len; /* total length of fragments */ member
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_rx.c | 799 u16 tot_len; in mlx5e_lro_update_hdr() local 804 tot_len = cqe_bcnt - network_depth; in mlx5e_lro_update_hdr() 814 ipv4->tot_len = cpu_to_be16(tot_len); in mlx5e_lro_update_hdr() 824 tot_len - sizeof(struct iphdr), in mlx5e_lro_update_hdr() 827 u16 payload_len = tot_len - sizeof(struct ipv6hdr); in mlx5e_lro_update_hdr() 954 pkt_len = network_depth + ntohs(ip4->tot_len); in mlx5e_skb_csum_fixup()
|
D | en_selftest.c | 148 iph->tot_len = htons(iplen); in mlx5e_test_get_udp_skb()
|
/drivers/media/cec/core/ |
D | cec-pin.c | 511 unsigned int tot_len = pin->tx_msg.len + in cec_pin_tx_states() local 515 v = !pin->tx_post_eom && tx_byte_idx == tot_len - 1; in cec_pin_tx_states() 516 if (tot_len > 1 && tx_byte_idx == tot_len - 2 && in cec_pin_tx_states()
|
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ipsec_rxtx.c | 128 ipv4hdr->tot_len = htons(ntohs(ipv4hdr->tot_len) - trailer_len); in mlx5e_ipsec_remove_trailer()
|
D | tls_rxtx.c | 161 iph->tot_len = htons(nskb->len - skb_network_offset(nskb)); in mlx5e_tls_complete_sync_skb()
|
/drivers/net/ethernet/cavium/thunder/ |
D | q_struct.h | 543 u64 tot_len:20; /* W0 */ member 552 u64 tot_len:20;
|
D | nicvf_queues.c | 1198 atomic64_add(hdr->tot_len, in nicvf_sq_free_used_descs() 1230 hdr->tot_len = len; in nicvf_xdp_sq_add_hdr_subdesc() 1357 hdr->tot_len = len; in nicvf_sq_add_hdr_subdesc() 1453 hdr->tot_len = 1; in nicvf_sq_add_cqe_subdesc()
|
/drivers/crypto/ |
D | n2_core.c | 854 unsigned int tot_len; in n2_compute_chunks() local 873 tot_len = 0; in n2_compute_chunks() 892 tot_len + this_len > (1 << 16)) { in n2_compute_chunks() 906 tot_len = 0; in n2_compute_chunks() 914 tot_len += this_len; in n2_compute_chunks()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_ah.c | 132 ipv4.tot_len = htons(0); in set_av_attr()
|
/drivers/infiniband/hw/cxgb4/ |
D | device.c | 1117 skb = alloc_skb(gl->tot_len + sizeof(struct cpl_pass_accept_req) + in copy_gl_to_skb_pkt() 1122 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) + in copy_gl_to_skb_pkt() 1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 1197 gl->tot_len); in c4iw_uld_rx_handler()
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_tx.c | 298 ip.v4->tot_len = 0; in offload_tso() 330 ip.v4->tot_len = 0; in offload_tso()
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cudbg_lib.c | 1368 unsigned long tot_len, in cudbg_read_fw_mem() argument 1398 bytes_left = tot_len; in cudbg_read_fw_mem() 3373 u32 i, j, cur_off, tot_len; in cudbg_collect_qdesc() local 3379 tot_len = size; in cudbg_collect_qdesc() 3537 while (tot_len) { in cudbg_collect_qdesc() 3538 u32 chunk_size = min_t(u32, tot_len, CUDBG_CHUNK_SIZE); in cudbg_collect_qdesc() 3548 tot_len -= chunk_size; in cudbg_collect_qdesc()
|
/drivers/net/ethernet/qualcomm/rmnet/ |
D | rmnet_map_data.c | 93 ntohs(ip4h->tot_len) - ip4h->ihl * 4, in rmnet_map_ipv4_dl_csum_trailer()
|
/drivers/net/ethernet/netronome/nfp/flower/ |
D | offload.c | 98 u32 meta_len, key_len, mask_len, act_len, tot_len; in nfp_flower_xmit_flow() local 107 tot_len = meta_len + key_len + mask_len + act_len; in nfp_flower_xmit_flow() 116 skb = nfp_flower_cmsg_alloc(app, tot_len, mtype, GFP_KERNEL); in nfp_flower_xmit_flow()
|