/drivers/infiniband/sw/rxe/ |
D | rxe_net.c | 151 struct in_addr *daddr) in rxe_find_route4() argument 159 memcpy(&fl.daddr, daddr, sizeof(*daddr)); in rxe_find_route4() 164 pr_err_ratelimited("no route to %pI4\n", &daddr->s_addr); in rxe_find_route4() 174 struct in6_addr *daddr) in rxe_find_route6() argument 182 memcpy(&fl6.daddr, daddr, sizeof(*daddr)); in rxe_find_route6() 187 pr_err_ratelimited("no route to %pI6\n", daddr); in rxe_find_route6() 192 pr_err("no route to %pI6\n", daddr); in rxe_find_route6() 206 struct in6_addr *daddr) in rxe_find_route6() argument 296 __be32 saddr, __be32 daddr, __u8 proto, in prepare_ipv4_hdr() argument 317 iph->daddr = daddr; in prepare_ipv4_hdr() [all …]
|
D | rxe_recv.c | 171 struct in_addr *daddr = in check_addr() local 174 if (ip_hdr(skb)->daddr != saddr->s_addr) { in check_addr() 176 &ip_hdr(skb)->daddr, in check_addr() 181 if (ip_hdr(skb)->saddr != daddr->s_addr) { in check_addr() 184 &daddr->s_addr); in check_addr() 191 struct in6_addr *daddr = in check_addr() local 194 if (memcmp(&ipv6_hdr(skb)->daddr, saddr, sizeof(*saddr))) { in check_addr() 196 &ipv6_hdr(skb)->daddr, saddr); in check_addr() 200 if (memcmp(&ipv6_hdr(skb)->saddr, daddr, sizeof(*daddr))) { in check_addr() 202 &ipv6_hdr(skb)->saddr, daddr); in check_addr() [all …]
|
/drivers/staging/wlan-ng/ |
D | p80211conv.c | 181 memcpy(p80211_hdr->a3.a1, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 189 memcpy(p80211_hdr->a3.a3, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 193 memcpy(p80211_hdr->a3.a1, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 284 u8 daddr[ETH_ALEN]; in skb_p80211_to_ether() local 301 ether_addr_copy(daddr, w_hdr->a3.a1); in skb_p80211_to_ether() 305 ether_addr_copy(daddr, w_hdr->a3.a1); in skb_p80211_to_ether() 309 ether_addr_copy(daddr, w_hdr->a3.a3); in skb_p80211_to_ether() 318 ether_addr_copy(daddr, w_hdr->a4.a3); in skb_p80211_to_ether() 364 ((!ether_addr_equal_unaligned(daddr, e_hdr->daddr)) || in skb_p80211_to_ether() 411 ether_addr_copy(e_hdr->daddr, daddr); in skb_p80211_to_ether() [all …]
|
/drivers/gpu/drm/radeon/ |
D | radeon_benchmark.c | 36 uint64_t saddr, uint64_t daddr, in radeon_benchmark_do_move() argument 49 fence = radeon_copy_dma(rdev, saddr, daddr, in radeon_benchmark_do_move() 54 fence = radeon_copy_blit(rdev, saddr, daddr, in radeon_benchmark_do_move() 92 uint64_t saddr, daddr; in radeon_benchmark_move() local 116 r = radeon_bo_pin(dobj, ddomain, &daddr); in radeon_benchmark_move() 123 time = radeon_benchmark_do_move(rdev, size, saddr, daddr, in radeon_benchmark_move() 134 time = radeon_benchmark_do_move(rdev, size, saddr, daddr, in radeon_benchmark_move()
|
/drivers/net/ethernet/chelsio/libcxgb/ |
D | libcxgb_cm.c | 58 __func__, ntohl(ip->saddr), ntohl(ip->daddr), in cxgb_get_4tuple() 62 memcpy(local_ip, &ip->daddr, 4); in cxgb_get_4tuple() 65 __func__, ip6->saddr.s6_addr, ip6->daddr.s6_addr, in cxgb_get_4tuple() 69 memcpy(local_ip, ip6->daddr.s6_addr, 16); in cxgb_get_4tuple() 131 memcpy(&fl6.daddr, peer_ip, 16); in cxgb_find_route6() 133 if (ipv6_addr_type(&fl6.daddr) & IPV6_ADDR_LINKLOCAL) in cxgb_find_route6()
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_benchmark.c | 32 uint64_t saddr, uint64_t daddr, int n) in amdgpu_benchmark_do_move() argument 42 r = amdgpu_copy_buffer(ring, saddr, daddr, size, NULL, &fence, in amdgpu_benchmark_do_move() 78 uint64_t saddr, daddr; in amdgpu_benchmark_move() local 104 r = amdgpu_bo_pin(dobj, ddomain, &daddr); in amdgpu_benchmark_move() 111 time = amdgpu_benchmark_do_move(adev, size, saddr, daddr, n); in amdgpu_benchmark_move()
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_main.c | 731 static void push_pseudo_header(struct sk_buff *skb, const char *daddr) in push_pseudo_header() argument 736 memcpy(phdr->hwaddr, daddr, INFINIBAND_ALEN); in push_pseudo_header() 844 if (ipoib_cm_enabled(dev, neigh->daddr)) { in path_rec_completion() 938 static struct ipoib_neigh *neigh_add_path(struct sk_buff *skb, u8 *daddr, in neigh_add_path() argument 947 neigh = ipoib_neigh_alloc(daddr, dev); in neigh_add_path() 963 path = __path_find(dev, daddr + 4); in neigh_add_path() 965 path = path_rec_create(dev, daddr + 4); in neigh_add_path() 978 if (ipoib_cm_enabled(dev, neigh->daddr)) { in neigh_add_path() 987 push_pseudo_header(skb, neigh->daddr); in neigh_add_path() 996 ipoib_send(dev, skb, path->ah, IPOIB_QPN(daddr)); in neigh_add_path() [all …]
|
/drivers/target/ |
D | target_core_sbc.c | 1262 void *daddr, *paddr; in sbc_dif_generate() local 1268 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1277 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1283 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1288 crc = crc_t10dif(daddr + offset, avail); in sbc_dif_generate() 1290 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1296 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1298 crc = crc_t10dif_update(crc, daddr, offset); in sbc_dif_generate() 1318 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1417 void *daddr, *paddr; in sbc_dif_verify() local [all …]
|
/drivers/misc/genwqe/ |
D | card_utils.c | 258 dma_addr_t daddr; in genwqe_map_pages() local 261 daddr = pci_map_page(pci_dev, page_list[i], in genwqe_map_pages() 266 if (pci_dma_mapping_error(pci_dev, daddr)) { in genwqe_map_pages() 269 __func__, (long long)daddr); in genwqe_map_pages() 273 dma_list[i] = daddr; in genwqe_map_pages() 394 dma_addr_t daddr; in genwqe_setup_sgl() local 409 daddr = sgl->fpage_dma_addr + map_offs; in genwqe_setup_sgl() 413 daddr = sgl->lpage_dma_addr; in genwqe_setup_sgl() 415 daddr = dma_list[p] + map_offs; in genwqe_setup_sgl() 421 if (prev_daddr == daddr) { in genwqe_setup_sgl() [all …]
|
/drivers/infiniband/core/ |
D | addr.c | 148 const void *daddr, in ib_nl_ip_send_msg() argument 185 nla_put(skb, attrtype, size, daddr); in ib_nl_ip_send_msg() 342 const void *daddr, u32 seq, u16 family) in ib_nl_fetch_ha() argument 349 return ib_nl_ip_send_msg(dev_addr, daddr, seq, family); in ib_nl_fetch_ha() 353 const void *daddr) in dst_fetch_ha() argument 358 n = dst_neigh_lookup(dst, daddr); in dst_fetch_ha() 397 const void *daddr = (dst_in->sa_family == AF_INET) ? in fetch_ha() local 404 return ib_nl_fetch_ha(dst, dev_addr, daddr, seq, family); in fetch_ha() 406 return dst_fetch_ha(dst, dev_addr, daddr); in fetch_ha() 421 fl4.daddr = dst_ip; in addr4_resolve() [all …]
|
/drivers/infiniband/hw/usnic/ |
D | usnic_fwd.h | 111 uint32_t daddr, uint16_t dport) in usnic_fwd_init_udp_filter() argument 117 if (daddr) { in usnic_fwd_init_udp_filter() 119 filter->u.ipv4.dst_addr = daddr; in usnic_fwd_init_udp_filter()
|
/drivers/scsi/cxgbi/ |
D | libcxgbi.c | 587 __be32 saddr, __be32 daddr, in find_route_ipv4() argument 592 rt = ip_route_output_ports(&init_net, fl4, NULL, daddr, saddr, in find_route_ipv4() 602 struct sockaddr_in *daddr = (struct sockaddr_in *)dst_addr; in cxgbi_check_route() local 614 rt = find_route_ipv4(&fl4, 0, daddr->sin_addr.s_addr, 0, daddr->sin_port, 0); in cxgbi_check_route() 617 be32_to_cpu(daddr->sin_addr.s_addr), in cxgbi_check_route() 618 be16_to_cpu(daddr->sin_port)); in cxgbi_check_route() 623 n = dst_neigh_lookup(dst, &daddr->sin_addr.s_addr); in cxgbi_check_route() 632 &daddr->sin_addr.s_addr, ntohs(daddr->sin_port), in cxgbi_check_route() 639 ndev = ip_dev_find(&init_net, daddr->sin_addr.s_addr); in cxgbi_check_route() 648 &daddr->sin_addr.s_addr, ndev->name); in cxgbi_check_route() [all …]
|
/drivers/misc/mic/vop/ |
D | vop_vringh.c | 527 size_t len, u64 daddr, size_t dlen, in vop_virtio_copy_to_user() argument 531 void __iomem *dbuf = vpdev->hw_ops->ioremap(vpdev, daddr, len); in vop_virtio_copy_to_user() 551 dma_offset = daddr - round_down(daddr, dma_alignment); in vop_virtio_copy_to_user() 552 daddr -= dma_offset; in vop_virtio_copy_to_user() 561 daddr += vpdev->aper->pa; in vop_virtio_copy_to_user() 564 err = vop_sync_dma(vdev, vvr->buf_da, daddr, in vop_virtio_copy_to_user() 578 daddr += partlen; in vop_virtio_copy_to_user() 602 size_t len, u64 daddr, size_t dlen, in vop_virtio_copy_from_user() argument 606 void __iomem *dbuf = vpdev->hw_ops->ioremap(vpdev, daddr, len); in vop_virtio_copy_from_user() 615 if (daddr & (dma_alignment - 1)) { in vop_virtio_copy_from_user() [all …]
|
/drivers/video/fbdev/ |
D | hitfb.c | 79 u32 saddr, daddr; in hitfb_accel_bitblt() local 87 daddr = WIDTH * (dy + height) + dx + width; in hitfb_accel_bitblt() 100 daddr = WIDTH * dy + dx; in hitfb_accel_bitblt() 109 daddr <<= 1; in hitfb_accel_bitblt() 115 fb_writew(daddr & 0xffff, HD64461_BBTDSARL); in hitfb_accel_bitblt() 116 fb_writew(daddr >> 16, HD64461_BBTDSARH); in hitfb_accel_bitblt()
|
/drivers/net/ipvlan/ |
D | ipvlan_core.c | 326 i6addr = use_dest ? &ip6h->daddr : &ip6h->saddr; in ipvlan_addr_lookup() 345 i4addr = use_dest ? &ip4h->daddr : &ip4h->saddr; in ipvlan_addr_lookup() 378 .daddr = ip4h->daddr, in ipvlan_process_v4_outbound() 413 .daddr = ip6h->daddr, in ipvlan_process_v6_outbound() 725 err = ip_route_input_noref(skb, ip4h->daddr, ip4h->saddr, in ipvlan_l3_rcv() 738 .daddr = ip6h->daddr, in ipvlan_l3_rcv()
|
/drivers/net/arcnet/ |
D | arcnet.c | 64 unsigned short type, uint8_t daddr); 104 unsigned short type, const void *daddr, 544 unsigned short type, const void *daddr, in arcnet_header() argument 554 daddr ? *(uint8_t *)daddr : -1, in arcnet_header() 566 _daddr = daddr ? *(uint8_t *)daddr : 0; in arcnet_header() 567 } else if (!daddr) { in arcnet_header() 581 _daddr = *(uint8_t *)daddr; in arcnet_header() 1075 unsigned short type, uint8_t daddr) in null_build_header() argument 1081 lp->default_proto[daddr]); in null_build_header()
|
D | rfc1051.c | 43 unsigned short type, uint8_t daddr); 162 unsigned short type, uint8_t daddr) in build_header() argument 202 pkt->hard.dest = daddr; in build_header()
|
D | arc-rawmode.c | 85 unsigned short type, uint8_t daddr) in build_header() argument 108 pkt->hard.dest = daddr; in build_header()
|
D | capmode.c | 101 uint8_t daddr) in build_header() argument 127 pkt->hard.dest = daddr; in build_header()
|
/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 219 req->peer_port = csk->daddr.sin_port; in send_act_open_req() 221 req->peer_ip = csk->daddr.sin_addr.s_addr; in send_act_open_req() 243 req->peer_port = csk->daddr.sin_port; in send_act_open_req() 245 req->peer_ip = csk->daddr.sin_addr.s_addr; in send_act_open_req() 267 (&csk->saddr), (&csk->daddr), t4 ? 4 : 5, csk, in send_act_open_req() 354 &csk->daddr6.sin6_addr, ntohs(csk->daddr.sin_port), in send_act_open_req6() 741 (&csk->saddr), (&csk->daddr), in do_act_establish() 887 "csk 0x%p,%u,0x%lx. ", (&csk->saddr), (&csk->daddr), in do_act_open_rpl() 931 (&csk->saddr), (&csk->daddr), in do_peer_close() 952 (&csk->saddr), (&csk->daddr), in do_close_con_rpl() [all …]
|
/drivers/net/ |
D | vrf.c | 162 .daddr = iph->daddr, in vrf_process_v6_outbound() 269 .daddr = ip4h->daddr, in vrf_process_v4_outbound() 385 nexthop = rt6_nexthop((struct rt6_info *)dst, &ipv6_hdr(skb)->daddr); in vrf_finish_output6() 424 if (rt6_need_strict(&ipv6_hdr(skb)->daddr)) in vrf_ip6_out() 584 nexthop = (__force u32)rt_nexthop(rt, ip_hdr(skb)->daddr); in vrf_finish_output() 933 .daddr = iph->daddr, in vrf_ip6_input_dst() 974 need_strict = rt6_need_strict(&ipv6_hdr(skb)->daddr); in vrf_ip6_rcv()
|
D | geneve.c | 751 fl4->daddr = info->key.u.ipv4.dst; in geneve_get_v4_rt() 765 fl4->daddr = geneve->remote.sin.sin_addr.s_addr; in geneve_get_v4_rt() 777 netdev_dbg(dev, "no route to %pI4\n", &fl4->daddr); in geneve_get_v4_rt() 781 netdev_dbg(dev, "circular route to %pI4\n", &fl4->daddr); in geneve_get_v4_rt() 812 fl6->daddr = info->key.u.ipv6.dst; in geneve_get_v6_dst() 828 fl6->daddr = geneve->remote.sin6.sin6_addr; in geneve_get_v6_dst() 839 netdev_dbg(dev, "no route to %pI6\n", &fl6->daddr); in geneve_get_v6_dst() 843 netdev_dbg(dev, "circular route to %pI6\n", &fl6->daddr); in geneve_get_v6_dst() 934 if (!ttl && IN_MULTICAST(ntohl(fl4.daddr))) in geneve_xmit_skb() 939 udp_tunnel_xmit_skb(rt, gs4->sock->sk, skb, fl4.saddr, fl4.daddr, in geneve_xmit_skb() [all …]
|
D | vxlan.c | 1488 u8 *daddr; in vxlan_na_create() local 1510 daddr = eth_hdr(request)->h_source; in vxlan_na_create() 1514 daddr = ns->opt + i + sizeof(struct nd_opt_hdr); in vxlan_na_create() 1520 ether_addr_copy(eth_hdr(reply)->h_dest, daddr); in vxlan_na_create() 1537 pip6->daddr = ipv6_hdr(request)->saddr; in vxlan_na_create() 1557 &pip6->daddr, sizeof(*na)+na_olen, IPPROTO_ICMPV6, in vxlan_na_create() 1574 const struct in6_addr *saddr, *daddr; in neigh_reduce() local 1584 daddr = &iphdr->daddr; in neigh_reduce() 1591 if (ipv6_addr_loopback(daddr) || in neigh_reduce() 1656 n = neigh_lookup(&arp_tbl, &pip->daddr, dev); in route_shortcircuit() [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 195 dma_addr_t daddr; in _c4iw_write_mem_dma() local 198 daddr = dma_map_single(&rdev->lldi.pdev->dev, data, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma() 199 if (dma_mapping_error(&rdev->lldi.pdev->dev, daddr)) in _c4iw_write_mem_dma() 201 save = daddr; in _c4iw_write_mem_dma() 212 ret = _c4iw_write_mem_dma_aligned(rdev, addr, dmalen, daddr, in _c4iw_write_mem_dma() 218 daddr += dmalen; in _c4iw_write_mem_dma()
|
/drivers/gpu/drm/i915/ |
D | i915_gem_gtt.c | 340 p->daddr = dma_map_page(kdev, in __setup_page_dma() 343 if (dma_mapping_error(kdev, p->daddr)) { in __setup_page_dma() 363 dma_unmap_page(&pdev->dev, p->daddr, 4096, PCI_DMA_BIDIRECTIONAL); in cleanup_page_dma() 474 scratch_pte = gen8_pte_encode(vm->scratch_page.daddr, in gen8_initialize_pt() 485 WARN_ON(vm->scratch_page.daddr == 0); in gen6_initialize_pt() 487 scratch_pte = vm->pte_encode(vm->scratch_page.daddr, in gen6_initialize_pt() 771 gen8_pte_t scratch_pte = gen8_pte_encode(vm->scratch_page.daddr, in gen8_ppgtt_clear_range() 925 u64 daddr = px_dma(&ppgtt->pml4); in gen8_ppgtt_notify_vgt() local 927 I915_WRITE(vgtif_reg(pdp[0].lo), lower_32_bits(daddr)); in gen8_ppgtt_notify_vgt() 928 I915_WRITE(vgtif_reg(pdp[0].hi), upper_32_bits(daddr)); in gen8_ppgtt_notify_vgt() [all …]
|