/drivers/edac/ |
D | pnd2_edac.c | 93 struct dram_addr *daddr, char *msg); 927 struct dram_addr *daddr, char *msg) in apl_pmi2mem() argument 984 daddr->col = column; in apl_pmi2mem() 985 daddr->bank = bank; in apl_pmi2mem() 986 daddr->row = row; in apl_pmi2mem() 987 daddr->rank = rank; in apl_pmi2mem() 988 daddr->dimm = 0; in apl_pmi2mem() 997 struct dram_addr *daddr, char *msg) in dnv_pmi2mem() argument 1000 daddr->rank = dnv_get_bit(pmiaddr, dmap[pmiidx].rs0 + 13, 0); in dnv_pmi2mem() 1002 daddr->rank |= dnv_get_bit(pmiaddr, dmap[pmiidx].rs1 + 13, 1); in dnv_pmi2mem() [all …]
|
/drivers/infiniband/sw/rxe/ |
D | rxe_net.c | 86 struct in_addr *daddr) in rxe_find_route4() argument 94 memcpy(&fl.daddr, daddr, sizeof(*daddr)); in rxe_find_route4() 99 pr_err_ratelimited("no route to %pI4\n", &daddr->s_addr); in rxe_find_route4() 109 struct in6_addr *daddr) in rxe_find_route6() argument 117 memcpy(&fl6.daddr, daddr, sizeof(*daddr)); in rxe_find_route6() 124 pr_err_ratelimited("no route to %pI6\n", daddr); in rxe_find_route6() 129 pr_err("no route to %pI6\n", daddr); in rxe_find_route6() 143 struct in6_addr *daddr) in rxe_find_route6() argument 165 struct in_addr *daddr; in rxe_find_route() local 168 daddr = &av->dgid_addr._sockaddr_in.sin_addr; in rxe_find_route() [all …]
|
D | rxe_recv.c | 171 struct in_addr *daddr = in check_addr() local 174 if (ip_hdr(skb)->daddr != saddr->s_addr) { in check_addr() 176 &ip_hdr(skb)->daddr, in check_addr() 181 if (ip_hdr(skb)->saddr != daddr->s_addr) { in check_addr() 184 &daddr->s_addr); in check_addr() 191 struct in6_addr *daddr = in check_addr() local 194 if (memcmp(&ipv6_hdr(skb)->daddr, saddr, sizeof(*saddr))) { in check_addr() 196 &ipv6_hdr(skb)->daddr, saddr); in check_addr() 200 if (memcmp(&ipv6_hdr(skb)->saddr, daddr, sizeof(*daddr))) { in check_addr() 202 &ipv6_hdr(skb)->saddr, daddr); in check_addr() [all …]
|
/drivers/staging/wlan-ng/ |
D | p80211conv.c | 178 memcpy(p80211_hdr->a3.a1, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 186 memcpy(p80211_hdr->a3.a3, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 190 memcpy(p80211_hdr->a3.a1, &e_hdr.daddr, ETH_ALEN); in skb_ether_to_p80211() 282 u8 daddr[ETH_ALEN]; in skb_p80211_to_ether() local 299 ether_addr_copy(daddr, w_hdr->a3.a1); in skb_p80211_to_ether() 303 ether_addr_copy(daddr, w_hdr->a3.a1); in skb_p80211_to_ether() 307 ether_addr_copy(daddr, w_hdr->a3.a3); in skb_p80211_to_ether() 316 ether_addr_copy(daddr, w_hdr->a4.a3); in skb_p80211_to_ether() 362 ((!ether_addr_equal_unaligned(daddr, e_hdr->daddr)) || in skb_p80211_to_ether() 409 ether_addr_copy(e_hdr->daddr, daddr); in skb_p80211_to_ether() [all …]
|
/drivers/net/ethernet/intel/ixgbevf/ |
D | ipsec.c | 34 memcpy(sam->addr, &xs->id.daddr.a6, sizeof(xs->id.daddr.a6)); in ixgbevf_ipsec_set_pf_sa() 36 memcpy(sam->addr, &xs->id.daddr.a4, sizeof(xs->id.daddr.a4)); in ixgbevf_ipsec_set_pf_sa() 179 __be32 *daddr, u8 proto, in ixgbevf_ipsec_find_rx_state() argument 189 ((ip4 && *daddr == rsa->xs->id.daddr.a4) || in ixgbevf_ipsec_find_rx_state() 190 (!ip4 && !memcmp(daddr, &rsa->xs->id.daddr.a6, in ixgbevf_ipsec_find_rx_state() 191 sizeof(rsa->xs->id.daddr.a6)))) && in ixgbevf_ipsec_find_rx_state() 307 memcpy(rsa.ipaddr, &xs->id.daddr.a6, 16); in ixgbevf_ipsec_add_sa() 309 memcpy(&rsa.ipaddr[3], &xs->id.daddr.a4, 4); in ixgbevf_ipsec_add_sa() 552 void *daddr; in ixgbevf_ipsec_rx() local 564 daddr = &ip4->daddr; in ixgbevf_ipsec_rx() [all …]
|
/drivers/gpu/drm/radeon/ |
D | radeon_benchmark.c | 36 uint64_t saddr, uint64_t daddr, in radeon_benchmark_do_move() argument 49 fence = radeon_copy_dma(rdev, saddr, daddr, in radeon_benchmark_do_move() 54 fence = radeon_copy_blit(rdev, saddr, daddr, in radeon_benchmark_do_move() 92 uint64_t saddr, daddr; in radeon_benchmark_move() local 116 r = radeon_bo_pin(dobj, ddomain, &daddr); in radeon_benchmark_move() 123 time = radeon_benchmark_do_move(rdev, size, saddr, daddr, in radeon_benchmark_move() 134 time = radeon_benchmark_do_move(rdev, size, saddr, daddr, in radeon_benchmark_move()
|
/drivers/net/ethernet/chelsio/libcxgb/ |
D | libcxgb_cm.c | 58 __func__, ntohl(ip->saddr), ntohl(ip->daddr), in cxgb_get_4tuple() 62 memcpy(local_ip, &ip->daddr, 4); in cxgb_get_4tuple() 65 __func__, ip6->saddr.s6_addr, ip6->daddr.s6_addr, in cxgb_get_4tuple() 69 memcpy(local_ip, ip6->daddr.s6_addr, 16); in cxgb_get_4tuple() 131 memcpy(&fl6.daddr, peer_ip, 16); in cxgb_find_route6() 133 if (ipv6_addr_type(&fl6.daddr) & IPV6_ADDR_LINKLOCAL) in cxgb_find_route6()
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_ipsec.c | 393 __be32 *daddr, u8 proto, in ixgbe_ipsec_find_rx_state() argument 405 ((ip4 && *daddr == rsa->xs->id.daddr.a4) || in ixgbe_ipsec_find_rx_state() 406 (!ip4 && !memcmp(daddr, &rsa->xs->id.daddr.a6, in ixgbe_ipsec_find_rx_state() 407 sizeof(rsa->xs->id.daddr.a6)))) && in ixgbe_ipsec_find_rx_state() 514 if (reg == xs->id.daddr.a4) in ixgbe_ipsec_check_mgmt_ip() 521 if (reg == xs->id.daddr.a4) in ixgbe_ipsec_check_mgmt_ip() 536 if (reg != xs->id.daddr.a6[j]) in ixgbe_ipsec_check_mgmt_ip() 546 if (reg != xs->id.daddr.a6[j]) in ixgbe_ipsec_check_mgmt_ip() 615 memcpy(rsa.ipaddr, &xs->id.daddr.a6, 16); in ixgbe_ipsec_add_sa() 617 memcpy(&rsa.ipaddr[3], &xs->id.daddr.a4, 4); in ixgbe_ipsec_add_sa() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_benchmark.c | 32 uint64_t saddr, uint64_t daddr, int n) in amdgpu_benchmark_do_move() argument 42 r = amdgpu_copy_buffer(ring, saddr, daddr, size, NULL, &fence, in amdgpu_benchmark_do_move() 79 uint64_t saddr, daddr; in amdgpu_benchmark_move() local 127 daddr = amdgpu_bo_gpu_offset(dobj); in amdgpu_benchmark_move() 130 time = amdgpu_benchmark_do_move(adev, size, saddr, daddr, n); in amdgpu_benchmark_move()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_span.c | 238 union mlxsw_sp_l3addr daddr, in mlxsw_sp_span_entry_tunnel_parms_common() argument 248 gw = daddr; in mlxsw_sp_span_entry_tunnel_parms_common() 284 sparmsp->daddr = daddr; in mlxsw_sp_span_entry_tunnel_parms_common() 336 union mlxsw_sp_l3addr daddr = { .addr4 = tparm.iph.daddr }; in mlxsw_sp_span_entry_gretap4_parms() local 339 union mlxsw_sp_l3addr gw = daddr; in mlxsw_sp_span_entry_gretap4_parms() 348 mlxsw_sp_l3addr_is_zero(daddr)) in mlxsw_sp_span_entry_gretap4_parms() 352 return mlxsw_sp_span_entry_tunnel_parms_common(l3edev, saddr, daddr, gw, in mlxsw_sp_span_entry_gretap4_parms() 377 be32_to_cpu(sparms.daddr.addr4)); in mlxsw_sp_span_entry_gretap4_configure() 413 if (!ip6_tnl_xmit_ctl(t, &fl6.saddr, &fl6.daddr)) in mlxsw_sp_span_gretap6_route() 438 union mlxsw_sp_l3addr daddr = { .addr6 = tparm.raddr }; in mlxsw_sp_span_entry_gretap6_parms() local [all …]
|
/drivers/target/ |
D | target_core_sbc.c | 1228 void *daddr, *paddr; in sbc_dif_generate() local 1234 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1243 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1249 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1254 crc = crc_t10dif(daddr + offset, avail); in sbc_dif_generate() 1256 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1262 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1264 crc = crc_t10dif_update(crc, daddr, offset); in sbc_dif_generate() 1284 kunmap_atomic(daddr - dsg->offset); in sbc_dif_generate() 1383 void *daddr, *paddr; in sbc_dif_verify() local [all …]
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_main.c | 703 static void push_pseudo_header(struct sk_buff *skb, const char *daddr) in push_pseudo_header() argument 708 memcpy(phdr->hwaddr, daddr, INFINIBAND_ALEN); in push_pseudo_header() 821 if (ipoib_cm_enabled(dev, neigh->daddr)) { in path_rec_completion() 928 static void neigh_refresh_path(struct ipoib_neigh *neigh, u8 *daddr, in neigh_refresh_path() argument 937 path = __path_find(dev, daddr + 4); in neigh_refresh_path() 946 static struct ipoib_neigh *neigh_add_path(struct sk_buff *skb, u8 *daddr, in neigh_add_path() argument 956 neigh = ipoib_neigh_alloc(daddr, dev); in neigh_add_path() 972 path = __path_find(dev, daddr + 4); in neigh_add_path() 974 path = path_rec_create(dev, daddr + 4); in neigh_add_path() 987 if (ipoib_cm_enabled(dev, neigh->daddr)) { in neigh_add_path() [all …]
|
/drivers/misc/genwqe/ |
D | card_utils.c | 248 dma_addr_t daddr; in genwqe_map_pages() local 251 daddr = pci_map_page(pci_dev, page_list[i], in genwqe_map_pages() 256 if (pci_dma_mapping_error(pci_dev, daddr)) { in genwqe_map_pages() 259 __func__, (long long)daddr); in genwqe_map_pages() 263 dma_list[i] = daddr; in genwqe_map_pages() 386 dma_addr_t daddr; in genwqe_setup_sgl() local 401 daddr = sgl->fpage_dma_addr + map_offs; in genwqe_setup_sgl() 405 daddr = sgl->lpage_dma_addr; in genwqe_setup_sgl() 407 daddr = dma_list[p] + map_offs; in genwqe_setup_sgl() 413 if (prev_daddr == daddr) { in genwqe_setup_sgl() [all …]
|
/drivers/infiniband/core/ |
D | addr.c | 146 const void *daddr, in ib_nl_ip_send_msg() argument 182 nla_put(skb, attrtype, size, daddr); in ib_nl_ip_send_msg() 317 const void *daddr, u32 seq, u16 family) in ib_nl_fetch_ha() argument 322 return ib_nl_ip_send_msg(dev_addr, daddr, seq, family); in ib_nl_fetch_ha() 327 const void *daddr) in dst_fetch_ha() argument 332 n = dst_neigh_lookup(dst, daddr); in dst_fetch_ha() 369 const void *daddr = (dst_in->sa_family == AF_INET) ? in fetch_ha() local 376 return ib_nl_fetch_ha(dev_addr, daddr, seq, family); in fetch_ha() 378 return dst_fetch_ha(dst, dev_addr, daddr); in fetch_ha() 397 fl4.daddr = dst_ip; in addr4_resolve() [all …]
|
/drivers/staging/media/ipu3/ |
D | ipu3-css.h | 70 dma_addr_t daddr; member 207 unsigned int queue, dma_addr_t daddr) in imgu_css_buf_init() argument 211 b->daddr = daddr; in imgu_css_buf_init()
|
D | ipu3-dmamap.c | 137 map->daddr = iova_dma_addr(&imgu->iova_domain, iova); in imgu_dmamap_alloc() 141 size, &map->daddr, map->vma->addr); in imgu_dmamap_alloc() 165 iova_pfn(&imgu->iova_domain, map->daddr)); in imgu_dmamap_unmap() 183 __func__, map->size, &map->daddr, map->vaddr); in imgu_dmamap_free() 234 map->daddr = iova_dma_addr(&imgu->iova_domain, iova); in imgu_dmamap_map_sg()
|
/drivers/net/ethernet/aquantia/atlantic/ |
D | aq_ring.c | 21 dma_unmap_page(dev, rxpage->daddr, len, DMA_FROM_DEVICE); in aq_free_rxpage() 32 dma_addr_t daddr; in aq_get_rxpage() local 39 daddr = dma_map_page(dev, page, 0, PAGE_SIZE << order, in aq_get_rxpage() 42 if (unlikely(dma_mapping_error(dev, daddr))) in aq_get_rxpage() 46 rxpage->daddr = daddr; in aq_get_rxpage() 344 buff->rxdata.daddr, in aq_ring_rx_clean() 392 buff_->rxdata.daddr, in aq_ring_rx_clean()
|
/drivers/misc/mic/vop/ |
D | vop_vringh.c | 515 size_t len, u64 daddr, size_t dlen, in vop_virtio_copy_to_user() argument 519 void __iomem *dbuf = vpdev->hw_ops->remap(vpdev, daddr, len); in vop_virtio_copy_to_user() 542 dma_offset = daddr - round_down(daddr, dma_alignment); in vop_virtio_copy_to_user() 543 daddr -= dma_offset; in vop_virtio_copy_to_user() 552 daddr += vpdev->aper->pa; in vop_virtio_copy_to_user() 555 err = vop_sync_dma(vdev, vvr->buf_da, daddr, in vop_virtio_copy_to_user() 569 daddr += partlen; in vop_virtio_copy_to_user() 593 size_t len, u64 daddr, size_t dlen, in vop_virtio_copy_from_user() argument 597 void __iomem *dbuf = vpdev->hw_ops->remap(vpdev, daddr, len); in vop_virtio_copy_from_user() 610 if (daddr & (dma_alignment - 1)) { in vop_virtio_copy_from_user() [all …]
|
/drivers/gpu/drm/tegra/ |
D | falcon.c | 62 dma_addr_t daddr; in falcon_copy_firmware_image() local 71 daddr = dma_map_single(falcon->dev, firmware_vaddr, in falcon_copy_firmware_image() 73 err = dma_mapping_error(falcon->dev, daddr); in falcon_copy_firmware_image() 78 dma_sync_single_for_device(falcon->dev, daddr, in falcon_copy_firmware_image() 80 dma_unmap_single(falcon->dev, daddr, falcon->firmware.size, in falcon_copy_firmware_image()
|
/drivers/infiniband/hw/usnic/ |
D | usnic_fwd.h | 112 uint32_t daddr, uint16_t dport) in usnic_fwd_init_udp_filter() argument 118 if (daddr) { in usnic_fwd_init_udp_filter() 120 filter->u.ipv4.dst_addr = daddr; in usnic_fwd_init_udp_filter()
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | tc_tun.c | 116 n = dst_neigh_lookup(&rt->dst, &fl4->daddr); in mlx5e_route_lookup_ipv4() 164 n = dst_neigh_lookup(dst, &fl6->daddr); in mlx5e_route_lookup_ipv6() 225 fl4.daddr = tun_key->u.ipv4.dst; in mlx5e_tc_tun_create_header_ipv4() 284 ip->daddr = fl4.daddr; in mlx5e_tc_tun_create_header_ipv4() 345 fl6.daddr = tun_key->u.ipv6.dst; in mlx5e_tc_tun_create_header_ipv6() 402 ip6h->daddr = fl6.daddr; in mlx5e_tc_tun_create_header_ipv6()
|
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ipsec.c | 287 __be32 saddr[4] = {0}, daddr[4] = {0}, spi; in mlx5e_xfrm_add_state() local 336 daddr[3] = x->id.daddr.a4; in mlx5e_xfrm_add_state() 339 memcpy(daddr, x->id.daddr.a6, sizeof(daddr)); in mlx5e_xfrm_add_state() 346 saddr, daddr, spi, in mlx5e_xfrm_add_state()
|
/drivers/scsi/cxgbi/ |
D | libcxgbi.c | 587 __be32 saddr, __be32 daddr, in find_route_ipv4() argument 593 rt = ip_route_output_ports(&init_net, fl4, NULL, daddr, saddr, in find_route_ipv4() 604 struct sockaddr_in *daddr = (struct sockaddr_in *)dst_addr; in cxgbi_check_route() local 616 rt = find_route_ipv4(&fl4, 0, daddr->sin_addr.s_addr, 0, in cxgbi_check_route() 617 daddr->sin_port, 0, ifindex); in cxgbi_check_route() 620 be32_to_cpu(daddr->sin_addr.s_addr), in cxgbi_check_route() 621 be16_to_cpu(daddr->sin_port)); in cxgbi_check_route() 626 n = dst_neigh_lookup(dst, &daddr->sin_addr.s_addr); in cxgbi_check_route() 635 &daddr->sin_addr.s_addr, ntohs(daddr->sin_port), in cxgbi_check_route() 642 ndev = ip_dev_find(&init_net, daddr->sin_addr.s_addr); in cxgbi_check_route() [all …]
|
/drivers/video/fbdev/ |
D | hitfb.c | 79 u32 saddr, daddr; in hitfb_accel_bitblt() local 87 daddr = WIDTH * (dy + height) + dx + width; in hitfb_accel_bitblt() 100 daddr = WIDTH * dy + dx; in hitfb_accel_bitblt() 109 daddr <<= 1; in hitfb_accel_bitblt() 115 fb_writew(daddr & 0xffff, HD64461_BBTDSARL); in hitfb_accel_bitblt() 116 fb_writew(daddr >> 16, HD64461_BBTDSARH); in hitfb_accel_bitblt()
|
/drivers/net/ethernet/mellanox/mlx5/core/accel/ |
D | ipsec.h | 52 const __be32 daddr[4], 68 const __be32 daddr[4], in mlx5_accel_esp_create_hw_context()
|