/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_span.c | 140 const struct net_device *to_dev, in mlxsw_sp1_span_entry_cpu_parms() argument 169 const struct net_device *to_dev, in mlxsw_sp_span_entry_phys_parms() argument 172 sparmsp->dest_port = netdev_priv(to_dev); in mlxsw_sp_span_entry_phys_parms() 411 mlxsw_sp_span_gretap4_route(const struct net_device *to_dev, in mlxsw_sp_span_gretap4_route() argument 414 struct ip_tunnel *tun = netdev_priv(to_dev); in mlxsw_sp_span_gretap4_route() 423 parms = mlxsw_sp_ipip_netdev_parms4(to_dev); in mlxsw_sp_span_gretap4_route() 449 const struct net_device *to_dev, in mlxsw_sp_span_entry_gretap4_parms() argument 452 struct ip_tunnel_parm tparm = mlxsw_sp_ipip_netdev_parms4(to_dev); in mlxsw_sp_span_entry_gretap4_parms() 460 if (!(to_dev->flags & IFF_UP) || in mlxsw_sp_span_entry_gretap4_parms() 469 l3edev = mlxsw_sp_span_gretap4_route(to_dev, &saddr.addr4, &gw.addr4); in mlxsw_sp_span_entry_gretap4_parms() [all …]
|
D | spectrum_span.h | 41 const struct net_device *to_dev; member 55 const struct net_device *to_dev; member 64 bool (*can_handle)(const struct net_device *to_dev); 66 const struct net_device *to_dev, 79 const struct net_device *to_dev);
|
D | spectrum_matchall.c | 35 if (!mall_entry->mirror.to_dev) { in mlxsw_sp_mall_port_mirror_add() 40 agent_parms.to_dev = mall_entry->mirror.to_dev; in mlxsw_sp_mall_port_mirror_add() 238 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_mall_replace()
|
D | spectrum_acl_flex_actions.c | 144 agent_parms.to_dev = out_dev; in mlxsw_sp_act_mirror_add()
|
D | spectrum_qdisc.c | 1389 .to_dev = mall_entry->mirror.to_dev, in mlxsw_sp_qevent_mirror_configure() 1571 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_qevent_mall_replace()
|
D | spectrum.h | 1046 const struct net_device *to_dev; member
|
/drivers/dma/ioat/ |
D | dma.c | 80 dev_err(to_dev(ioat_chan), "Err(%d): %s\n", in ioat_print_chanerrs() 177 dev_dbg(to_dev(ioat_chan), in __ioat_issue_pending() 213 dev_err(to_dev(ioat_chan), in __ioat_start_null_desc() 218 dev_dbg(to_dev(ioat_chan), in __ioat_start_null_desc() 256 dev_dbg(to_dev(ioat_chan), in __ioat_restart_chan() 317 dev_dbg(to_dev(ioat_chan), "%s: cookie: %d\n", __func__, cookie); in ioat_tx_submit_unlock() 391 descs->virt = dma_alloc_coherent(to_dev(ioat_chan), in ioat_alloc_ring() 398 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring() 420 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring() 472 dev_dbg(to_dev(ioat_chan), "%s: num_descs: %d (%x:%x:%x)\n", in ioat_check_space_lock() [all …]
|
D | prep.c | 279 struct device *dev = to_dev(ioat_chan); in dump_pq_desc_dbg() 306 struct device *dev = to_dev(ioat_chan); in dump_pq16_desc_dbg() 358 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq_lock() 480 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq16_lock() 507 dev_err(to_dev(ioat_chan), in __ioat_prep_pq16_lock()
|
D | dma.h | 23 #define to_dev(ioat_chan) (&(ioat_chan)->ioat_dma->pdev->dev) macro 223 struct device *dev = to_dev(ioat_chan); in __dump_desc_dbg()
|
D | sysfs.c | 88 dev_warn(to_dev(ioat_chan), in ioat_kobject_add()
|
D | init.c | 637 dev_dbg(to_dev(ioat_chan), "freeing %d idle descriptors\n", descs); in ioat_free_chan_resources() 644 dev_err(to_dev(ioat_chan), "Freeing %d in use descriptors!\n", in ioat_free_chan_resources() 654 dma_free_coherent(to_dev(ioat_chan), IOAT_CHUNK_SIZE, in ioat_free_chan_resources() 758 dev_WARN(to_dev(ioat_chan), in ioat_alloc_chan_resources()
|
/drivers/nvdimm/ |
D | btt.c | 27 static struct device *to_dev(struct arena_info *arena) in to_dev() function 68 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->infooff, 512), in btt_info_write() 70 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->info2off, 512), in btt_info_write() 100 dev_err_ratelimited(to_dev(arena), in __btt_map_write() 141 dev_err_ratelimited(to_dev(arena), in btt_map_write() 159 dev_err_ratelimited(to_dev(arena), in btt_map_read() 340 dev_err(to_dev(arena), in btt_log_read() 423 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->mapoff, 512), in btt_map_init() 429 dev_WARN_ONCE(to_dev(arena), size < 512, in btt_map_init() 467 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->logoff, 512), in btt_log_init() [all …]
|
D | pmem.c | 33 static struct device *to_dev(struct pmem_device *pmem) in to_dev() function 44 return to_nd_region(to_dev(pmem)->parent); in to_region() 74 struct device *dev = to_dev(pmem); in pmem_clear_poison()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_crat.c | 355 struct kfd_topology_device *dev, *to_dev; in kfd_parse_subtype_iolink() local 406 to_dev = kfd_topology_device_by_proximity_domain(id_to); in kfd_parse_subtype_iolink() 407 if (!to_dev) in kfd_parse_subtype_iolink() 414 to_dev->io_link_count++; in kfd_parse_subtype_iolink() 415 to_dev->node_props.io_links_count++; in kfd_parse_subtype_iolink() 416 list_add_tail(&props2->list, &to_dev->io_link_props); in kfd_parse_subtype_iolink()
|
/drivers/media/i2c/ |
D | mt9m032.c | 155 #define to_dev(sensor) \ macro 175 dev_dbg(to_dev(sensor), "MT9M032 line time: %u ns\n", ns); in mt9m032_row_time()
|