| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_span.c | 140 const struct net_device *to_dev, in mlxsw_sp1_span_entry_cpu_parms() argument 451 const struct net_device *to_dev, in mlxsw_sp_span_entry_gretap4_parms() argument 462 if (!(to_dev->flags & IFF_UP) || in mlxsw_sp_span_entry_gretap4_parms() 567 if (!(to_dev->flags & IFF_UP) || in mlxsw_sp_span_entry_gretap6_parms() 640 if (!(to_dev->flags & IFF_UP)) in mlxsw_sp_span_entry_vlan_parms() 882 span_entry->to_dev = to_dev; in mlxsw_sp_span_entry_create() 906 if (refcount_read(&curr->ref_count) && curr->to_dev == to_dev) in mlxsw_sp_span_entry_find_by_port() 943 if (refcount_read(&curr->ref_count) && curr->to_dev == to_dev && in mlxsw_sp_span_entry_find_by_parms() 954 const struct net_device *to_dev, in mlxsw_sp_span_entry_get() argument 1018 const struct net_device *to_dev) in mlxsw_sp_span_entry_ops() argument [all …]
|
| A D | spectrum_span.h | 56 const struct net_device *to_dev; member 71 const struct net_device *to_dev; member 80 bool (*can_handle)(const struct net_device *to_dev); 82 const struct net_device *to_dev, 95 const struct net_device *to_dev);
|
| A D | spectrum_matchall.c | 36 if (!mall_entry->mirror.to_dev) { in mlxsw_sp_mall_port_mirror_add() 41 agent_parms.to_dev = mall_entry->mirror.to_dev; in mlxsw_sp_mall_port_mirror_add() 271 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_mall_replace() 418 .to_dev = NULL, /* Mirror to CPU. */ in mlxsw_sp2_mall_sample_add()
|
| A D | spectrum_acl_flex_actions.c | 144 agent_parms.to_dev = out_dev; in mlxsw_sp_act_mirror_add()
|
| A D | spectrum_qdisc.c | 1834 .to_dev = mall_entry->mirror.to_dev, in mlxsw_sp_qevent_mirror_configure() 2031 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_qevent_mall_replace()
|
| A D | spectrum.h | 1196 const struct net_device *to_dev; member
|
| /linux/drivers/dma/ioat/ |
| A D | dma.c | 80 dev_err(to_dev(ioat_chan), "Err(%d): %s\n", in ioat_print_chanerrs() 177 dev_dbg(to_dev(ioat_chan), in __ioat_issue_pending() 213 dev_err(to_dev(ioat_chan), in __ioat_start_null_desc() 218 dev_dbg(to_dev(ioat_chan), in __ioat_start_null_desc() 256 dev_dbg(to_dev(ioat_chan), in __ioat_restart_chan() 398 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring() 420 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring() 480 dev_dbg_ratelimited(to_dev(ioat_chan), in ioat_check_space_lock() 919 dev_err(to_dev(ioat_chan), "Errors:\n"); in ioat_timer_event() 960 dev_err(to_dev(ioat_chan), "Errors:\n"); in ioat_timer_event() [all …]
|
| A D | prep.c | 279 struct device *dev = to_dev(ioat_chan); in dump_pq_desc_dbg() 306 struct device *dev = to_dev(ioat_chan); in dump_pq16_desc_dbg() 358 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq_lock() 480 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq16_lock() 507 dev_err(to_dev(ioat_chan), in __ioat_prep_pq16_lock()
|
| A D | dma.h | 23 #define to_dev(ioat_chan) (&(ioat_chan)->ioat_dma->pdev->dev) macro 222 struct device *dev = to_dev(ioat_chan); in __dump_desc_dbg()
|
| A D | sysfs.c | 88 dev_warn(to_dev(ioat_chan), in ioat_kobject_add()
|
| A D | init.c | 626 dev_dbg(to_dev(ioat_chan), "freeing %d idle descriptors\n", descs); in ioat_free_chan_resources() 633 dev_err(to_dev(ioat_chan), "Freeing %d in use descriptors!\n", in ioat_free_chan_resources() 643 dma_free_coherent(to_dev(ioat_chan), IOAT_CHUNK_SIZE, in ioat_free_chan_resources() 747 dev_WARN(to_dev(ioat_chan), in ioat_alloc_chan_resources()
|
| /linux/samples/bpf/ |
| A D | xdp_sample.bpf.c | 207 const struct net_device *to_dev, int sent, int drops, int err) in BPF_PROG() argument 214 idx_out = to_dev->ifindex; in BPF_PROG() 238 const struct net_device *to_dev, int sent, int drops, int err) in BPF_PROG() argument 246 idx_out = to_dev->ifindex; in BPF_PROG()
|
| /linux/drivers/nvdimm/ |
| A D | btt.c | 29 static struct device *to_dev(struct arena_info *arena) in to_dev() function 102 dev_err_ratelimited(to_dev(arena), in __btt_map_write() 143 dev_err_ratelimited(to_dev(arena), in btt_map_write() 161 dev_err_ratelimited(to_dev(arena), in btt_map_read() 342 dev_err(to_dev(arena), in btt_log_read() 431 dev_WARN_ONCE(to_dev(arena), size < 512, in btt_map_init() 475 dev_WARN_ONCE(to_dev(arena), size < 512, in btt_log_init() 570 dev_err_ratelimited(to_dev(arena), in btt_freelist_init() 878 dev_err(to_dev(arena), in discover_arenas() 890 dev_err(to_dev(arena), in discover_arenas() [all …]
|
| A D | pmem.c | 35 static struct device *to_dev(struct pmem_device *pmem) in to_dev() function 46 return to_nd_region(to_dev(pmem)->parent); in to_region() 102 long cleared = nvdimm_clear_poison(to_dev(pmem), phys, len); in __pmem_clear_poison()
|
| /linux/tools/testing/selftests/net/forwarding/ |
| A D | mirror_lib.sh | 7 local to_dev=$1; shift 12 action mirred egress mirror dev $to_dev
|
| /linux/include/trace/events/ |
| A D | xdp.h | 271 const struct net_device *to_dev, 274 TP_ARGS(from_dev, to_dev, sent, drops, err), 288 __entry->to_ifindex = to_dev->ifindex;
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | core_kern.c | 44 *from_dev, const struct net_device *to_dev, int sent, int drops, in BPF_PROG() argument
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_topology.c | 1248 static void kfd_set_iolink_non_coherent(struct kfd_topology_device *to_dev, in kfd_set_iolink_non_coherent() argument 1253 if (!to_dev->gpu && in kfd_set_iolink_non_coherent() 1257 if (to_dev->gpu) { in kfd_set_iolink_non_coherent() 1263 KFD_GC_VERSION(to_dev->gpu) == IP_VERSION(9, 4, 0))) { in kfd_set_iolink_non_coherent() 1281 static void kfd_set_recommended_sdma_engines(struct kfd_topology_device *to_dev, in kfd_set_recommended_sdma_engines() argument 1288 bool support_rec_eng = !amdgpu_sriov_vf(adev) && to_dev->gpu && in kfd_set_recommended_sdma_engines() 1295 int dst_socket_id = to_dev->gpu->adev->gmc.xgmi.physical_node_id; in kfd_set_recommended_sdma_engines() 1306 kfd_get_num_xgmi_sdma_engines(gpu) && to_dev->gpu) { in kfd_set_recommended_sdma_engines()
|
| A D | kfd_crat.c | 1233 struct kfd_topology_device *dev, *to_dev; in kfd_parse_subtype_iolink() local 1283 to_dev = kfd_topology_device_by_proximity_domain_no_lock(id_to); in kfd_parse_subtype_iolink() 1284 if (!to_dev) in kfd_parse_subtype_iolink() 1294 to_dev->node_props.io_links_count++; in kfd_parse_subtype_iolink() 1295 list_add_tail(&props2->list, &to_dev->io_link_props); in kfd_parse_subtype_iolink()
|