Home
last modified time | relevance | path

Searched refs:master (Results 1 – 25 of 41) sorted by relevance

12

/net/sched/
A Dsch_teql.c136 if (!master) in teql_destroy()
139 prev = master->slaves; in teql_destroy()
145 if (q == master->slaves) { in teql_destroy()
152 master->slaves = NULL; in teql_destroy()
286 start = master->slaves; in teql_master_xmit()
319 master->tx_packets++; in teql_master_xmit()
347 master->tx_errors++; in teql_master_xmit()
350 master->tx_dropped++; in teql_master_xmit()
444 master->dev = dev; in teql_master_setup()
492 master = netdev_priv(dev); in teql_init()
[all …]
/net/hsr/
A Dhsr_main.c34 struct hsr_port *port, *master; in hsr_netdev_notify() local
75 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_netdev_notify()
78 eth_hw_addr_set(master->dev, dev->dev_addr); in hsr_netdev_notify()
80 master->dev); in hsr_netdev_notify()
95 master->dev->dev_addr, in hsr_netdev_notify()
98 master->dev->dev_addr); in hsr_netdev_notify()
100 netdev_warn(master->dev, in hsr_netdev_notify()
108 WRITE_ONCE(master->dev->mtu, mtu_max); in hsr_netdev_notify()
114 if (hsr_slave_empty(master->hsr)) { in hsr_netdev_notify()
117 ops = master->dev->rtnl_link_ops; in hsr_netdev_notify()
[all …]
A Dhsr_device.c54 netif_carrier_on(master->dev); in hsr_check_carrier()
59 netif_carrier_off(master->dev); in hsr_check_carrier()
90 struct hsr_port *master; in hsr_check_carrier_and_operstate() local
227 struct hsr_port *master; in hsr_dev_xmit() local
230 if (master) { in hsr_dev_xmit()
231 skb->dev = master->dev; in hsr_dev_xmit()
235 hsr_forward_skb(skb, master); in hsr_dev_xmit()
270 skb->dev = master->dev; in hsr_init_skb()
399 hsr_forward_skb(skb, master); in send_prp_supervision_frame()
408 struct hsr_port *master; in hsr_announce() local
[all …]
A Dhsr_slave.c141 struct hsr_port *master; in hsr_portdev_setup() local
153 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_portdev_setup()
154 hsr_dev = master->dev; in hsr_portdev_setup()
179 struct hsr_port *port, *master; in hsr_add_port() local
209 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_add_port()
210 netdev_update_features(master->dev); in hsr_add_port()
211 dev_set_mtu(master->dev, hsr_get_max_mtu(hsr)); in hsr_add_port()
223 struct hsr_port *master; in hsr_del_port() local
226 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_del_port()
229 if (port != master) { in hsr_del_port()
[all …]
A Dhsr_netlink.c215 struct hsr_port *master; in hsr_nl_ringerror() local
245 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_nl_ringerror()
246 netdev_warn(master->dev, "Could not send HSR ring error message\n"); in hsr_nl_ringerror()
257 struct hsr_port *master; in hsr_nl_nodedown() local
282 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_nl_nodedown()
283 netdev_warn(master->dev, "Could not send HSR node down\n"); in hsr_nl_nodedown()
/net/netfilter/
A Dnf_conntrack_expect.c199 if (!nf_ct_is_confirmed(exp->master)) in nf_ct_find_expectation()
222 nf_ct_put(exp->master); in nf_ct_find_expectation()
263 net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && in expect_clash()
264 nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); in expect_clash()
272 net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && in expect_matches()
273 nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); in expect_matches()
283 return a->master == b->master; in master_matches()
306 new->master = me; in nf_ct_expect_alloc()
437 struct nf_conn *master = expect->master; in __nf_ct_expect_check() local
454 i->master != expect->master) in __nf_ct_expect_check()
[all …]
A Dxt_quota.c30 struct xt_quota_priv *priv = q->master; in quota_mt()
53 q->master = kmalloc(sizeof(*q->master), GFP_KERNEL); in quota_mt_check()
54 if (q->master == NULL) in quota_mt_check()
57 spin_lock_init(&q->master->lock); in quota_mt_check()
58 q->master->quota = q->quota; in quota_mt_check()
66 kfree(q->master); in quota_mt_destroy()
77 .usersize = offsetof(struct xt_quota_info, master),
A Dxt_statistic.c42 oval = atomic_read(&info->master->count); in statistic_mt()
44 } while (atomic_cmpxchg(&info->master->count, oval, nval) != oval); in statistic_mt()
61 info->master = kzalloc(sizeof(*info->master), GFP_KERNEL); in statistic_mt_check()
62 if (info->master == NULL) in statistic_mt_check()
64 atomic_set(&info->master->count, info->u.nth.count); in statistic_mt_check()
73 kfree(info->master); in statistic_mt_destroy()
84 .usersize = offsetof(struct xt_statistic_info, master),
A Dxt_limit.c66 struct xt_limit_priv *priv = r->master; in limit_mt()
123 r->master = priv; in limit_mt_check()
140 kfree(info->master); in limit_mt_destroy()
152 u_int32_t master; member
163 .prev = cm->prev | (unsigned long)cm->master << 32, in limit_mt_compat_from_user()
181 .master = m->prev >> 32, in limit_mt_compat_to_user()
A Dnf_nat_amanda.c49 nf_ct_helper_log(skb, exp->master, "all ports in use"); in help()
54 if (!nf_nat_mangle_udp_packet(skb, exp->master, ctinfo, in help()
57 nf_ct_helper_log(skb, exp->master, "cannot mangle packet"); in help()
A Dnf_nat_tftp.c27 const struct nf_conn *ct = exp->master; in help()
34 nf_ct_helper_log(skb, exp->master, "cannot add expectation"); in help()
A Dxt_helper.c34 if (!ct || !ct->master) in helper_mt()
37 master_help = nfct_help(ct->master); in helper_mt()
A Dnf_nat_ftp.c74 struct nf_conn *ct = exp->master; in nf_nat_ftp()
91 nf_ct_helper_log(skb, exp->master, "all ports in use"); in nf_nat_ftp()
A Dxt_cluster.c113 if (ct->master) in xt_cluster_mt()
114 hash = xt_cluster_hash(ct->master, info); in xt_cluster_mt()
A Dnf_conntrack_netlink.c1965 if (ct->master) { in ctnetlink_change_helper()
2376 struct nf_conntrack_tuple master; in ctnetlink_create_conntrack() local
2392 ct->master = master_ct; in ctnetlink_create_conntrack()
2407 if (ct->master) in ctnetlink_create_conntrack()
2408 nf_ct_put(ct->master); in ctnetlink_create_conntrack()
3017 c = (unsigned long)exp->master; in nf_expect_get_id()
3031 struct nf_conn *master = exp->master; in ctnetlink_exp_dump_expect() local
3078 help = nfct_help(master); in ctnetlink_exp_dump_expect()
3178 id += nf_ct_get_id(exp->master); in ctnetlink_exp_id()
3394 m_help = nfct_help(exp->master); in expect_iter_name()
[all …]
A Dnf_nat_helper.c190 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in nf_nat_follow_master()
197 = ct->master->tuplehash[!exp->dir].tuple.src.u3; in nf_nat_follow_master()
A Dnf_conntrack_core.c492 b = (unsigned long)ct->master; in nf_ct_get_id()
567 struct nf_conn *master = ct->master; in destroy_gre_conntrack() local
569 if (master) in destroy_gre_conntrack()
570 nf_ct_gre_keymap_destroy(master); in destroy_gre_conntrack()
595 if (ct->master) in nf_ct_destroy()
596 nf_ct_put(ct->master); in nf_ct_destroy()
1820 ct->master = exp->master; in init_conntrack()
1828 ct->mark = READ_ONCE(exp->master->mark); in init_conntrack()
1831 ct->secmark = exp->master->secmark; in init_conntrack()
A Dnf_nat_irc.c40 struct nf_conn *ct = exp->master; in help()
A Dnf_nat_sip.c321 struct nf_conn_help *help = nfct_help(ct->master); in nf_nat_sip_expected()
361 &ct->master->tuplehash[exp->dir].tuple.src.u3)) { in nf_nat_sip_expected()
364 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in nf_nat_sip_expected()
/net/l3mdev/
A Dl3mdev.c120 struct net_device *master; in l3mdev_master_ifindex_rcu() local
129 master = netdev_master_upper_dev_get_rcu(_dev); in l3mdev_master_ifindex_rcu()
130 if (master) in l3mdev_master_ifindex_rcu()
131 ifindex = master->ifindex; in l3mdev_master_ifindex_rcu()
177 const struct net_device *master; in l3mdev_fib_table_rcu() local
179 master = netdev_master_upper_dev_get_rcu(_dev); in l3mdev_fib_table_rcu()
180 if (master && in l3mdev_fib_table_rcu()
181 master->l3mdev_ops->l3mdev_fib_table) in l3mdev_fib_table_rcu()
182 tb_id = master->l3mdev_ops->l3mdev_fib_table(master); in l3mdev_fib_table_rcu()
A DKconfig3 # Configuration for L3 master device support
11 drivers to support L3 master devices like VRF.
/net/ipv4/netfilter/
A Dnf_nat_pptp.c47 const struct nf_conn *master = ct->master; in pptp_nat_expected() local
60 ct_pptp_info = nfct_help_data(master); in pptp_nat_expected()
67 t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip; in pptp_nat_expected()
69 t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip; in pptp_nat_expected()
76 t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip; in pptp_nat_expected()
78 t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip; in pptp_nat_expected()
100 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in pptp_nat_expected()
110 = ct->master->tuplehash[!exp->dir].tuple.src.u3; in pptp_nat_expected()
202 const struct nf_conn *ct = expect_orig->master; in pptp_exp_gre()
/net/ax25/
A DKconfig57 networking. A DAMA server (called "master") accepts incoming traffic
72 networking. A DAMA server (called "master") accepts incoming traffic
74 If you say Y here, your Linux box will act as a DAMA master; this is
/net/mac80211/
A Diface.c1284 struct ieee80211_sub_if_data *master; in ieee80211_do_open() local
1291 master = container_of(sdata->bss, in ieee80211_do_open()
1294 master->control_port_protocol; in ieee80211_do_open()
1296 master->control_port_no_encrypt; in ieee80211_do_open()
1298 master->control_port_over_nl80211; in ieee80211_do_open()
1300 master->control_port_no_preauth; in ieee80211_do_open()
1301 sdata->vif.cab_queue = master->vif.cab_queue; in ieee80211_do_open()
1302 memcpy(sdata->vif.hw_queue, master->vif.hw_queue, in ieee80211_do_open()
1304 sdata->vif.bss_conf.chanreq = master->vif.bss_conf.chanreq; in ieee80211_do_open()
1307 master->crypto_tx_tailroom_needed_cnt; in ieee80211_do_open()
/net/openvswitch/
A Dconntrack.c174 if (ct->master && !nf_ct_is_confirmed(ct)) in ovs_ct_get_labels()
175 ct = ct->master; in ovs_ct_get_labels()
211 if (ct->master) in __ovs_ct_update_key()
212 ct = ct->master; in __ovs_ct_update_key()
261 if (ct->master) in ovs_ct_update_key()
389 master_cl = ct->master ? nf_ct_labels_find(ct->master) : NULL; in ovs_ct_init_labels()

Completed in 45 milliseconds

12