| /net/sched/ |
| A D | sch_teql.c | 136 if (!master) in teql_destroy() 139 prev = master->slaves; in teql_destroy() 145 if (q == master->slaves) { in teql_destroy() 152 master->slaves = NULL; in teql_destroy() 286 start = master->slaves; in teql_master_xmit() 319 master->tx_packets++; in teql_master_xmit() 347 master->tx_errors++; in teql_master_xmit() 350 master->tx_dropped++; in teql_master_xmit() 444 master->dev = dev; in teql_master_setup() 492 master = netdev_priv(dev); in teql_init() [all …]
|
| /net/hsr/ |
| A D | hsr_main.c | 34 struct hsr_port *port, *master; in hsr_netdev_notify() local 75 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_netdev_notify() 78 eth_hw_addr_set(master->dev, dev->dev_addr); in hsr_netdev_notify() 80 master->dev); in hsr_netdev_notify() 95 master->dev->dev_addr, in hsr_netdev_notify() 98 master->dev->dev_addr); in hsr_netdev_notify() 100 netdev_warn(master->dev, in hsr_netdev_notify() 108 WRITE_ONCE(master->dev->mtu, mtu_max); in hsr_netdev_notify() 114 if (hsr_slave_empty(master->hsr)) { in hsr_netdev_notify() 117 ops = master->dev->rtnl_link_ops; in hsr_netdev_notify() [all …]
|
| A D | hsr_device.c | 54 netif_carrier_on(master->dev); in hsr_check_carrier() 59 netif_carrier_off(master->dev); in hsr_check_carrier() 90 struct hsr_port *master; in hsr_check_carrier_and_operstate() local 227 struct hsr_port *master; in hsr_dev_xmit() local 230 if (master) { in hsr_dev_xmit() 231 skb->dev = master->dev; in hsr_dev_xmit() 235 hsr_forward_skb(skb, master); in hsr_dev_xmit() 270 skb->dev = master->dev; in hsr_init_skb() 399 hsr_forward_skb(skb, master); in send_prp_supervision_frame() 408 struct hsr_port *master; in hsr_announce() local [all …]
|
| A D | hsr_slave.c | 141 struct hsr_port *master; in hsr_portdev_setup() local 153 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_portdev_setup() 154 hsr_dev = master->dev; in hsr_portdev_setup() 179 struct hsr_port *port, *master; in hsr_add_port() local 209 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_add_port() 210 netdev_update_features(master->dev); in hsr_add_port() 211 dev_set_mtu(master->dev, hsr_get_max_mtu(hsr)); in hsr_add_port() 223 struct hsr_port *master; in hsr_del_port() local 226 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_del_port() 229 if (port != master) { in hsr_del_port() [all …]
|
| A D | hsr_netlink.c | 215 struct hsr_port *master; in hsr_nl_ringerror() local 245 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_nl_ringerror() 246 netdev_warn(master->dev, "Could not send HSR ring error message\n"); in hsr_nl_ringerror() 257 struct hsr_port *master; in hsr_nl_nodedown() local 282 master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); in hsr_nl_nodedown() 283 netdev_warn(master->dev, "Could not send HSR node down\n"); in hsr_nl_nodedown()
|
| /net/netfilter/ |
| A D | nf_conntrack_expect.c | 199 if (!nf_ct_is_confirmed(exp->master)) in nf_ct_find_expectation() 222 nf_ct_put(exp->master); in nf_ct_find_expectation() 263 net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && in expect_clash() 264 nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); in expect_clash() 272 net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && in expect_matches() 273 nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); in expect_matches() 283 return a->master == b->master; in master_matches() 306 new->master = me; in nf_ct_expect_alloc() 437 struct nf_conn *master = expect->master; in __nf_ct_expect_check() local 454 i->master != expect->master) in __nf_ct_expect_check() [all …]
|
| A D | xt_quota.c | 30 struct xt_quota_priv *priv = q->master; in quota_mt() 53 q->master = kmalloc(sizeof(*q->master), GFP_KERNEL); in quota_mt_check() 54 if (q->master == NULL) in quota_mt_check() 57 spin_lock_init(&q->master->lock); in quota_mt_check() 58 q->master->quota = q->quota; in quota_mt_check() 66 kfree(q->master); in quota_mt_destroy() 77 .usersize = offsetof(struct xt_quota_info, master),
|
| A D | xt_statistic.c | 42 oval = atomic_read(&info->master->count); in statistic_mt() 44 } while (atomic_cmpxchg(&info->master->count, oval, nval) != oval); in statistic_mt() 61 info->master = kzalloc(sizeof(*info->master), GFP_KERNEL); in statistic_mt_check() 62 if (info->master == NULL) in statistic_mt_check() 64 atomic_set(&info->master->count, info->u.nth.count); in statistic_mt_check() 73 kfree(info->master); in statistic_mt_destroy() 84 .usersize = offsetof(struct xt_statistic_info, master),
|
| A D | xt_limit.c | 66 struct xt_limit_priv *priv = r->master; in limit_mt() 123 r->master = priv; in limit_mt_check() 140 kfree(info->master); in limit_mt_destroy() 152 u_int32_t master; member 163 .prev = cm->prev | (unsigned long)cm->master << 32, in limit_mt_compat_from_user() 181 .master = m->prev >> 32, in limit_mt_compat_to_user()
|
| A D | nf_nat_amanda.c | 49 nf_ct_helper_log(skb, exp->master, "all ports in use"); in help() 54 if (!nf_nat_mangle_udp_packet(skb, exp->master, ctinfo, in help() 57 nf_ct_helper_log(skb, exp->master, "cannot mangle packet"); in help()
|
| A D | nf_nat_tftp.c | 27 const struct nf_conn *ct = exp->master; in help() 34 nf_ct_helper_log(skb, exp->master, "cannot add expectation"); in help()
|
| A D | xt_helper.c | 34 if (!ct || !ct->master) in helper_mt() 37 master_help = nfct_help(ct->master); in helper_mt()
|
| A D | nf_nat_ftp.c | 74 struct nf_conn *ct = exp->master; in nf_nat_ftp() 91 nf_ct_helper_log(skb, exp->master, "all ports in use"); in nf_nat_ftp()
|
| A D | xt_cluster.c | 113 if (ct->master) in xt_cluster_mt() 114 hash = xt_cluster_hash(ct->master, info); in xt_cluster_mt()
|
| A D | nf_conntrack_netlink.c | 1965 if (ct->master) { in ctnetlink_change_helper() 2376 struct nf_conntrack_tuple master; in ctnetlink_create_conntrack() local 2392 ct->master = master_ct; in ctnetlink_create_conntrack() 2407 if (ct->master) in ctnetlink_create_conntrack() 2408 nf_ct_put(ct->master); in ctnetlink_create_conntrack() 3017 c = (unsigned long)exp->master; in nf_expect_get_id() 3031 struct nf_conn *master = exp->master; in ctnetlink_exp_dump_expect() local 3078 help = nfct_help(master); in ctnetlink_exp_dump_expect() 3178 id += nf_ct_get_id(exp->master); in ctnetlink_exp_id() 3394 m_help = nfct_help(exp->master); in expect_iter_name() [all …]
|
| A D | nf_nat_helper.c | 190 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in nf_nat_follow_master() 197 = ct->master->tuplehash[!exp->dir].tuple.src.u3; in nf_nat_follow_master()
|
| A D | nf_conntrack_core.c | 492 b = (unsigned long)ct->master; in nf_ct_get_id() 567 struct nf_conn *master = ct->master; in destroy_gre_conntrack() local 569 if (master) in destroy_gre_conntrack() 570 nf_ct_gre_keymap_destroy(master); in destroy_gre_conntrack() 595 if (ct->master) in nf_ct_destroy() 596 nf_ct_put(ct->master); in nf_ct_destroy() 1820 ct->master = exp->master; in init_conntrack() 1828 ct->mark = READ_ONCE(exp->master->mark); in init_conntrack() 1831 ct->secmark = exp->master->secmark; in init_conntrack()
|
| A D | nf_nat_irc.c | 40 struct nf_conn *ct = exp->master; in help()
|
| A D | nf_nat_sip.c | 321 struct nf_conn_help *help = nfct_help(ct->master); in nf_nat_sip_expected() 361 &ct->master->tuplehash[exp->dir].tuple.src.u3)) { in nf_nat_sip_expected() 364 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in nf_nat_sip_expected()
|
| /net/l3mdev/ |
| A D | l3mdev.c | 120 struct net_device *master; in l3mdev_master_ifindex_rcu() local 129 master = netdev_master_upper_dev_get_rcu(_dev); in l3mdev_master_ifindex_rcu() 130 if (master) in l3mdev_master_ifindex_rcu() 131 ifindex = master->ifindex; in l3mdev_master_ifindex_rcu() 177 const struct net_device *master; in l3mdev_fib_table_rcu() local 179 master = netdev_master_upper_dev_get_rcu(_dev); in l3mdev_fib_table_rcu() 180 if (master && in l3mdev_fib_table_rcu() 181 master->l3mdev_ops->l3mdev_fib_table) in l3mdev_fib_table_rcu() 182 tb_id = master->l3mdev_ops->l3mdev_fib_table(master); in l3mdev_fib_table_rcu()
|
| A D | Kconfig | 3 # Configuration for L3 master device support 11 drivers to support L3 master devices like VRF.
|
| /net/ipv4/netfilter/ |
| A D | nf_nat_pptp.c | 47 const struct nf_conn *master = ct->master; in pptp_nat_expected() local 60 ct_pptp_info = nfct_help_data(master); in pptp_nat_expected() 67 t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip; in pptp_nat_expected() 69 t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip; in pptp_nat_expected() 76 t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip; in pptp_nat_expected() 78 t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip; in pptp_nat_expected() 100 = ct->master->tuplehash[!exp->dir].tuple.dst.u3; in pptp_nat_expected() 110 = ct->master->tuplehash[!exp->dir].tuple.src.u3; in pptp_nat_expected() 202 const struct nf_conn *ct = expect_orig->master; in pptp_exp_gre()
|
| /net/ax25/ |
| A D | Kconfig | 57 networking. A DAMA server (called "master") accepts incoming traffic 72 networking. A DAMA server (called "master") accepts incoming traffic 74 If you say Y here, your Linux box will act as a DAMA master; this is
|
| /net/mac80211/ |
| A D | iface.c | 1284 struct ieee80211_sub_if_data *master; in ieee80211_do_open() local 1291 master = container_of(sdata->bss, in ieee80211_do_open() 1294 master->control_port_protocol; in ieee80211_do_open() 1296 master->control_port_no_encrypt; in ieee80211_do_open() 1298 master->control_port_over_nl80211; in ieee80211_do_open() 1300 master->control_port_no_preauth; in ieee80211_do_open() 1301 sdata->vif.cab_queue = master->vif.cab_queue; in ieee80211_do_open() 1302 memcpy(sdata->vif.hw_queue, master->vif.hw_queue, in ieee80211_do_open() 1304 sdata->vif.bss_conf.chanreq = master->vif.bss_conf.chanreq; in ieee80211_do_open() 1307 master->crypto_tx_tailroom_needed_cnt; in ieee80211_do_open()
|
| /net/openvswitch/ |
| A D | conntrack.c | 174 if (ct->master && !nf_ct_is_confirmed(ct)) in ovs_ct_get_labels() 175 ct = ct->master; in ovs_ct_get_labels() 211 if (ct->master) in __ovs_ct_update_key() 212 ct = ct->master; in __ovs_ct_update_key() 261 if (ct->master) in ovs_ct_update_key() 389 master_cl = ct->master ? nf_ct_labels_find(ct->master) : NULL; in ovs_ct_init_labels()
|