| /net/sunrpc/ |
| A D | xdr.c | 637 base += len; in xdr_buf_head_copy_right() 643 base -= pglen; in xdr_buf_head_copy_right() 647 base -= len; in xdr_buf_head_copy_right() 719 base += hdlen; in xdr_buf_tail_copy_left() 733 base += pglen; in xdr_buf_tail_copy_left() 794 len += base; in xdr_buf_pages_shift_left() 1300 base = len; in xdr_set_iov() 1671 base = 0; in xdr_buf_subsegment() 1684 base = 0; in xdr_buf_subsegment() 1697 base = 0; in xdr_buf_subsegment() [all …]
|
| A D | socklib.c | 190 size_t base) in xprt_send_rm_and_kvec() argument 202 return xprt_sendmsg(sock, msg, base); in xprt_send_rm_and_kvec() 234 if (base < want) { in xprt_sock_sendmsg() 235 unsigned int len = want - base; in xprt_sock_sendmsg() 242 &xdr->head[0], base); in xprt_sock_sendmsg() 248 base = 0; in xprt_sock_sendmsg() 250 base -= want; in xprt_sock_sendmsg() 253 if (base < xdr->page_len) { in xprt_sock_sendmsg() 263 base = 0; in xprt_sock_sendmsg() 265 base -= xdr->page_len; in xprt_sock_sendmsg() [all …]
|
| /net/sctp/ |
| A D | endpointola.c | 69 ep->base.dead = false; in sctp_endpoint_init() 112 ep->base.sk = sk; in sctp_endpoint_init() 113 ep->base.net = sock_net(sk); in sctp_endpoint_init() 114 sock_hold(ep->base.sk); in sctp_endpoint_init() 176 ep->base.dead = true; in sctp_endpoint_free() 223 sk = ep->base.sk; in sctp_endpoint_destroy() 259 sctp_sk(ep->base.sk))) in sctp_endpoint_is_match() 309 bp = &ep->base.bind_addr; in sctp_endpoint_is_peeled_off() 341 if (ep->base.dead) in sctp_endpoint_bh_rcv() 345 inqueue = &ep->base.inqueue; in sctp_endpoint_bh_rcv() [all …]
|
| A D | associola.c | 67 asoc->base.net = sock_net(sk); in sctp_association_init() 70 sock_hold(asoc->base.sk); in sctp_association_init() 79 sctp_bind_addr_init(&asoc->base.bind_addr, ep->base.bind_addr.port); in sctp_association_init() 280 sock_put(asoc->base.sk); in sctp_association_init() 337 asoc->base.dead = true; in sctp_association_free() 408 sock_put(asoc->base.sk); in sctp_association_destroy() 592 sp = sctp_sk(asoc->base.sk); in sctp_assoc_add_peer() 1040 if (asoc->base.dead) in sctp_assoc_bh_rcv() 1070 sock_put(assoc->base.sk); in sctp_assoc_migrate() 1077 assoc->base.sk = newsk; in sctp_assoc_migrate() [all …]
|
| A D | input.c | 171 rcvr = asoc ? &asoc->base : &ep->base; in sctp_rcv() 494 sk = asoc->base.sk; in sctp_err_lookup() 552 struct sock *sk = asoc->base.sk; in sctp_v4_err_handle() 735 struct sock *sk = ep->base.sk; in __sctp_hash_endpoint() 756 struct sock *sk2 = ep2->base.sk; in __sctp_hash_endpoint() 803 struct sock *sk = ep->base.sk; in __sctp_unhash_endpoint() 864 sk = ep->base.sk; in __sctp_rcv_lookup_endpoint() 897 if (!net_eq(t->asoc->base.net, x->net)) in sctp_hash_cmp() 912 return sctp_hashfn(t->asoc->base.net, in sctp_hash_obj() 952 arg.net = t->asoc->base.net; in sctp_hash_transport() [all …]
|
| A D | ulpqueue.c | 165 struct sctp_sock *sp = sctp_sk(ulpq->asoc->base.sk); in sctp_ulpq_set_pd() 176 return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); in sctp_ulpq_clear_pd() 181 struct sock *sk = ulpq->asoc->base.sk; in sctp_ulpq_tail_event() 476 if (!sctp_sk(asoc->base.sk)->frag_interleave && in sctp_ulpq_retrieve_reassembled() 477 atomic_read(&sctp_sk(asoc->base.sk)->pd_mode)) in sctp_ulpq_retrieve_reassembled() 481 pd_point = sctp_sk(asoc->base.sk)->pd_point; in sctp_ulpq_retrieve_reassembled() 483 retval = sctp_make_reassembled_event(asoc->base.net, in sctp_ulpq_retrieve_reassembled() 493 retval = sctp_make_reassembled_event(ulpq->asoc->base.net, in sctp_ulpq_retrieve_reassembled() 1029 sp = sctp_sk(asoc->base.sk); in sctp_ulpq_partial_delivery() 1079 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege() [all …]
|
| A D | diag.c | 34 laddr = list_entry(asoc->base.bind_addr.address_list.next, in inet_diag_msg_sctpasoc_fill() 40 r->id.idiag_sport = htons(asoc->base.bind_addr.port); in inet_diag_msg_sctpasoc_fill() 197 addr_list = asoc ? &asoc->base.bind_addr.address_list in inet_sctp_diag_fill() 198 : &ep->base.bind_addr.address_list; in inet_sctp_diag_fill() 232 list_for_each_entry_rcu(laddr, &asoc->base.bind_addr.address_list, in inet_assoc_attr_size() 249 struct sock *sk = ep->base.sk; in sctp_sock_dump_one() 289 struct sock *sk = ep->base.sk; in sctp_sock_dump() 303 if (r->id.idiag_sport != htons(assoc->base.bind_addr.port) && in sctp_sock_dump() 344 struct sock *sk = ep->base.sk; in sctp_sock_filter() 360 struct sock *sk = ep->base.sk; in sctp_ep_dump()
|
| /net/netfilter/ |
| A D | nf_conntrack_h323_asn1.c | 526 base = (base && (f->attr & DECODE)) ? base + f->offset : NULL; in decode_seq() 539 if (base) in decode_seq() 598 if (base) in decode_seq() 659 base = (base && (f->attr & DECODE)) ? base + f->offset : NULL; in decode_seqof() 692 if (base) { in decode_seqof() 700 if (base) in decode_seqof() 721 base : NULL, in decode_seqof() 732 base : NULL, in decode_seqof() 737 if (base) in decode_seqof() 755 base = (base && (f->attr & DECODE)) ? base + f->offset : NULL; in decode_choice() [all …]
|
| A D | nft_payload.c | 171 switch (priv->base) { in nft_payload_eval() 265 if (priv->base != payload->base || in nft_payload_reduce() 558 switch (priv->base) { in nft_payload_offload() 607 switch (priv->base) { in nft_payload_inner_eval() 651 u32 base; in nft_payload_inner_init() local 658 switch (base) { in nft_payload_inner_init() 668 priv->base = base; in nft_payload_inner_init() 864 switch (priv->base) { in nft_payload_set_eval() 1043 enum nft_payload_bases base; in nft_payload_select_ops() local 1053 switch (base) { in nft_payload_select_ops() [all …]
|
| A D | nft_set_pipapo.c | 998 base[len - 1 - step / BITS_PER_BYTE]); in pipapo_step_diff() 1019 memcpy(tmp, base, len); in pipapo_step_after_end() 1050 base[i]++; 1052 base[i] += 1 << (step % BITS_PER_BYTE); 1054 if (base[i]) 1077 u8 base[NFT_PIPAPO_MAX_BYTES]; local 1079 memcpy(base, start, bytes); 1080 while (memcmp(base, end, bytes) <= 0) { 1091 err = pipapo_insert(f, base, 0); 1100 err = pipapo_insert(f, base, len - step); [all …]
|
| /net/ipv4/ |
| A D | inetpeer.c | 89 struct inet_peer_base *base, in lookup() argument 100 pp = &base->rb_root.rb_node; in lookup() 147 if (base->total >= peer_threshold) in inet_peer_gc() 151 base->total / peer_threshold * HZ; in inet_peer_gc() 164 base->total--; in inet_peer_gc() 181 seq = read_seqbegin(&base->lock); in inet_getpeer() 191 write_seqlock_bh(&base->lock); in inet_getpeer() 212 base->total++; in inet_getpeer() 216 inet_peer_gc(base, gc_stack, gc_cnt); in inet_getpeer() 217 write_sequnlock_bh(&base->lock); in inet_getpeer() [all …]
|
| /net/rxrpc/ |
| A D | peer_event.c | 245 time64_t base, in rxrpc_peer_keepalive_dispatch() argument 269 slot = keepalive_at - base; in rxrpc_peer_keepalive_dispatch() 273 if (keepalive_at <= base || in rxrpc_peer_keepalive_dispatch() 306 time64_t base, now, delay; in rxrpc_peer_keepalive_worker() local 311 base = rxnet->peer_keepalive_base; in rxrpc_peer_keepalive_worker() 313 _enter("%lld,%u", base - now, cursor); in rxrpc_peer_keepalive_worker() 332 base++; in rxrpc_peer_keepalive_worker() 336 base = now; in rxrpc_peer_keepalive_worker() 339 rxnet->peer_keepalive_base = base; in rxrpc_peer_keepalive_worker() 350 base++; in rxrpc_peer_keepalive_worker() [all …]
|
| /net/ethtool/ |
| A D | linkmodes.c | 10 struct ethnl_req_info base; member 14 struct ethnl_reply_data base; member 21 container_of(__reply_base, struct linkmodes_reply_data, base) 36 data->lsettings = &data->ksettings.base; in linkmodes_prepare_data() 65 const struct ethtool_link_settings *lsettings = &ksettings->base; in linkmodes_reply_size() 106 const struct ethtool_link_settings *lsettings = &ksettings->base; in linkmodes_fill_reply() 189 (!req_speed || info->speed == ksettings->base.speed) && in ethnl_auto_linkmodes() 191 (!req_duplex || info->duplex == ksettings->base.duplex)) in ethnl_auto_linkmodes() 240 struct ethtool_link_settings *lsettings = &ksettings->base; in ethnl_update_linkmodes()
|
| A D | tsconfig.c | 13 struct ethnl_req_info base; member 17 struct ethnl_reply_data base; member 27 container_of(__reply_base, struct tsconfig_reply_data, base) 215 reply_data->base.dev = dev; in tsconfig_send_reply() 216 ret = tsconfig_prepare_data(&req_info->base, &reply_data->base, info); in tsconfig_send_reply() 220 ret = tsconfig_reply_size(&req_info->base, &reply_data->base); in tsconfig_send_reply() 230 ret = tsconfig_fill_reply(rskb, &req_info->base, &reply_data->base); in tsconfig_send_reply()
|
| A D | tsinfo.c | 15 struct ethnl_req_info base; member 20 struct ethnl_reply_data base; member 26 container_of(__req_base, struct tsinfo_req_info, base) 29 container_of(__reply_base, struct tsinfo_reply_data, base) 309 reply_data->base.dev = dev; in ethnl_tsinfo_prepare_dump() 331 ret = tsinfo_fill_reply(skb, &req_info->base, &reply_data->base); in ethnl_tsinfo_end_dump() 335 reply_data->base.dev = NULL; in ethnl_tsinfo_end_dump() 475 if (ctx->req_info->base.dev) { in ethnl_tsinfo_dumpit() 476 dev = ctx->req_info->base.dev; in ethnl_tsinfo_dumpit() 517 ret = ethnl_parse_header_dev_get(&req_info->base, in ethnl_tsinfo_start() [all …]
|
| A D | linkinfo.c | 7 struct ethnl_req_info base; member 11 struct ethnl_reply_data base; member 17 container_of(__reply_base, struct linkinfo_reply_data, base) 32 data->lsettings = &data->ksettings.base; in linkinfo_prepare_data() 112 lsettings = &ksettings.base; in ethnl_set_linkinfo()
|
| A D | rss.c | 9 struct ethnl_req_info base; member 14 struct ethnl_reply_data base; member 421 ret = rss_fill_reply(skb, &req.base, &data.base); in rss_dump_one_ctx() 426 rss_cleanup_data(&data.base); in rss_dump_one_ctx() 430 rss_cleanup_data(&data.base); in rss_dump_one_ctx() 531 ethnl_notify(dev, type, &req_info.base); in ethtool_rss_notify() 839 data.base.dev = dev; in ethnl_rss_set() 910 rss_cleanup_data(&data.base); in ethnl_rss_set() 1022 dev = req.base.dev; in ethnl_rss_create_doit() 1094 ntf_fail |= rss_fill_reply(rsp, &req.base, &data.base); in ethnl_rss_create_doit() [all …]
|
| A D | eeprom.c | 9 struct ethnl_req_info base; member 18 struct ethnl_reply_data base; member 24 container_of(__req_base, struct eeprom_req_info, base) 27 container_of(__reply_base, struct eeprom_reply_data, base) 56 struct net_device *dev = reply->base.dev; in eeprom_fallback()
|
| A D | phc_vclocks.c | 9 struct ethnl_req_info base; member 13 struct ethnl_reply_data base; member 19 container_of(__reply_base, struct phc_vclocks_reply_data, base)
|
| /net/bridge/ |
| A D | br_device.c | 259 cmd->base.duplex = DUPLEX_UNKNOWN; in br_get_link_ksettings() 260 cmd->base.port = PORT_OTHER; in br_get_link_ksettings() 261 cmd->base.speed = SPEED_UNKNOWN; in br_get_link_ksettings() 273 if (ecmd.base.speed == (__u32)SPEED_UNKNOWN) in br_get_link_ksettings() 276 if (cmd->base.speed == (__u32)SPEED_UNKNOWN || in br_get_link_ksettings() 277 cmd->base.speed < ecmd.base.speed) in br_get_link_ksettings() 278 cmd->base.speed = ecmd.base.speed; in br_get_link_ksettings()
|
| /net/sched/ |
| A D | act_gate.c | 38 ktime_t now, base, cycle; in gate_get_start_time() local 41 base = ns_to_ktime(param->tcfg_basetime); in gate_get_start_time() 44 if (ktime_after(base, now)) { in gate_get_start_time() 45 *start = base; in gate_get_start_time() 51 n = div64_u64(ktime_sub_ns(now, base), cycle); in gate_get_start_time() 52 *start = ktime_add_ns(base, (n + 1) * cycle); in gate_get_start_time() 99 ktime_t cycle, base; in gate_timer_func() local 103 base = ns_to_ktime(p->tcfg_basetime); in gate_timer_func() 104 n = div64_u64(ktime_sub_ns(now, base), cycle); in gate_timer_func() 105 close_time = ktime_add_ns(base, (n + 1) * cycle); in gate_timer_func()
|
| A D | cls_basic.c | 148 struct basic_filter *f, unsigned long base, in basic_set_parms() argument 165 tcf_bind_filter(tp, &f->res, base); in basic_set_parms() 173 struct tcf_proto *tp, unsigned long base, u32 handle, in basic_change() argument 221 err = basic_set_parms(net, tp, fnew, base, tb, tca[TCA_RATE], flags, in basic_change() 262 unsigned long base) in basic_bind_class() argument 266 tc_cls_bind_class(classid, cl, q, &f->res, base); in basic_bind_class()
|
| A D | cls_fw.c | 203 struct nlattr **tca, unsigned long base, u32 flags, in fw_set_parms() argument 233 tcf_bind_filter(tp, &f->res, base); in fw_set_parms() 240 struct tcf_proto *tp, unsigned long base, in fw_change() argument 280 err = fw_set_parms(net, tp, fnew, tb, tca, base, flags, extack); in fw_change() 329 err = fw_set_parms(net, tp, f, tb, tca, base, flags, extack); in fw_change() 416 unsigned long base) in fw_bind_class() argument 420 tc_cls_bind_class(classid, cl, q, &f->res, base); in fw_bind_class()
|
| /net/ipv6/ |
| A D | ip6_output.c | 1355 cork->base.dst = &rt->dst; in ip6_setup_cork() 1404 cork->base.fragsize = mtu; in ip6_setup_cork() 1406 cork->base.tx_flags = 0; in ip6_setup_cork() 1414 cork->base.length = 0; in ip6_setup_cork() 1870 cork->base.dst = NULL; in ip6_cork_steal_dst() 1888 if (cork->base.dst) { in ip6_cork_release() 1890 cork->base.dst = NULL; in ip6_cork_release() 1951 skb->mark = cork->base.mark; in __ip6_make_skb() 2052 cork->base.flags = 0; in ip6_make_skb() 2053 cork->base.addr = 0; in ip6_make_skb() [all …]
|
| /net/unix/ |
| A D | unix_bpf.c | 106 static void unix_dgram_bpf_rebuild_protos(struct proto *prot, const struct proto *base) in unix_dgram_bpf_rebuild_protos() argument 108 *prot = *base; in unix_dgram_bpf_rebuild_protos() 115 const struct proto *base) in unix_stream_bpf_rebuild_protos() argument 117 *prot = *base; in unix_stream_bpf_rebuild_protos()
|