| /net/netfilter/ |
| A D | nft_payload.c | 66 offset = ETH_HLEN; in nft_payload_copy_vlan() 68 offset -= VLAN_HLEN; in nft_payload_copy_vlan() 166 int offset; in nft_payload_eval() local 195 if (offset < 0) in nft_payload_eval() 202 offset += priv->offset; in nft_payload_eval() 266 priv->offset != payload->offset || in nft_payload_reduce() 602 int offset; in nft_payload_inner_eval() local 636 offset += priv->offset; in nft_payload_inner_eval() 800 u8 offset; member 892 if (offset < 0) in nft_payload_set_eval() [all …]
|
| A D | nft_exthdr.c | 21 u8 offset; member 32 if (opt[offset] <= TCPOPT_NOP || opt[offset + 1] == 0) in optlen() 65 offset += priv->offset; in nft_exthdr_ipv6_eval() 159 offset += priv->offset; in nft_exthdr_ipv4_eval() 213 offset = i + priv->offset; in nft_exthdr_tcp_eval() 266 offset = i + priv->offset; in nft_exthdr_tcp_set_eval() 396 if (nft_skb_copy_to_reg(pkt->skb, offset + priv->offset, in nft_exthdr_sctp_eval() 536 priv->offset = offset; in nft_exthdr_init() 571 if (offset < 2) in nft_exthdr_tcp_set_init() 586 priv->offset = offset; in nft_exthdr_tcp_set_init() [all …]
|
| A D | nft_set_pipapo_avx2.c | 101 int offset = start % BITS_PER_LONG; in nft_pipapo_avx2_fill() local 107 *data |= BIT(offset); in nft_pipapo_avx2_fill() 113 *data |= GENMASK(len - 1 + offset, offset); in nft_pipapo_avx2_fill() 117 *data |= ~0UL << offset; in nft_pipapo_avx2_fill() 118 len -= BITS_PER_LONG - offset; in nft_pipapo_avx2_fill() 216 int offset, const u8 *pkt, in nft_pipapo_avx2_lookup_4b_2() argument 279 int offset, const u8 *pkt, in nft_pipapo_avx2_lookup_4b_4() argument 452 int offset, const u8 *pkt, in nft_pipapo_avx2_lookup_4b_12() argument 542 int offset, const u8 *pkt, in nft_pipapo_avx2_lookup_4b_32() argument 963 int offset, const u8 *pkt, in nft_pipapo_avx2_lookup_8b_16() argument [all …]
|
| A D | nft_hash.c | 23 u32 offset; member 37 regs->data[priv->dreg] = h + priv->offset; in nft_jhash_eval() 43 u32 offset; member 57 regs->data[priv->dreg] = h + priv->offset; in nft_symhash_eval() 85 priv->offset = ntohl(nla_get_be32(tb[NFTA_HASH_OFFSET])); in nft_jhash_init() 103 if (priv->offset + priv->modulus - 1 < priv->offset) in nft_jhash_init() 128 priv->offset = ntohl(nla_get_be32(tb[NFTA_HASH_OFFSET])); in nft_symhash_init() 134 if (priv->offset + priv->modulus - 1 < priv->offset) in nft_symhash_init() 158 if (priv->offset != 0) in nft_jhash_dump() 188 if (priv->offset != 0) in nft_symhash_dump() [all …]
|
| A D | nft_numgen.c | 21 u32 offset; member 33 return nval + priv->offset; in nft_ng_inc_gen() 60 priv->offset = ntohl(nla_get_be32(tb[NFTA_NG_OFFSET])); in nft_ng_inc_init() 66 if (priv->offset + priv->modulus - 1 < priv->offset) in nft_ng_inc_init() 98 u32 modulus, enum nft_ng_types type, u32 offset) in nft_ng_dump() argument 106 if (nla_put_be32(skb, NFTA_NG_OFFSET, htonl(offset))) in nft_ng_dump() 121 priv->offset); in nft_ng_inc_dump() 135 u32 offset; member 159 priv->offset = ntohl(nla_get_be32(tb[NFTA_NG_OFFSET])); in nft_ng_random_init() 165 if (priv->offset + priv->modulus - 1 < priv->offset) in nft_ng_random_init() [all …]
|
| /net/tls/ |
| A D | tls_strp.c | 43 int offset, int len) in tls_strp_skb_copy() argument 80 rxm->offset = 0; in tls_strp_msg_make_copy() 165 offset = strp->stm.offset; in tls_strp_msg_hold() 171 offset -= iter->len; in tls_strp_msg_hold() 176 offset = 0; in tls_strp_msg_hold() 252 offset += chunk; in tls_strp_copyin_frag() 273 offset += chunk; in tls_strp_copyin_frag() 417 strp->stm.offset = 0; in tls_strp_read_copy() 459 u32 offset; in tls_strp_load_anchor_with_queue() local 475 strp->stm.offset = offset; in tls_strp_load_anchor_with_queue() [all …]
|
| /net/rds/ |
| A D | info.c | 67 unsigned long offset; member 75 int offset = optname - RDS_INFO_FIRST; in rds_info_register_func() local 80 BUG_ON(rds_info_funcs[offset]); in rds_info_register_func() 81 rds_info_funcs[offset] = func; in rds_info_register_func() 88 int offset = optname - RDS_INFO_FIRST; in rds_info_deregister_func() local 93 BUG_ON(rds_info_funcs[offset] != func); in rds_info_deregister_func() 94 rds_info_funcs[offset] = NULL; in rds_info_deregister_func() 128 iter->offset, this, data, bytes); in rds_info_copy() 134 iter->offset += this; in rds_info_copy() 136 if (iter->offset == PAGE_SIZE) { in rds_info_copy() [all …]
|
| /net/sunrpc/auth_gss/ |
| A D | gss_krb5_crypto.c | 261 sg->offset); in encryptor() 263 sg->offset); in encryptor() 322 sg->offset); in decryptor() 401 u32 len = buf->len - offset; in gss_krb5_cts_crypt() 486 nbytes = buf->len - offset; in krb5_cbc_cts_encrypt() 498 desc.pos = offset; in krb5_cbc_cts_encrypt() 543 u32 offset, struct xdr_buf *buf) in krb5_cbc_cts_decrypt() argument 605 offset += GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt() 609 offset -= GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt() 834 offset += GSS_KRB5_TOK_HDR_LEN; in krb5_etm_encrypt() [all …]
|
| A D | gss_krb5_internal.h | 41 u32 (*encrypt)(struct krb5_ctx *kctx, u32 offset, 43 u32 (*decrypt)(struct krb5_ctx *kctx, u32 offset, u32 len, 49 u32 (*wrap)(struct krb5_ctx *kctx, int offset, 51 u32 (*unwrap)(struct krb5_ctx *kctx, int offset, int len, 93 u32 gss_krb5_wrap_v2(struct krb5_ctx *kctx, int offset, 96 u32 gss_krb5_unwrap_v2(struct krb5_ctx *kctx, int offset, int len, 168 u32 gss_krb5_aes_encrypt(struct krb5_ctx *kctx, u32 offset, 171 u32 gss_krb5_aes_decrypt(struct krb5_ctx *kctx, u32 offset, u32 len, 177 u32 krb5_etm_decrypt(struct krb5_ctx *kctx, u32 offset, u32 len, 184 struct crypto_sync_skcipher *cbc_tfm, u32 offset, [all …]
|
| A D | gss_krb5_wrap.c | 94 gss_krb5_wrap_v2(struct krb5_ctx *kctx, int offset, in gss_krb5_wrap_v2() argument 107 if (xdr_extend_head(buf, offset, GSS_KRB5_TOK_HDR_LEN)) in gss_krb5_wrap_v2() 111 ptr = buf->head[0].iov_base + offset; in gss_krb5_wrap_v2() 133 err = (*kctx->gk5e->encrypt)(kctx, offset, buf, pages); in gss_krb5_wrap_v2() 142 gss_krb5_unwrap_v2(struct krb5_ctx *kctx, int offset, int len, in gss_krb5_unwrap_v2() argument 158 ptr = buf->head[0].iov_base + offset; in gss_krb5_unwrap_v2() 185 rotate_left(offset + 16, buf, rrc); in gss_krb5_unwrap_v2() 187 err = (*kctx->gk5e->decrypt)(kctx, offset, len, buf, in gss_krb5_unwrap_v2() 224 movelen -= offset + GSS_KRB5_TOK_HDR_LEN + headskip; in gss_krb5_unwrap_v2() 225 BUG_ON(offset + GSS_KRB5_TOK_HDR_LEN + headskip + movelen > in gss_krb5_unwrap_v2()
|
| /net/ipv6/ |
| A D | exthdrs_core.c | 123 if (offset + 2 > packet_len) in ipv6_find_tlv() 125 hdr = (struct ipv6_opt_hdr *)(nh + offset); in ipv6_find_tlv() 128 if (offset + len > packet_len) in ipv6_find_tlv() 131 offset += 2; in ipv6_find_tlv() 135 int opttype = nh[offset]; in ipv6_find_tlv() 139 return offset; in ipv6_find_tlv() 148 optlen = nh[offset + 1] + 2; in ipv6_find_tlv() 153 offset += optlen; in ipv6_find_tlv() 198 if (*offset) { in ipv6_find_hdr() 204 start = *offset + sizeof(struct ipv6hdr); in ipv6_find_hdr() [all …]
|
| A D | mcast_snoop.c | 18 unsigned int offset = skb_network_offset(skb) + sizeof(*ip6h); in ipv6_mc_check_ip6hdr() local 20 if (!pskb_may_pull(skb, offset)) in ipv6_mc_check_ip6hdr() 28 len = offset + ntohs(ip6h->payload_len); in ipv6_mc_check_ip6hdr() 29 if (skb->len < len || len <= offset) in ipv6_mc_check_ip6hdr() 32 skb_set_transport_header(skb, offset); in ipv6_mc_check_ip6hdr() 40 int offset; in ipv6_mc_check_exthdrs() local 50 offset = skb_network_offset(skb) + sizeof(*ip6h); in ipv6_mc_check_exthdrs() 51 offset = ipv6_skip_exthdr(skb, offset, &nexthdr, &frag_off); in ipv6_mc_check_exthdrs() 53 if (offset < 0) in ipv6_mc_check_exthdrs() 59 skb_set_transport_header(skb, offset); in ipv6_mc_check_exthdrs()
|
| A D | calipso.c | 810 if (offset < sizeof(*opt) || offset >= opt_len) in calipso_tlv_len() 851 offset = sizeof(*hop); in calipso_opt_find() 858 switch (opt[offset]) { in calipso_opt_find() 862 offset_e = offset; in calipso_opt_find() 866 offset_e = offset; in calipso_opt_find() 870 offset_s = offset; in calipso_opt_find() 874 offset += tlv_len; in calipso_opt_find() 1088 offset = sizeof(*hop); in calipso_sock_getattr() 1104 offset += len; in calipso_sock_getattr() 1291 int offset; in calipso_skbuff_optptr() local [all …]
|
| A D | output_core.c | 60 unsigned int offset = sizeof(struct ipv6hdr); in ip6_find_1stfragopt() local 66 while (offset <= packet_len) { in ip6_find_1stfragopt() 78 if (ipv6_find_tlv(skb, offset, IPV6_TLV_HAO) >= 0) in ip6_find_1stfragopt() 82 return offset; in ip6_find_1stfragopt() 85 return offset; in ip6_find_1stfragopt() 88 if (offset + sizeof(struct ipv6_opt_hdr) > packet_len) in ip6_find_1stfragopt() 92 offset); in ip6_find_1stfragopt() 93 offset += ipv6_optlen(exthdr); in ip6_find_1stfragopt() 94 if (offset > IPV6_MAXPLEN) in ip6_find_1stfragopt()
|
| /net/ethtool/ |
| A D | eeprom.c | 10 u32 offset; member 33 u32 offset = request->offset; in fallback_set_params() local 37 offset = request->page * ETH_MODULE_EEPROM_PAGE_LEN + offset; in fallback_set_params() 41 offset += ETH_MODULE_EEPROM_PAGE_LEN * 2; in fallback_set_params() 43 if (offset >= modinfo->eeprom_len) in fallback_set_params() 48 eeprom->offset = offset; in fallback_set_params() 119 page_data.offset = request->offset; in eeprom_prepare_data() 164 request->offset = nla_get_u32(tb[ETHTOOL_A_MODULE_EEPROM_OFFSET]); in eeprom_parse_request() 174 if (request->page && request->offset < ETH_MODULE_EEPROM_PAGE_LEN) { in eeprom_parse_request() 180 if (request->offset < ETH_MODULE_EEPROM_PAGE_LEN && in eeprom_parse_request() [all …]
|
| A D | cmis_fw_update.c | 164 u32 offset, max_block_size, max_lpl_len; in cmis_fw_update_write_image_lpl() local 175 for (offset = start; offset < image_size; offset += max_block_size) { in cmis_fw_update_write_image_lpl() 177 .block_address = cpu_to_be32(offset - start), in cmis_fw_update_write_image_lpl() 184 offset - start, in cmis_fw_update_write_image_lpl() 186 block_size = min_t(u32, max_block_size, image_size - offset); in cmis_fw_update_write_image_lpl() 223 u32 offset, lpl_len; in cmis_fw_update_write_image_epl() local 229 for (offset = start; offset < image_size; in cmis_fw_update_write_image_epl() 230 offset += ETHTOOL_CMIS_CDB_EPL_MAX_PL_LENGTH) { in cmis_fw_update_write_image_epl() 232 .block_address = cpu_to_be32(offset - start), in cmis_fw_update_write_image_epl() 240 offset - start, in cmis_fw_update_write_image_epl() [all …]
|
| A D | cmis_cdb.c | 44 u8 page, u32 offset, u32 length) in ethtool_cmis_page_init() argument 47 page_data->offset = offset; in ethtool_cmis_page_init() 374 u16 max_duration, u32 offset, in ethtool_cmis_wait_for_cond() argument 571 u16 offset = CMIS_CDB_EPL_FW_BLOCK_OFFSET_START; in ethtool_cmis_cdb_execute_epl_cmd() local 584 page, offset, in ethtool_cmis_cdb_execute_epl_cmd() 590 offset += bytes_to_write; in ethtool_cmis_cdb_execute_epl_cmd() 614 u32 offset; in ethtool_cmis_cdb_execute_cmd() local 633 offset = CMIS_CDB_CMD_ID_OFFSET + in ethtool_cmis_cdb_execute_cmd() 636 ETHTOOL_CMIS_CDB_CMD_PAGE, offset, in ethtool_cmis_cdb_execute_cmd() 648 offset = CMIS_CDB_CMD_ID_OFFSET + in ethtool_cmis_cdb_execute_cmd() [all …]
|
| /net/batman-adv/ |
| A D | multicast_forw.c | 283 size_t offset; in batadv_mcast_forw_push_want_all() local 302 offset, num_dests, tvlv_len); in batadv_mcast_forw_push_want_all() 329 size_t offset; in batadv_mcast_forw_push_want_rtr() local 368 unsigned short offset, in batadv_mcast_forw_scrape() argument 378 memmove(to, from, offset); in batadv_mcast_forw_scrape() 420 memmove(to, from, offset); in batadv_mcast_forw_push_insert_padding() 421 memset(to + offset, 0, *tvlv_len - offset); in batadv_mcast_forw_push_insert_padding() 832 unsigned int offset; in batadv_mcast_forw_shrink_update_headers() local 857 return offset; in batadv_mcast_forw_shrink_update_headers() 895 unsigned int offset; in batadv_mcast_forw_shrink_tracker() local [all …]
|
| /net/core/ |
| A D | datagram.c | 395 int i, copy = start - offset, start_off = offset, n; in __skb_datagram_iter() 404 offset += n; in __skb_datagram_iter() 419 WARN_ON(start > offset + len); in __skb_datagram_iter() 440 offset += n; in __skb_datagram_iter() 452 WARN_ON(start > offset + len); in __skb_datagram_iter() 463 offset += copy; in __skb_datagram_iter() 554 int i, copy = start - offset; in skb_copy_datagram_from_iter() 565 offset += copy; in skb_copy_datagram_from_iter() 589 offset += copy; in skb_copy_datagram_from_iter() 604 offset - start, in skb_copy_datagram_from_iter() [all …]
|
| A D | skbuff.c | 2686 offset = end; in ___pskb_trim() 2721 offset = end; in ___pskb_trim() 2956 offset += copy; in skb_copy_bits() 3043 *offset = pfrag->offset; in linear_to_page() 3081 spd->partial[spd->nr_pages].offset = offset; in spd_fill_page() 3260 offset += ret; in __skb_send_sock() 3306 offset += ret; in __skb_send_sock() 3310 offset = 0; in __skb_send_sock() 3382 offset += copy; in skb_store_bits() 3462 offset += copy; in skb_checksum() [all …]
|
| /net/sched/ |
| A D | sch_mqprio_lib.c | 27 unsigned int last = qopt->offset[i] + qopt->count[i]; in mqprio_validate_queue_counts() 38 if (qopt->offset[i] >= dev->real_num_tx_queues || in mqprio_validate_queue_counts() 42 qopt->count[i], qopt->offset[i], in mqprio_validate_queue_counts() 52 if (intervals_overlap(qopt->offset[i], last, in mqprio_validate_queue_counts() 53 qopt->offset[j], in mqprio_validate_queue_counts() 54 qopt->offset[j] + in mqprio_validate_queue_counts() 58 i, qopt->count[i], qopt->offset[i], in mqprio_validate_queue_counts() 59 j, qopt->count[j], qopt->offset[j]); in mqprio_validate_queue_counts() 112 qopt->offset[tc] = dev->tc_to_txq[tc].offset; in mqprio_qopt_reconstruct()
|
| /net/netlabel/ |
| A D | netlabel_kapi.c | 555 u32 offset, in _netlbl_catmap_getnode() argument 726 u32 *offset, in netlbl_catmap_getlong() argument 730 u32 off = *offset; in netlbl_catmap_getlong() 739 *offset = (u32)-1; in netlbl_catmap_getlong() 745 *offset = off; in netlbl_catmap_getlong() 749 *offset = (u32)-1; in netlbl_catmap_getlong() 843 u32 offset, in netlbl_catmap_setlong() argument 858 offset -= iter->startbit; in netlbl_catmap_setlong() 889 if (offset >= bitmap_len) in netlbl_bitmap_walk() 891 byte_offset = offset / 8; in netlbl_bitmap_walk() [all …]
|
| /net/rxrpc/ |
| A D | rxgk.c | 331 txb->offset += offset + shdr; in rxgk_alloc_txbuf() 476 unsigned int offset = sp->offset, len = sp->len; in rxgk_verify_packet_integrity() local 506 sp->offset = offset; in rxgk_verify_packet_integrity() 525 unsigned int offset = sp->offset, len = sp->len; in rxgk_verify_packet_encrypted() local 550 offset += sizeof(hdr); in rxgk_verify_packet_encrypted() 564 sp->offset = offset; in rxgk_verify_packet_encrypted() 856 offset += sizeof(a); in rxgk_construct_authenticator() 867 offset += ret; in rxgk_construct_authenticator() 888 size_t offset, in rxgk_encrypt_authenticator() argument 944 offset = ret; in rxgk_construct_response() [all …]
|
| /net/xfrm/ |
| A D | espintcp.c | 121 if (skb->len < rxm->offset + 2) in espintcp_parse() 195 emsg->offset, emsg->len); in espintcp_sendskb_locked() 200 emsg->offset += ret; in espintcp_sendskb_locked() 225 int offset = sg->offset + emsg->offset; in espintcp_sendskmsg_locked() local 228 emsg->offset = 0; in espintcp_sendskmsg_locked() 239 emsg->offset = offset - sg->offset; in espintcp_sendskmsg_locked() 245 offset += ret; in espintcp_sendskmsg_locked() 295 int offset; in espintcp_push_skb() local 302 offset = skb_transport_offset(skb); in espintcp_push_skb() 303 len = skb->len - offset; in espintcp_push_skb() [all …]
|
| /net/sunrpc/ |
| A D | socklib.c | 29 unsigned int offset; member 52 csum = skb_copy_and_csum_bits(desc->skb, desc->offset, to, len); in xdr_skb_read_bits() 53 desc->csum = csum_block_add(desc->csum, csum, desc->offset); in xdr_skb_read_bits() 55 if (unlikely(skb_copy_bits(desc->skb, desc->offset, to, len))) in xdr_skb_read_bits() 60 desc->offset += len; in xdr_skb_read_bits() 130 .count = skb->len - desc.offset, in csum_partial_copy_to_xdr() 142 desc.csum = csum_partial(skb->data, desc.offset, skb->csum); in csum_partial_copy_to_xdr() 145 if (desc.offset != skb->len) { in csum_partial_copy_to_xdr() 147 csum2 = skb_checksum(skb, desc.offset, skb->len - desc.offset, 0); in csum_partial_copy_to_xdr() 148 desc.csum = csum_block_add(desc.csum, csum2, desc.offset); in csum_partial_copy_to_xdr()
|