| /net/rxrpc/ |
| A D | rxgk_kdf.c | 60 gfp_t gfp) in rxgk_derive_transport_key() argument 98 gfp_t gfp) in rxgk_set_up_ciphers() argument 108 buffer = kzalloc(krb5->key_bytes, gfp); in rxgk_set_up_ciphers() 137 krb5, &TK, RXGK_SERVER_MIC_PACKET, gfp); in rxgk_set_up_ciphers() 142 krb5, &TK, RXGK_CLIENT_MIC_PACKET, gfp); in rxgk_set_up_ciphers() 149 krb5, &TK, RXGK_SERVER_ENC_PACKET, gfp); in rxgk_set_up_ciphers() 154 krb5, &TK, RXGK_CLIENT_ENC_PACKET, gfp); in rxgk_set_up_ciphers() 164 krb5, &TK, RXGK_CLIENT_MIC_PACKET, gfp); in rxgk_set_up_ciphers() 169 krb5, &TK, RXGK_SERVER_MIC_PACKET, gfp); in rxgk_set_up_ciphers() 208 gfp_t gfp) in rxgk_generate_transport_key() argument [all …]
|
| A D | call_object.c | 121 struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *rx, gfp_t gfp, in rxrpc_alloc_call() argument 127 call = kmem_cache_zalloc(rxrpc_call_jar, gfp); in rxrpc_alloc_call() 191 gfp_t gfp, in rxrpc_alloc_client_call() argument 200 call = rxrpc_alloc_call(rx, gfp, debug_id); in rxrpc_alloc_client_call() 261 static struct semaphore *rxrpc_get_call_slot(struct rxrpc_call_params *p, gfp_t gfp) in rxrpc_get_call_slot() argument 291 static int rxrpc_connect_call(struct rxrpc_call *call, gfp_t gfp) in rxrpc_connect_call() argument 298 ret = rxrpc_look_up_bundle(call, gfp); in rxrpc_connect_call() 323 gfp_t gfp, in rxrpc_new_client_call() argument 341 limiter = rxrpc_get_call_slot(p, gfp); in rxrpc_new_client_call() 347 call = rxrpc_alloc_client_call(rx, cp, p, gfp, debug_id); in rxrpc_new_client_call() [all …]
|
| A D | call_accept.c | 37 unsigned long user_call_ID, gfp_t gfp, in rxrpc_service_prealloc_one() argument 73 peer = rxrpc_alloc_peer(rx->local, gfp, rxrpc_peer_new_prealloc); in rxrpc_service_prealloc_one() 86 conn = rxrpc_prealloc_service_connection(rxnet, gfp); in rxrpc_service_prealloc_one() 97 call = rxrpc_alloc_call(rx, gfp, debug_id); in rxrpc_service_prealloc_one() 162 int rxrpc_service_prealloc(struct rxrpc_sock *rx, gfp_t gfp) in rxrpc_service_prealloc() argument 167 b = kzalloc(sizeof(struct rxrpc_backlog), gfp); in rxrpc_service_prealloc() 471 unsigned long user_call_ID, gfp_t gfp, in rxrpc_kernel_charge_accept() argument 481 gfp, debug_id); in rxrpc_kernel_charge_accept()
|
| A D | txbuf.c | 20 size_t data_align, gfp_t gfp) in rxrpc_alloc_data_txbuf() argument 26 txb = kzalloc(sizeof(*txb), gfp); in rxrpc_alloc_data_txbuf() 38 buf = page_frag_alloc_align(&call->conn->tx_data_alloc, total, gfp, in rxrpc_alloc_data_txbuf()
|
| A D | peer_object.c | 222 struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *local, gfp_t gfp, in rxrpc_alloc_peer() argument 229 peer = kzalloc(sizeof(struct rxrpc_peer), gfp); in rxrpc_alloc_peer() 287 gfp_t gfp) in rxrpc_create_peer() argument 293 peer = rxrpc_alloc_peer(local, gfp, rxrpc_peer_new_client); in rxrpc_create_peer() 335 struct sockaddr_rxrpc *srx, gfp_t gfp) in rxrpc_lookup_peer() argument 354 candidate = rxrpc_create_peer(local, srx, hash_key, gfp); in rxrpc_lookup_peer()
|
| /net/wireless/ |
| A D | nl80211.h | 60 const u8 *buf, size_t len, gfp_t gfp); 67 bool reconnect, gfp_t gfp); 71 bool reconnect, gfp_t gfp); 74 const u8 *addr, gfp_t gfp); 77 const u8 *addr, gfp_t gfp); 81 gfp_t gfp); 84 struct cfg80211_roam_info *info, gfp_t gfp); 99 int key_id, const u8 *tsc, gfp_t gfp); 108 gfp_t gfp); 112 struct cfg80211_rx_info *info, gfp_t gfp); [all …]
|
| /net/sctp/ |
| A D | ulpevent.c | 49 gfp_t gfp) in sctp_ulpevent_new() argument 54 skb = alloc_skb(size, gfp); in sctp_ulpevent_new() 149 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_assoc_change() 251 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_peer_addr_change() 376 gfp_t gfp) in sctp_ulpevent_make_remote_error() argument 445 gfp); in sctp_ulpevent_make_send_failed() 579 __u16 flags, gfp_t gfp) in sctp_ulpevent_make_shutdown_event() argument 586 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_shutdown_event() 648 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_adaptation_indication() 677 gfp_t gfp) in sctp_ulpevent_make_rcvmsg() argument [all …]
|
| A D | auth.c | 188 gfp_t gfp) in sctp_auth_make_key_vector() argument 202 new = sctp_auth_create_key(len, gfp); in sctp_auth_make_key_vector() 223 gfp_t gfp) in sctp_auth_make_local_vector() argument 234 gfp_t gfp) in sctp_auth_make_peer_vector() argument 239 gfp); in sctp_auth_make_peer_vector() 256 gfp_t gfp) in sctp_auth_asoc_set_secret() argument 289 gfp_t gfp) in sctp_auth_asoc_create_secret() argument 341 gfp); in sctp_auth_asoc_create_secret() 355 gfp_t gfp) in sctp_auth_asoc_copy_shkeys() argument 466 gfp); in sctp_auth_init_hmacs() [all …]
|
| A D | bind_addr.c | 35 gfp_t gfp, int flags); 45 enum sctp_scope scope, gfp_t gfp, in sctp_bind_addr_copy() argument 57 gfp, flags); in sctp_bind_addr_copy() 69 SCTP_SCOPE_LINK, gfp, in sctp_bind_addr_copy() 96 gfp_t gfp) in sctp_bind_addr_dup() argument 106 1, gfp); in sctp_bind_addr_dup() 213 gfp_t gfp) in sctp_bind_addrs_to_raw() argument 240 retval.v = kmalloc(len, gfp); in sctp_bind_addrs_to_raw() 288 SCTP_ADDR_SRC, gfp); in sctp_raw_to_bind_addrs() 461 gfp_t gfp, int flags) in sctp_copy_one_addr() argument [all …]
|
| A D | stream_sched_prio.c | 41 struct sctp_stream *stream, int prio, gfp_t gfp) in sctp_sched_prio_new_head() argument 45 p = kmalloc(sizeof(*p), gfp); in sctp_sched_prio_new_head() 59 struct sctp_stream *stream, int prio, gfp_t gfp) in sctp_sched_prio_get_head() argument 90 return sctp_sched_prio_new_head(stream, prio, gfp); in sctp_sched_prio_get_head() 164 __u16 prio, gfp_t gfp) in sctp_sched_prio_set() argument 175 prio_head = sctp_sched_prio_get_head(stream, prio, gfp); in sctp_sched_prio_set() 203 gfp_t gfp) in sctp_sched_prio_init_sid() argument 206 return sctp_sched_prio_set(stream, sid, 0, gfp); in sctp_sched_prio_init_sid()
|
| A D | outqueue.c | 313 sctp_outq_flush(q, 0, gfp); in sctp_outq_tail() 675 sctp_packet_transmit(pkt, gfp); in __sctp_outq_flush_rtx() 764 sctp_outq_flush(q, 0, gfp); in sctp_outq_uncork() 768 struct sctp_chunk *chunk, gfp_t gfp) in sctp_packet_singleton() argument 796 gfp_t gfp; member 915 ctx->gfp); in sctp_outq_flush_ctrl() 946 chunk, ctx->gfp); in sctp_outq_flush_ctrl() 960 one_packet, ctx->gfp); in sctp_outq_flush_ctrl() 1013 &start_timer, ctx->gfp); in sctp_outq_flush_rtx() 1112 ctx->gfp); in sctp_outq_flush_data() [all …]
|
| A D | sm_sideeffect.c | 48 gfp_t gfp); 57 gfp_t gfp); 703 gfp_t gfp) in sctp_cmd_process_init() argument 1152 void *event_arg, gfp_t gfp) in sctp_do_sm() argument 1177 &commands, gfp); in sctp_do_sm() 1194 gfp_t gfp) in sctp_side_effects() argument 1207 commands, gfp))) in sctp_side_effects() 1283 gfp_t gfp) in sctp_cmd_interpreter() argument 1396 cmd->obj.init, gfp); in sctp_cmd_interpreter() 1493 sctp_packet_transmit(packet, gfp); in sctp_cmd_interpreter() [all …]
|
| A D | stream_sched.c | 26 __u16 value, gfp_t gfp) in sctp_sched_fcfs_set() argument 44 gfp_t gfp) in sctp_sched_fcfs_init_sid() argument 207 __u16 value, gfp_t gfp) in sctp_sched_set_value() argument 220 return asoc->outqueue.sched->set(&asoc->stream, sid, value, gfp); in sctp_sched_set_value() 264 int sctp_sched_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp) in sctp_sched_init_sid() argument 270 return sched->init_sid(stream, sid, gfp); in sctp_sched_init_sid()
|
| A D | endpointola.c | 43 gfp_t gfp) in sctp_endpoint_init() argument 48 ep->digest = kzalloc(SCTP_SIGNATURE_SIZE, gfp); in sctp_endpoint_init() 55 if (sctp_auth_init(ep, gfp)) in sctp_endpoint_init() 98 null_key = sctp_auth_shkey_create(0, gfp); in sctp_endpoint_init() 129 struct sctp_endpoint *sctp_endpoint_new(struct sock *sk, gfp_t gfp) in sctp_endpoint_new() argument 134 ep = kzalloc(sizeof(*ep), gfp); in sctp_endpoint_new() 138 if (!sctp_endpoint_init(ep, sk, gfp)) in sctp_endpoint_new()
|
| A D | sm_make_chunk.c | 51 gfp_t gfp); 53 __u8 flags, int paylen, gfp_t gfp); 56 gfp_t gfp); 66 gfp_t gfp); 200 gfp_t gfp, int vparam_len) in sctp_make_init() argument 1369 struct sock *sk, gfp_t gfp) in sctp_chunkify() argument 1422 gfp_t gfp) in _sctp_make_chunk() argument 1435 skb = alloc_skb(chunklen, gfp); in _sctp_make_chunk() 1478 gfp_t gfp) in sctp_make_control() argument 1622 gfp_t gfp) in sctp_make_temp_asoc() argument [all …]
|
| A D | associola.c | 55 enum sctp_scope scope, gfp_t gfp) in sctp_association_init() argument 256 if (sctp_auth_asoc_copy_shkeys(ep, asoc, gfp)) in sctp_association_init() 288 enum sctp_scope scope, gfp_t gfp) in sctp_association_new() argument 292 asoc = kzalloc(sizeof(*asoc), gfp); in sctp_association_new() 585 const gfp_t gfp, in sctp_assoc_add_peer() argument 1555 enum sctp_scope scope, gfp_t gfp) in sctp_assoc_set_bind_addr_from_ep() argument 1574 scope, gfp, flags); in sctp_assoc_set_bind_addr_from_ep() 1580 gfp_t gfp) in sctp_assoc_set_bind_addr_from_cookie() argument 1588 asoc->ep->base.bind_addr.port, gfp); in sctp_assoc_set_bind_addr_from_cookie() 1608 bool preload = gfpflags_allow_blocking(gfp); in sctp_assoc_set_id() [all …]
|
| A D | output.c | 182 int one_packet, gfp_t gfp) in sctp_packet_transmit_chunk() argument 194 error = sctp_packet_transmit(packet, gfp); in sctp_packet_transmit_chunk() 437 struct sk_buff *head, int gso, gfp_t gfp) in sctp_packet_pack() argument 472 nskb = alloc_skb(pkt_size + MAX_HEADER, gfp); in sctp_packet_pack() 521 packet->auth->shkey, gfp); in sctp_packet_pack() 569 int sctp_packet_transmit(struct sctp_packet *packet, gfp_t gfp) in sctp_packet_transmit() argument 599 MAX_HEADER, gfp); in sctp_packet_transmit() 621 pkt_count = sctp_packet_pack(packet, head, gso, gfp); in sctp_packet_transmit()
|
| /net/rds/ |
| A D | tcp_recv.c | 152 gfp_t gfp; member 175 arg->gfp); in rds_tcp_data_recv() 219 clone = pskb_extract(skb, offset, to_copy, arg->gfp); in rds_tcp_data_recv() 246 arg->gfp); in rds_tcp_data_recv() 263 static int rds_tcp_read_sock(struct rds_conn_path *cp, gfp_t gfp) in rds_tcp_read_sock() argument 272 arg.gfp = gfp; in rds_tcp_read_sock() 278 rdsdebug("tcp_read_sock for tc %p gfp 0x%x returned %d\n", tc, gfp, in rds_tcp_read_sock()
|
| A D | page.c | 71 gfp_t gfp) in rds_page_remainder_alloc() argument 77 gfp |= __GFP_HIGHMEM; in rds_page_remainder_alloc() 81 page = alloc_page(gfp); in rds_page_remainder_alloc() 124 page = alloc_page(gfp); in rds_page_remainder_alloc()
|
| A D | ib_recv.c | 107 cache->percpu = alloc_percpu_gfp(struct rds_ib_cache_head, gfp); in rds_ib_recv_alloc_cache() 122 int rds_ib_recv_alloc_caches(struct rds_ib_connection *ic, gfp_t gfp) in rds_ib_recv_alloc_caches() argument 126 ret = rds_ib_recv_alloc_cache(&ic->i_cache_incs, gfp); in rds_ib_recv_alloc_caches() 128 ret = rds_ib_recv_alloc_cache(&ic->i_cache_frags, gfp); in rds_ib_recv_alloc_caches() 309 struct rds_ib_recv_work *recv, gfp_t gfp) in rds_ib_recv_refill_one() argument 314 gfp_t slab_mask = gfp; in rds_ib_recv_refill_one() 315 gfp_t page_mask = gfp; in rds_ib_recv_refill_one() 317 if (gfp & __GFP_DIRECT_RECLAIM) { in rds_ib_recv_refill_one() 383 void rds_ib_recv_refill(struct rds_connection *conn, int prefill, gfp_t gfp) in rds_ib_recv_refill() argument 389 bool can_wait = !!(gfp & __GFP_DIRECT_RECLAIM); in rds_ib_recv_refill() [all …]
|
| /net/core/ |
| A D | page_pool.c | 498 PP_DMA_INDEX_LIMIT, gfp); in page_pool_dma_map() 501 PP_DMA_INDEX_LIMIT, gfp); in page_pool_dma_map() 522 gfp_t gfp) in __page_pool_alloc_page_order() argument 526 gfp |= __GFP_COMP; in __page_pool_alloc_page_order() 548 gfp_t gfp) in __page_pool_alloc_netmems_slow() argument 616 netmem = pool->mp_ops->alloc_netmems(pool, gfp); in page_pool_alloc_netmems() 618 netmem = __page_pool_alloc_netmems_slow(pool, gfp); in page_pool_alloc_netmems() 1017 gfp_t gfp) in page_pool_alloc_frag_netmem() argument 1038 netmem = page_pool_alloc_netmems(pool, gfp); in page_pool_alloc_frag_netmem() 1061 unsigned int size, gfp_t gfp) in page_pool_alloc_frag() argument [all …]
|
| A D | mp_dmabuf_devmem.h | 16 netmem_ref mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp); 28 mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp) in mp_dmabuf_devmem_alloc_netmems() argument
|
| /net/handshake/ |
| A D | handshake-test.c | 36 gfp_t gfp; member 75 .gfp = GFP_KERNEL, 81 .gfp = GFP_KERNEL, 87 .gfp = GFP_KERNEL, 93 .gfp = GFP_KERNEL, 99 .gfp = GFP_KERNEL, 105 .gfp = GFP_KERNEL | __GFP_NOWARN, 111 .gfp = GFP_KERNEL, 135 result = handshake_req_alloc(param->proto, param->gfp); in handshake_req_alloc_case()
|
| /net/ceph/ |
| A D | buffer.c | 12 struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp) in ceph_buffer_new() argument 16 b = kmalloc(sizeof(*b), gfp); in ceph_buffer_new() 20 b->vec.iov_base = kvmalloc(len, gfp); in ceph_buffer_new()
|
| /net/netfilter/ |
| A D | nft_limit.c | 153 const struct nft_limit_priv *priv_src, gfp_t gfp) in nft_limit_clone() argument 161 priv_dst->limit = kmalloc(sizeof(*priv_dst->limit), gfp); in nft_limit_clone() 227 gfp_t gfp) in nft_limit_pkts_clone() argument 234 return nft_limit_clone(&priv_dst->limit, &priv_src->limit, gfp); in nft_limit_pkts_clone() 286 gfp_t gfp) in nft_limit_bytes_clone() argument 291 return nft_limit_clone(priv_dst, priv_src, gfp); in nft_limit_bytes_clone()
|