| /net/rds/ |
| A D | message.c | 53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref() 139 if (rm->m_rs) { in rds_message_purge() 149 rm->m_rs = NULL; in rds_message_purge() 182 kfree(rm); in rds_message_put() 292 if (!rm) in rds_message_alloc() 305 return rm; in rds_message_alloc() 321 if (rm->m_used_sgs + nents > rm->m_total_sgs) { in rds_message_alloc_sgs() 323 rm->m_total_sgs, rm->m_used_sgs, nents); in rds_message_alloc_sgs() 342 if (!rm) in rds_message_map_pages() 348 rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs); in rds_message_map_pages() [all …]
|
| A D | send.c | 203 if (!rm) { in rds_send_xmit() 238 if (!rm) { in rds_send_xmit() 269 if (!rm) in rds_send_xmit() 310 rm->m_final_op = &rm->rdma; in rds_send_xmit() 326 rm->m_final_op = &rm->atomic; in rds_send_xmit() 353 if (rm->atomic.op_active && !rm->atomic.op_silent) in rds_send_xmit() 355 if (rm->rdma.op_active && !rm->rdma.op_silent) in rds_send_xmit() 364 rm->m_final_op = &rm->data; in rds_send_xmit() 1268 if (!rm) { in rds_sendmsg() 1407 if (rm) in rds_sendmsg() [all …]
|
| A D | tcp_send.c | 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() 98 (unsigned long long)rm->m_ack_seq); in rds_tcp_xmit() 115 while (sg < rm->data.op_nents) { in rds_tcp_xmit() 117 if (sg + 1 < rm->data.op_nents) in rds_tcp_xmit() 121 rm->data.op_sg[sg].length - off, in rds_tcp_xmit() 122 rm->data.op_sg[sg].offset + off); in rds_tcp_xmit() 124 rm->data.op_sg[sg].length - off); in rds_tcp_xmit() 127 rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, in rds_tcp_xmit() [all …]
|
| A D | rdma.c | 626 || rm->rdma.op_active) in rds_cmsg_rdma_args() 820 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest() 823 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest() 843 rm->rdma.op_rdma_mr = mr; in rds_cmsg_rdma_dest() 862 &rm->rdma.op_rdma_mr, rm->m_conn_path); in rds_cmsg_rdma_map() 876 || rm->atomic.op_active) in rds_cmsg_atomic() 913 rm->atomic.op_active = 1; in rds_cmsg_atomic() 915 rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1); in rds_cmsg_atomic() 934 if (rm->atomic.op_notify || rm->atomic.op_recverr) { in rds_cmsg_atomic() 940 rm->atomic.op_notifier = kmalloc(sizeof(*rm->atomic.op_notifier), GFP_KERNEL); in rds_cmsg_atomic() [all …]
|
| A D | ib_send.c | 187 return rm; in rds_ib_send_unmap_op() 288 rds_message_put(rm); in rds_ib_send_cqe_handler() 552 rm->data.op_sg, in rds_ib_xmit() 555 rdsdebug("ic %p mapping rm %p: %d\n", ic, rm, rm->data.op_count); in rds_ib_xmit() 563 rm->data.op_count = 0; in rds_ib_xmit() 567 rm->data.op_dmasg = 0; in rds_ib_xmit() 615 if (rm->rdma.op_active && rm->rdma.op_fence) in rds_ib_xmit() 650 && scat != &rm->data.op_sg[rm->data.op_count]) { in rds_ib_xmit() 664 rm->data.op_dmasg++; in rds_ib_xmit() 708 && scat != &rm->data.op_sg[rm->data.op_count]); in rds_ib_xmit() [all …]
|
| A D | loop.c | 75 static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_loop_xmit() argument 79 struct scatterlist *sgp = &rm->data.op_sg[sg]; in rds_loop_xmit() 81 be32_to_cpu(rm->m_inc.i_hdr.h_len); in rds_loop_xmit() 84 if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { in rds_loop_xmit() 92 rds_inc_init(&rm->m_inc, conn, &conn->c_laddr); in rds_loop_xmit() 94 rds_message_addref(rm); in rds_loop_xmit() 96 rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc, in rds_loop_xmit() 99 rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence), in rds_loop_xmit() 102 rds_inc_put(&rm->m_inc); in rds_loop_xmit() 113 struct rds_message *rm = container_of(inc, struct rds_message, m_inc); in rds_loop_inc_free() local [all …]
|
| A D | rds.h | 551 int (*xmit)(struct rds_connection *conn, struct rds_message *rm, 866 void rds_message_addref(struct rds_message *rm); 867 void rds_message_put(struct rds_message *rm); 868 void rds_message_wait(struct rds_message *rm); 869 void rds_message_unmapped(struct rds_message *rm); 916 typedef int (*is_acked_func)(struct rds_message *rm, uint64_t ack); 932 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, 934 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, 937 int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, 941 void rds_rdma_send_complete(struct rds_message *rm, int wc_status); [all …]
|
| A D | cong.c | 421 struct rds_message *rm; in rds_cong_update_alloc() local 423 rm = rds_message_map_pages(map->m_page_addrs, RDS_CONG_MAP_BYTES); in rds_cong_update_alloc() 424 if (!IS_ERR(rm)) in rds_cong_update_alloc() 425 rm->m_inc.i_hdr.h_flags = RDS_FLAG_CONG_BITMAP; in rds_cong_update_alloc() 427 return rm; in rds_cong_update_alloc()
|
| A D | connection.c | 443 struct rds_message *rm, *rtmp; in rds_conn_path_destroy() local 456 list_for_each_entry_safe(rm, rtmp, in rds_conn_path_destroy() 459 list_del_init(&rm->m_conn_item); in rds_conn_path_destroy() 460 BUG_ON(!list_empty(&rm->m_sock_item)); in rds_conn_path_destroy() 461 rds_message_put(rm); in rds_conn_path_destroy() 542 struct rds_message *rm; in rds_conn_message_info_cmn() local 577 list_for_each_entry(rm, list, m_conn_item) { in rds_conn_message_info_cmn() 580 __rds_inc_msg_cp(&rm->m_inc, in rds_conn_message_info_cmn()
|
| A D | tcp.h | 87 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
|
| A D | recv.c | 137 struct rds_message *rm, *tmp; in rds_conn_peer_gen_update() local 151 list_for_each_entry_safe(rm, tmp, in rds_conn_peer_gen_update() 154 set_bit(RDS_MSG_FLUSH, &rm->m_flags); in rds_conn_peer_gen_update()
|
| A D | ib_cm.c | 1163 struct rds_message *rm; in rds_ib_conn_path_shutdown() local 1165 rm = container_of(ic->i_data_op, struct rds_message, data); in rds_ib_conn_path_shutdown() 1166 rds_message_put(rm); in rds_ib_conn_path_shutdown()
|
| A D | ib.h | 424 int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm,
|
| /net/wireless/ |
| A D | sme.c | 1210 ev->rm.req_ie = next; in cfg80211_roamed() 1211 ev->rm.req_ie_len = info->req_ie_len; in cfg80211_roamed() 1216 ev->rm.resp_ie = next; in cfg80211_roamed() 1217 ev->rm.resp_ie_len = info->resp_ie_len; in cfg80211_roamed() 1223 ev->rm.fils.kek = next; in cfg80211_roamed() 1230 ev->rm.fils.pmk = next; in cfg80211_roamed() 1237 ev->rm.fils.pmkid = next; in cfg80211_roamed() 1246 ev->rm.ap_mld_addr = next; in cfg80211_roamed() 1251 ev->rm.valid_links = info->valid_links; in cfg80211_roamed() 1256 ev->rm.links[link].addr = next; in cfg80211_roamed() [all …]
|
| A D | core.h | 275 struct cfg80211_roam_info rm; member
|
| A D | util.c | 1161 __cfg80211_roamed(wdev, &ev->rm); in cfg80211_process_wdev_events()
|
| /net/ipv4/ |
| A D | ipconfig.c | 412 struct rtentry rm; in ic_setup_routes() local 415 memset(&rm, 0, sizeof(rm)); in ic_setup_routes() 420 set_sockaddr((struct sockaddr_in *) &rm.rt_dst, 0, 0); in ic_setup_routes() 421 set_sockaddr((struct sockaddr_in *) &rm.rt_genmask, 0, 0); in ic_setup_routes() 422 set_sockaddr((struct sockaddr_in *) &rm.rt_gateway, ic_gateway, 0); in ic_setup_routes() 423 rm.rt_flags = RTF_UP | RTF_GATEWAY; in ic_setup_routes() 424 if ((err = ip_rt_ioctl(&init_net, SIOCADDRT, &rm)) < 0) { in ic_setup_routes()
|
| /net/tipc/ |
| A D | group.c | 595 struct tipc_member *m, *rm, *pm; in tipc_group_update_rcv_win() local 621 rm = list_first_entry(active, struct tipc_member, list); in tipc_group_update_rcv_win() 622 rm->state = MBR_RECLAIMING; in tipc_group_update_rcv_win() 623 list_del_init(&rm->list); in tipc_group_update_rcv_win() 624 tipc_group_proto_xmit(grp, rm, GRP_RECLAIM_MSG, xmitq); in tipc_group_update_rcv_win()
|
| /net/sunrpc/ |
| A D | xprtsock.c | 955 rpc_fraghdr rm = xs_stream_record_marker(xdr); in xs_local_send_request() local 956 unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen; in xs_local_send_request() 977 transport->xmit.offset, rm, &sent); in xs_local_send_request() 1111 rpc_fraghdr rm = xs_stream_record_marker(xdr); in xs_tcp_send_request() local 1112 unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen; in xs_tcp_send_request() 1148 transport->xmit.offset, rm, &sent); in xs_tcp_send_request()
|
| /net/xfrm/ |
| A D | xfrm_policy.c | 1118 static void xfrm_policy_inexact_gc_tree(struct rb_root *r, bool rm) in xfrm_policy_inexact_gc_tree() argument 1126 xfrm_policy_inexact_gc_tree(&node->root, rm); in xfrm_policy_inexact_gc_tree() 1130 WARN_ON_ONCE(rm); in xfrm_policy_inexact_gc_tree()
|