Home
last modified time | relevance | path

Searched refs:rn (Results 1 – 25 of 49) sorted by relevance

12

/drivers/scsi/csiostor/
A Dcsio_rnode.c221 if (!rn) in csio_alloc_rnode()
230 return rn; in csio_alloc_rnode()
263 if (!rn) { in csio_get_rnode()
265 if (!rn) in csio_get_rnode()
271 return rn; in csio_get_rnode()
322 if (!rn) { in csio_confirm_rnode()
334 if (!rn) in csio_confirm_rnode()
421 return rn; in csio_confirm_rnode()
425 if (!rn) in csio_confirm_rnode()
433 return rn; in csio_confirm_rnode()
[all …]
A Dcsio_rnode.h120 #define csio_rn_flowid(rn) ((rn)->flowid) argument
121 #define csio_rn_wwpn(rn) ((rn)->rn_sparm.wwpn) argument
122 #define csio_rn_wwnn(rn) ((rn)->rn_sparm.wwnn) argument
123 #define csio_rnode_to_lnode(rn) ((rn)->lnp) argument
125 int csio_is_rnode_ready(struct csio_rnode *rn);
126 void csio_rnode_state_to_str(struct csio_rnode *rn, int8_t *str);
132 void csio_rnode_fwevt_handler(struct csio_rnode *rn, uint8_t fwevt);
134 void csio_put_rnode(struct csio_lnode *ln, struct csio_rnode *rn);
A Dcsio_attr.c72 if (rn->role & CSIO_RNFR_INITIATOR || rn->role & CSIO_RNFR_TARGET) { in csio_reg_rnode()
73 rport = rn->rport; in csio_reg_rnode()
79 if (!rn->rport) { in csio_reg_rnode()
81 rn->nport_id); in csio_reg_rnode()
86 rport = rn->rport; in csio_reg_rnode()
91 sp = &rn->rn_sparm; in csio_reg_rnode()
109 rn->nport_id, ids.roles); in csio_reg_rnode()
316 if (rn) in csio_get_host_fabric_name()
683 struct csio_rnode *rn; in csio_dev_loss_tmo_callbk() local
698 rn, rn->nport_id, csio_rn_flowid(rn)); in csio_dev_loss_tmo_callbk()
[all …]
A Dcsio_scsi.c777 struct csio_rnode *rn; in csio_scsis_io_active() local
800 rn = req->rnode; in csio_scsis_io_active()
806 csio_is_rnode_ready(rn)) { in csio_scsis_io_active()
810 &rn->host_cmpl_q); in csio_scsis_io_active()
1643 rn->flowid); in csio_scsi_err_handler()
1835 ioreq->rnode = rn; in csio_queuecommand()
1932 if (!rn) in csio_eh_abort_handler()
2076 if (!rn) in csio_eh_lun_reset_handler()
2080 cmnd->device->lun, rn->flowid, rn->scsi_id); in csio_eh_lun_reset_handler()
2122 ioreq->rnode = rn; in csio_eh_lun_reset_handler()
[all …]
A Dcsio_lnode.c939 struct csio_rnode *rn; in csio_post_event_rns() local
942 rn = (struct csio_rnode *) tmp; in csio_post_event_rns()
943 csio_post_event(&rn->sm, evt); in csio_post_event_rns()
961 struct csio_rnode *rn; in csio_cleanup_rns() local
964 rn = (struct csio_rnode *) tmp; in csio_cleanup_rns()
965 csio_put_rnode(ln, rn); in csio_cleanup_rns()
1486 struct csio_rnode *rn; in csio_fcoe_fwevt_handler() local
1564 if (!rn) { in csio_fcoe_fwevt_handler()
1730 struct csio_rnode *rn = io_req->rnode; in csio_ln_mgmt_submit_wr() local
1763 ln->nport_id, rn->nport_id, in csio_ln_mgmt_submit_wr()
[all …]
/drivers/s390/char/
A Dsclp_cmd.c177 u16 rn; member
183 u16 rn; member
207 sccb->rn = rn; in do_assign_storage()
234 start = rn2addr(rn); in sclp_assign_storage()
292 istart = rn2addr(incr->rn); in sclp_mem_change_state()
313 istart = rn2addr(incr->rn); in contains_standby_increment()
413 if (rn && first_rn && (first_rn + num == rn)) { in add_memory_merged()
434 first_rn = rn; in add_memory_merged()
457 new_incr->rn = rn; in insert_increment()
462 if (assigned && incr->rn > rn) in insert_increment()
[all …]
A Dsclp_early_core.c358 unsigned long rn, rzm; in sclp_early_read_storage_info() local
385 rn = sccb->entries[sn] >> 16; in sclp_early_read_storage_info()
386 add_physmem_online_range((rn - 1) * rzm, rn * rzm); in sclp_early_read_storage_info()
/drivers/w1/
A Dw1.c416 rn->id = id; in w1_atoreg_num()
423 rn->family, (unsigned long long)rn->id, rn->crc); in w1_atoreg_num()
737 rn->family, (unsigned long long)rn->id, rn->crc); in w1_attach_slave_device()
749 rn->family, rn->family, in w1_attach_slave_device()
750 (unsigned long long)rn->id, rn->crc); in w1_attach_slave_device()
890 memcpy(&rn, &sl->reg_num, sizeof(rn)); in w1_reconnect_slaves()
954 if (rn && w1_addr_crc_is_valid(dev, rn)) in w1_slave_found()
995 last_rn = rn; in w1_search()
996 rn = 0; in w1_search()
1019 cb(dev, rn); in w1_search()
[all …]
A Dw1_internal.h44 void w1_slave_found(struct w1_master *dev, u64 rn);
48 struct w1_reg_num *rn);
57 int w1_attach_slave_device(struct w1_master *dev, struct w1_reg_num *rn);
A Dw1_netlink.c234 static void w1_send_slave(struct w1_master *dev, u64 rn) in w1_send_slave() argument
250 *data = rn; in w1_send_slave()
256 static void w1_found_send_slave(struct w1_master *dev, u64 rn) in w1_found_send_slave() argument
259 w1_slave_found(dev, rn); in w1_found_send_slave()
261 w1_send_slave(dev, rn); in w1_found_send_slave()
273 u64 rn; in w1_get_slaves() local
276 memcpy(&rn, &sl->reg_num, sizeof(rn)); in w1_get_slaves()
277 w1_send_slave(dev, rn); in w1_get_slaves()
A Dw1_io.c396 u64 rn = le64_to_cpu(*((u64*)&sl->reg_num)); in w1_reset_select_slave() local
398 memcpy(&match[1], &rn, 8); in w1_reset_select_slave()
/drivers/infiniband/hw/hfi1/
A Dipoib_main.c183 struct rdma_netdev *rn = netdev_priv(netdev); in hfi1_ipoib_setup_rn() local
187 rn->send = hfi1_ipoib_send; in hfi1_ipoib_setup_rn()
188 rn->tx_timeout = hfi1_ipoib_tx_timeout; in hfi1_ipoib_setup_rn()
189 rn->attach_mcast = hfi1_ipoib_mcast_attach; in hfi1_ipoib_setup_rn()
190 rn->detach_mcast = hfi1_ipoib_mcast_detach; in hfi1_ipoib_setup_rn()
191 rn->set_id = hfi1_ipoib_set_id; in hfi1_ipoib_setup_rn()
192 rn->hca = device; in hfi1_ipoib_setup_rn()
193 rn->port_num = port_num; in hfi1_ipoib_setup_rn()
194 rn->mtu = netdev->mtu; in hfi1_ipoib_setup_rn()
A Dvnic_main.c563 struct rdma_netdev *rn; in hfi1_vnic_alloc_rn() local
582 rn = netdev_priv(netdev); in hfi1_vnic_alloc_rn()
588 rn->free_rdma_netdev = hfi1_vnic_free_rn; in hfi1_vnic_alloc_rn()
589 rn->set_id = hfi1_vnic_set_vesw_id; in hfi1_vnic_alloc_rn()
A Dipoib.h136 struct rdma_netdev rn; /* keep this first */ member
/drivers/infiniband/ulp/opa_vnic/
A Dopa_vnic_netdev.c137 struct rdma_netdev *rn = netdev_priv(adapter->netdev); in opa_vnic_process_vema_config() local
157 rn->set_id(netdev, info->vesw.vesw_id); in opa_vnic_process_vema_config()
327 struct rdma_netdev *rn; in opa_vnic_add_netdev() local
339 rn = netdev_priv(netdev); in opa_vnic_add_netdev()
346 rn->clnt_priv = adapter; in opa_vnic_add_netdev()
347 rn->hca = ibdev; in opa_vnic_add_netdev()
348 rn->port_num = port_num; in opa_vnic_add_netdev()
382 rn->free_rdma_netdev(netdev); in opa_vnic_add_netdev()
391 struct rdma_netdev *rn = netdev_priv(netdev); in opa_vnic_rem_netdev() local
399 rn->free_rdma_netdev(netdev); in opa_vnic_rem_netdev()
/drivers/md/persistent-data/
A Ddm-btree.c659 struct btree_node *ln, *rn, *pn; in split_one_into_two() local
669 rn = dm_block_data(right); in split_one_into_two()
671 rn->header.flags = ln->header.flags; in split_one_into_two()
675 redistribute2(ln, rn); in split_one_into_two()
691 if (key < le64_to_cpu(rn->keys[0])) { in split_one_into_two()
769 rn = dm_block_data(right); in split_two_into_three()
776 redistribute3(ln, mn, rn); in split_two_into_three()
844 struct btree_node *pn, *ln, *rn; in btree_split_beneath() local
875 rn = dm_block_data(right); in btree_split_beneath()
878 rn->header.flags = pn->header.flags; in btree_split_beneath()
[all …]
/drivers/gpu/drm/i915/gt/
A Dintel_gt_requests.c19 struct i915_request *rq, *rn; in retire_requests() local
21 list_for_each_entry_safe(rq, rn, &tl->requests, link) in retire_requests()
243 struct i915_request *rq, *rn; in intel_gt_watchdog_work() local
250 llist_for_each_entry_safe(rq, rn, first, watchdog.link) { in intel_gt_watchdog_work()
A Dintel_timeline.c424 struct i915_request *rq, *rn; in intel_gt_show_timelines() local
441 list_for_each_entry_safe(rq, rn, &tl->requests, link) { in intel_gt_show_timelines()
466 list_for_each_entry_safe(rq, rn, &tl->requests, link) in intel_gt_show_timelines()
/drivers/net/ethernet/mellanox/mlx5/core/ipoib/
A Dipoib.c772 struct rdma_netdev *rn; in mlx5_rdma_setup_rn() local
807 rn = &ipriv->rn; in mlx5_rdma_setup_rn()
808 rn->hca = ibdev; in mlx5_rdma_setup_rn()
809 rn->send = mlx5i_xmit; in mlx5_rdma_setup_rn()
810 rn->attach_mcast = mlx5i_attach_mcast; in mlx5_rdma_setup_rn()
811 rn->detach_mcast = mlx5i_detach_mcast; in mlx5_rdma_setup_rn()
812 rn->set_id = mlx5i_set_pkey_index; in mlx5_rdma_setup_rn()
A Dipoib.h54 struct rdma_netdev rn; /* keep this first */ member
/drivers/infiniband/ulp/ipoib/
A Dipoib_main.c992 struct rdma_netdev *rn = netdev_priv(dev); in neigh_add_path() local
1085 struct rdma_netdev *rn = netdev_priv(dev); in unicast_arp_send() local
1232 if (rn->tx_timeout) { in ipoib_timeout()
1233 rn->tx_timeout(dev, txqueue); in ipoib_timeout()
2008 rn->mtu = priv->mcast_mtu; in ipoib_ndo_init()
2266 rn->send = ipoib_send; in ipoib_intf_init()
2267 rn->attach_mcast = ipoib_mcast_attach; in ipoib_intf_init()
2268 rn->detach_mcast = ipoib_mcast_detach; in ipoib_intf_init()
2269 rn->hca = hca; in ipoib_intf_init()
2287 rn->clnt_priv = priv; in ipoib_intf_init()
[all …]
A Dipoib_multicast.c215 struct rdma_netdev *rn = netdev_priv(dev); in ipoib_mcast_join_finish() local
248 rn->mtu = priv->mcast_mtu; in ipoib_mcast_join_finish()
264 ret = rn->attach_mcast(dev, priv->ca, &mcast->mcmember.mgid, in ipoib_mcast_join_finish()
680 struct rdma_netdev *rn = netdev_priv(dev); in ipoib_mcast_leave() local
694 ret = rn->detach_mcast(dev, priv->ca, &mcast->mcmember.mgid, in ipoib_mcast_leave()
745 struct rdma_netdev *rn = netdev_priv(dev); in ipoib_mcast_send() local
812 mcast->ah->last_send = rn->send(dev, skb, mcast->ah->ah, in ipoib_mcast_send()
A Dipoib_vlan.c102 struct rdma_netdev *rn = netdev_priv(ndev); in __ipoib_vlan_add() local
121 rn->mtu = priv->mcast_mtu; in __ipoib_vlan_add()
/drivers/gpu/drm/msm/
A Dmsm_gpu.h512 unsigned rn, sp; in msm_gpu_convert_priority() local
514 rn = div_u64_rem(prio, NR_SCHED_PRIORITIES, &sp); in msm_gpu_convert_priority()
521 if (rn >= gpu->nr_rings) in msm_gpu_convert_priority()
524 *ring_nr = rn; in msm_gpu_convert_priority()
/drivers/w1/slaves/
A Dw1_ds2408.c295 u64 rn = le64_to_cpu(*((u64 *)&sl->reg_num)); in w1_f29_disable_test_mode() local
297 memcpy(&magic[1], &rn, 8); in w1_f29_disable_test_mode()

Completed in 76 milliseconds

12