| /drivers/mtd/nand/raw/ |
| A D | nandsim.c | 659 ns->geom.pgszoob = ns->geom.pgsz + ns->geom.oobsz; in ns_init() 661 ns->geom.totszoob = ns->geom.totsz + (uint64_t)ns->geom.pgnum * ns->geom.oobsz; in ns_init() 664 ns->geom.pgsec = ns->geom.secsz / ns->geom.pgsz; in ns_init() 665 ns->geom.secszoob = ns->geom.secsz + ns->geom.oobsz * ns->geom.pgsec; in ns_init() 1266 ns->state = ns->op[ns->stateidx]; in ns_find_operation() 1267 ns->nxstate = ns->op[ns->stateidx + 1]; in ns_find_operation() 1296 ns->pstates[ns->npstates++] = ns->state; in ns_find_operation() 1525 tx = ns_write_file(ns, ns->cfile, ns->file_buf, in ns_prog_page() 1678 num, ns->regs.row, ns->regs.column, NS_RAW_OFFSET(ns) + ns->regs.off); in ns_do_state_action() 1731 ns->nxstate = ns->op[ns->stateidx + 1]; in ns_switch_state() [all …]
|
| /drivers/net/netdevsim/ |
| A D | ethtool.c | 52 memcpy(coal, &ns->ethtool.coalesce, sizeof(ns->ethtool.coalesce)); in nsim_get_coalesce() 63 memcpy(&ns->ethtool.coalesce, coal, sizeof(ns->ethtool.coalesce)); in nsim_set_coalesce() 74 memcpy(ring, &ns->ethtool.ring, sizeof(ns->ethtool.ring)); in nsim_get_ringparam() 145 if (ns->ethtool.get_err) in nsim_get_fecparam() 146 return -ns->ethtool.get_err; in nsim_get_fecparam() 147 memcpy(fecparam, &ns->ethtool.fec, sizeof(ns->ethtool.fec)); in nsim_get_fecparam() 157 if (ns->ethtool.set_err) in nsim_set_fecparam() 158 return -ns->ethtool.set_err; in nsim_set_fecparam() 159 memcpy(&ns->ethtool.fec, fecparam, sizeof(ns->ethtool.fec)); in nsim_set_fecparam() 220 nsim_ethtool_ring_init(ns); in nsim_ethtool_init() [all …]
|
| A D | netdev.c | 170 if (ns->xdp.prog && !ns->xdp.prog->aux->xdp_has_frags && in nsim_change_mtu() 354 ns, ns, true); in nsim_setup_tc() 894 if (ns->page) in nsim_pp_hold_read() 920 ns->page = page_pool_dev_alloc_pages(ns->rq[0]->page_pool); in nsim_pp_hold_write() 971 ns->rq = kcalloc(dev->num_rx_queues, sizeof(*ns->rq), in nsim_queue_init() 973 if (!ns->rq) in nsim_queue_init() 987 kfree(ns->rq); in nsim_queue_init() 999 kfree(ns->rq); in nsim_queue_uninit() 1000 ns->rq = NULL; in nsim_queue_uninit() 1116 return ns; in nsim_create() [all …]
|
| A D | udp_tunnels.c | 18 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_set_port() 19 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_set_port() 44 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_unset_port() 45 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_unset_port() 74 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_sync_table() 78 ns->udp_ports.ports[table][i] = in nsim_udp_tunnel_sync_table() 111 memset(ns->udp_ports.ports, 0, sizeof(ns->udp_ports.__ports)); in nsim_udp_tunnels_info_reset_write() 138 ns->udp_ports.ports = ns->udp_ports.__ports; in nsim_udp_tunnels_info_create() 143 ns->nsim_dev_port->ddir); in nsim_udp_tunnels_info_create() 148 ns->udp_ports.dfs_ports[0].array = ns->udp_ports.ports[0]; in nsim_udp_tunnels_info_create() [all …]
|
| A D | macsec.c | 11 if (ns->macsec.nsim_secy[i].sci == sci) in nsim_macsec_find_secy() 39 if (!ns->macsec.nsim_secy[idx].used) in nsim_macsec_add_secy() 54 ns->macsec.nsim_secy_count++; in nsim_macsec_add_secy() 93 memset(&ns->macsec.nsim_secy[idx], 0, sizeof(ns->macsec.nsim_secy[idx])); in nsim_macsec_del_secy() 94 ns->macsec.nsim_secy_count--; in nsim_macsec_del_secy() 111 secy = &ns->macsec.nsim_secy[idx]; in nsim_macsec_add_rxsc() 146 secy = &ns->macsec.nsim_secy[idx]; in nsim_macsec_upd_rxsc() 173 secy = &ns->macsec.nsim_secy[idx]; in nsim_macsec_del_rxsc() 204 secy = &ns->macsec.nsim_secy[idx]; in nsim_macsec_add_rxsa() 231 secy = &ns->macsec.nsim_secy[idx]; in nsim_macsec_upd_rxsa() [all …]
|
| A D | bpf.c | 41 struct netdevsim *ns; member 89 return ns->xdp_hw.prog; in nsim_xdp_offload_active() 111 ns->bpf_offloaded = prog; in nsim_bpf_offload() 175 if (ns->bpf_offloaded && !nsim_xdp_offload_active(ns)) in nsim_bpf_disable_tc() 184 if (!nsim_xdp_offload_active(ns) && bpf->prog && ns->bpf_offloaded) { in nsim_xdp_offload_prog() 189 return nsim_bpf_offload(ns, bpf->prog, nsim_xdp_offload_active(ns)); in nsim_xdp_offload_prog() 503 nmap->ns = ns; in nsim_bpf_map_alloc() 561 return nsim_xdp_set_prog(ns, bpf, &ns->xdp); in nsim_bpf() 567 return nsim_xdp_set_prog(ns, bpf, &ns->xdp_hw); in nsim_bpf() 622 ns->netdev); in nsim_bpf_init() [all …]
|
| A D | ipsec.c | 16 struct netdevsim *ns = filp->private_data; in nsim_dbg_netdev_ops_read() local 17 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_dbg_netdev_ops_read() 137 struct netdevsim *ns; in nsim_ipsec_add_sa() local 142 ns = netdev_priv(dev); in nsim_ipsec_add_sa() 143 ipsec = &ns->ipsec; in nsim_ipsec_add_sa() 204 struct netdevsim *ns = netdev_priv(dev); in nsim_ipsec_del_sa() local 205 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_del_sa() 227 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_tx() 271 void nsim_ipsec_init(struct netdevsim *ns) in nsim_ipsec_init() argument 283 ns->nsim_dev_port->ddir, ns, in nsim_ipsec_init() [all …]
|
| /drivers/block/drbd/ |
| A D | drbd_state.c | 857 ns.role == R_PRIMARY && ns.conn < C_CONNECTED && ns.pdsk >= D_UNKNOWN) in is_valid_state() 1078 if (ns.conn == C_STANDALONE && ns.disk == D_DISKLESS && ns.role == R_SECONDARY) in sanitize_state() 1083 if (ns.conn > C_CONNECTED && (ns.disk <= D_FAILED || ns.pdsk <= D_FAILED)) { in sanitize_state() 1194 (ns.role == R_PRIMARY && ns.conn < C_CONNECTED && ns.pdsk > D_OUTDATED) && in sanitize_state() 1203 if (ns.aftr_isp || ns.peer_isp || ns.user_isp) { in sanitize_state() 1276 ns = sanitize_state(device, os, ns, &ssw); in _drbd_set_state() 1474 ascw->ns = ns; in _drbd_set_state() 1684 && (ns < D_INCONSISTENT || ns == D_UNKNOWN || ns == D_OUTDATED)) in lost_contact_to_peer_data() 1694 && (ns == D_DISKLESS || ns == D_FAILED || ns == D_OUTDATED)) in lost_contact_to_peer_data() 1721 sib.ns = ns; in after_state_ch() [all …]
|
| /drivers/nvme/host/ |
| A D | multipath.c | 355 if (ns) in nvme_next_ns() 356 return ns; in nvme_next_ns() 377 ns && ns != old; in nvme_round_robin_path() 378 ns = nvme_next_ns(head, ns)) { in nvme_round_robin_path() 383 found = ns; in nvme_round_robin_path() 387 found = ns; in nvme_round_robin_path() 461 return ns; in nvme_numa_path() 569 if (ns) in nvme_ns_head_get_unique_id() 585 if (ns) in nvme_ns_head_report_zones() 1286 if (blk_queue_is_zoned(ns->queue) && ns->head->disk) in nvme_mpath_add_disk() [all …]
|
| A D | zns.c | 50 dev_warn(ns->ctrl->device, in nvme_query_zone_info() 52 ns->head->ns_id); in nvme_query_zone_info() 55 dev_warn(ns->ctrl->device, in nvme_query_zone_info() 57 ns->head->ns_id); in nvme_query_zone_info() 61 if (!ns->ctrl->max_zone_append) { in nvme_query_zone_info() 85 dev_warn(ns->ctrl->device, in nvme_query_zone_info() 94 dev_warn(ns->ctrl->device, in nvme_query_zone_info() 96 zi->zone_size, ns->head->ns_id); in nvme_query_zone_info() 130 get_capacity(ns->disk) >> ilog2(ns->head->zsze)); in nvme_zns_alloc_report_buffer() 198 sector &= ~(ns->head->zsze - 1); in nvme_ns_report_zones() [all …]
|
| A D | ioctl.c | 43 if (!ns) { in nvme_cmd_allowed() 62 effects = nvme_command_effects(ns->ctrl, ns, c->common.opcode); in nvme_cmd_allowed() 121 struct block_device *bdev = ns ? ns->disk->part0 : NULL; in nvme_map_user_request() 235 (ns->head->ms == ns->head->pi_size)) { in nvme_submit_io() 276 if (ns && nsid != ns->head->ns_id) { in nvme_validate_passthru_nsid() 322 status = nvme_submit_user_cmd(ns ? ns->queue : ctrl->admin_q, &c, in nvme_user_cmd() 448 struct request_queue *q = ns ? ns->queue : ctrl->admin_q; in nvme_uring_cmd_io() 600 return nvme_user_cmd64(ns->ctrl, ns, argp, flags, in nvme_ns_ioctl() 720 if (!ns) in nvme_ns_head_ioctl() 751 if (!ns) in nvme_ns_head_chr_ioctl() [all …]
|
| A D | core.c | 342 if (ns) { in nvme_log_error() 344 ns->disk ? ns->disk->disk_name : "?", in nvme_log_error() 695 kfree(ns); in nvme_free_ns() 1206 if (ns) { in nvme_command_effects() 1271 if (ns) in nvme_passthru_end() 3866 ns->cdev_device.parent = ns->ctrl->device; in nvme_add_ns_cdev() 3868 ns->ctrl->instance, ns->head->instance); in nvme_add_ns_cdev() 4096 list_add_rcu(&ns->list, &ns->ctrl->namespaces); in nvme_ns_add_to_ctrl_list() 4108 if (!ns) in nvme_alloc_ns() 4185 dev_set_drvdata(disk_to_dev(ns->disk), ns); in nvme_alloc_ns() [all …]
|
| /drivers/nvme/target/ |
| A D | io-cmd-bdev.c | 57 if (ns->bdev_file) { in nvmet_bdev_ns_disable() 58 fput(ns->bdev_file); in nvmet_bdev_ns_disable() 59 ns->bdev = NULL; in nvmet_bdev_ns_disable() 91 if (ns->buffered_io) in nvmet_bdev_ns_enable() 94 ns->bdev_file = bdev_file_open_by_path(ns->device_path, in nvmet_bdev_ns_enable() 105 ns->bdev = file_bdev(ns->bdev_file); in nvmet_bdev_ns_enable() 106 ns->size = bdev_nr_bytes(ns->bdev); in nvmet_bdev_ns_enable() 107 ns->blksize_shift = blksize_bits(bdev_logical_block_size(ns->bdev)); in nvmet_bdev_ns_enable() 109 ns->pi_type = 0; in nvmet_bdev_ns_enable() 127 ns->size = bdev_nr_bytes(ns->bdev); in nvmet_bdev_ns_revalidate() [all …]
|
| A D | io-cmd-file.c | 18 ns->size = i_size_read(ns->file->f_mapping->host); in nvmet_file_ns_revalidate() 23 if (ns->file) { in nvmet_file_ns_disable() 24 if (ns->buffered_io) in nvmet_file_ns_disable() 28 fput(ns->file); in nvmet_file_ns_disable() 29 ns->file = NULL; in nvmet_file_ns_disable() 38 if (!ns->buffered_io) in nvmet_file_ns_enable() 41 ns->file = filp_open(ns->device_path, flags, 0); in nvmet_file_ns_enable() 46 ns->file = NULL; in nvmet_file_ns_enable() 69 fput(ns->file); in nvmet_file_ns_enable() 70 ns->file = NULL; in nvmet_file_ns_enable() [all …]
|
| A D | zns.c | 41 if (ns->subsys->zasl) { in nvmet_bdev_zns_enable() 42 if (ns->subsys->zasl > zasl) in nvmet_bdev_zns_enable() 45 ns->subsys->zasl = zasl; in nvmet_bdev_zns_enable() 59 ret = blkdev_report_zones(ns->bdev, 0, bdev_nr_zones(ns->bdev), in nvmet_bdev_zns_enable() 64 ns->blksize_shift = blksize_bits(bdev_logical_block_size(ns->bdev)); in nvmet_bdev_zns_enable() 117 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns() 118 mutex_lock(&req->ns->subsys->lock); in nvmet_execute_identify_ns_zns() 119 nvmet_ns_changed(req->ns->subsys, req->ns->nsid); in nvmet_execute_identify_ns_zns() 123 if (!bdev_is_zoned(req->ns->bdev)) { in nvmet_execute_identify_ns_zns() 130 req->ns->blksize_shift; in nvmet_execute_identify_ns_zns() [all …]
|
| A D | core.c | 440 if (unlikely(!req->ns || !req->ns->enabled)) { in nvmet_req_find_ns() 552 ns->nsid); in nvmet_p2pmem_ns_add_p2p() 559 if (ns->bdev) in nvmet_ns_revalidate() 684 kfree(ns); in nvmet_ns_free() 696 ns = kzalloc(sizeof(*ns), GFP_KERNEL); in nvmet_ns_alloc() 697 if (!ns) in nvmet_ns_alloc() 724 return ns; in nvmet_ns_alloc() 727 kfree(ns); in nvmet_ns_alloc() 777 struct nvmet_ns *ns = req->ns; in __nvmet_req_complete() local 794 if (ns) in __nvmet_req_complete() [all …]
|
| A D | pr.c | 43 struct nvmet_ns *ns; in nvmet_set_feat_resv_notif_mask() local 56 if (!req->ns->pr.enable) in nvmet_set_feat_resv_notif_mask() 64 if (ns->pr.enable) in nvmet_set_feat_resv_notif_mask() 81 if (!req->ns->pr.enable) in nvmet_get_feat_resv_notif_mask() 447 struct nvmet_ns *ns = req->ns; in nvmet_pr_set_ctrl_to_abort() local 595 struct nvmet_ns *ns = req->ns; in nvmet_pr_do_abort() local 610 up(&ns->pr.pr_sem); in nvmet_pr_do_abort() 938 struct nvmet_ns *ns = req->ns; in nvmet_pr_check_cmd_access() local 1060 if (ns->pr.enable) { in nvmet_ctrl_init_pr() 1071 if (ns->pr.enable) { in nvmet_ctrl_init_pr() [all …]
|
| A D | admin-cmd.c | 241 if (!req->ns->bdev) in nvmet_get_smart_log_nsid() 264 struct nvmet_ns *ns; in nvmet_get_smart_log_all() local 271 if (!ns->bdev) in nvmet_get_smart_log_all() 301 if (!req->ns->bdev || bdev_nonrot(req->ns->bdev)) { in nvmet_execute_get_log_page_rmi() 827 nvmet_ns_changed(req->ns->subsys, req->ns->nsid); in nvmet_execute_identify_ns() 836 cpu_to_le64(req->ns->size >> req->ns->blksize_shift); in nvmet_execute_identify_ns() 846 if (req->ns->bdev) in nvmet_execute_identify_ns() 1011 if (memchr_inv(&req->ns->uuid, 0, sizeof(req->ns->uuid))) { in nvmet_execute_identify_desclist() 1018 if (memchr_inv(req->ns->nguid, 0, sizeof(req->ns->nguid))) { in nvmet_execute_identify_desclist() 1087 if (req->ns->bdev && !bdev_nonrot(req->ns->bdev)) in nvmet_execute_id_cs_indep() [all …]
|
| /drivers/clk/qcom/ |
| A D | clk-rcg.c | 22 return ns; in ns_to_src() 41 u32 ns; in clk_rcg_get_parent() local 47 ns = ns_to_src(&rcg->s, ns); in clk_rcg_get_parent() 82 ns = ns_to_src(s, ns); in clk_dyn_rcg_get_parent() 97 u32 ns; in clk_rcg_set_parent() local 151 ns = ~ns >> mn->n_val_shift; in ns_m_to_n() 240 ns = mn_to_ns(mn, f->m, f->n, ns); in configure_bank() 247 ns = mn_to_reg(mn, f->m, f->n, ns); in configure_bank() 501 ns = mn_to_reg(mn, f->m, f->n, ns); in __clk_rcg_set_rate() 503 ns = mn_to_ns(mn, f->m, f->n, ns); in __clk_rcg_set_rate() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | fs_cmd.h | 39 int (*create_flow_table)(struct mlx5_flow_root_namespace *ns, 43 int (*destroy_flow_table)(struct mlx5_flow_root_namespace *ns, 46 int (*modify_flow_table)(struct mlx5_flow_root_namespace *ns, 50 int (*create_flow_group)(struct mlx5_flow_root_namespace *ns, 59 int (*create_fte)(struct mlx5_flow_root_namespace *ns, 64 int (*update_fte)(struct mlx5_flow_root_namespace *ns, 70 int (*delete_fte)(struct mlx5_flow_root_namespace *ns, 74 int (*update_root_ft)(struct mlx5_flow_root_namespace *ns, 95 int (*set_peer)(struct mlx5_flow_root_namespace *ns, 99 int (*create_ns)(struct mlx5_flow_root_namespace *ns); [all …]
|
| A D | fs_cmd.c | 172 struct mlx5_flow_namespace *ns; in mlx5_cmd_set_slave_root_fdb() local 187 ns = mlx5_get_flow_namespace(slave, in mlx5_cmd_set_slave_root_fdb() 189 root = find_root(&ns->node); in mlx5_cmd_set_slave_root_fdb() 216 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_update_root_ft() 258 ns->root_ft->id); in mlx5_cmd_update_root_ft() 265 ns->root_ft->id); in mlx5_cmd_update_root_ft() 287 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_create_flow_table() 343 mlx5_ft_pool_put_sz(ns->dev, size); in mlx5_cmd_create_flow_table() 353 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_destroy_flow_table() 376 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_modify_flow_table() [all …]
|
| /drivers/scsi/bfa/ |
| A D | bfa_fcs_lport.c | 3648 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_offline() 3669 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_plogi_sending() 3692 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_plogi() 3703 &ns->timer, bfa_fcs_lport_ns_timeout, ns, in bfa_fcs_lport_ns_sm_plogi() 3727 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_plogi_retry() 3753 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_sending_rnn_id() 3775 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_rnn_id() 3792 &ns->timer, bfa_fcs_lport_ns_timeout, ns, in bfa_fcs_lport_ns_sm_rnn_id() 3815 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_rnn_id_retry() 3838 bfa_trc(ns->port->fcs, ns->port->port_cfg.pwwn); in bfa_fcs_lport_ns_sm_sending_rsnn_nn() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| A D | fs_dr.c | 41 mlx5_core_err(ns->dev, in set_miss_action() 86 err = set_miss_action(ns, ft, next_ft); in mlx5_cmd_dr_create_flow_table() 133 return set_miss_action(ns, ft, next_ft); in mlx5_cmd_dr_modify_flow_table() 162 mlx5_core_err(ns->dev, "Failed creating matcher\n"); in mlx5_cmd_dr_create_flow_group() 261 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_dr_create_fte() 775 ret = mlx5_cmd_dr_create_fte(ns, ft, group, fte); in mlx5_cmd_dr_update_fte() 779 ret = mlx5_cmd_dr_delete_fte(ns, ft, &fte_tmp); in mlx5_cmd_dr_update_fte() 803 ns->fs_dr_domain.dr_domain = in mlx5_cmd_dr_create_ns() 804 mlx5dr_domain_create(ns->dev, in mlx5_cmd_dr_create_ns() 806 if (!ns->fs_dr_domain.dr_domain) { in mlx5_cmd_dr_create_ns() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | fs_hws.c | 139 ns->fs_hws_context.hws_ctx = in mlx5_cmd_hws_create_ns() 141 if (!ns->fs_hws_context.hws_ctx) { in mlx5_cmd_hws_create_ns() 145 err = mlx5_fs_init_hws_actions_pool(ns->dev, &ns->fs_hws_context); in mlx5_cmd_hws_create_ns() 180 if (!ns->fs_hws_context.hws_ctx) in mlx5_fs_set_ft_default_miss() 262 if (ns->table_type != FS_FT_FDB) { in mlx5_cmd_hws_create_flow_table() 264 ns->table_type); in mlx5_cmd_hws_create_flow_table() 749 struct mlx5_core_dev *dev = ns->dev; in mlx5_fs_fte_get_hws_actions() 1284 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_hws_packet_reformat_alloc() 1375 struct mlx5_core_dev *dev = ns->dev; in mlx5_cmd_hws_packet_reformat_dealloc() 1549 struct mlx5_flow_root_namespace *ns; in mlx5_fs_hws_action_get_pkt_reformat_id() local [all …]
|
| /drivers/ptp/ |
| A D | ptp_dte.c | 73 s64 ns; in dte_read_nco() local 84 return ns; in dte_read_nco() 89 s64 ns; in dte_write_nco_delta() local 91 ns = dte_read_nco(ptp_dte->regs); in dte_write_nco_delta() 96 ns += DTE_NCO_MAX_NS + delta; in dte_write_nco_delta() 99 ns = 0; in dte_write_nco_delta() 102 ns += delta; in dte_write_nco_delta() 103 if (ns > DTE_NCO_MAX_NS) { in dte_write_nco_delta() 105 ns -= DTE_NCO_MAX_NS; in dte_write_nco_delta() 118 s64 ns = 0; in dte_read_nco_with_ovf() local [all …]
|