| /drivers/s390/cio/ |
| A D | qdio_debug.c | 63 struct qdio_dbf_entry *new_entry; in qdio_allocate_dbf() local 84 new_entry = kzalloc(sizeof(struct qdio_dbf_entry), GFP_KERNEL); in qdio_allocate_dbf() 85 if (!new_entry) { in qdio_allocate_dbf() 89 strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); in qdio_allocate_dbf() 90 new_entry->dbf_info = irq_ptr->debug_area; in qdio_allocate_dbf() 92 list_add(&new_entry->dbf_list, &qdio_dbf_list); in qdio_allocate_dbf()
|
| /drivers/net/ethernet/mellanox/mlx4/ |
| A D | mcg.c | 155 struct mlx4_steer_index *new_entry; in new_steering_entry() local 165 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in new_steering_entry() 166 if (!new_entry) in new_steering_entry() 169 INIT_LIST_HEAD(&new_entry->duplicates); in new_steering_entry() 170 new_entry->index = index; in new_steering_entry() 233 list_del(&new_entry->list); in new_steering_entry() 234 kfree(new_entry); in new_steering_entry() 1117 u8 new_entry = 0; in mlx4_qp_attach_common() local 1132 new_entry = 1; in mlx4_qp_attach_common() 1146 new_entry = 1; in mlx4_qp_attach_common() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | affinity.c | 588 bool new_entry = false; in hfi1_dev_affinity_init() local 609 new_entry = true; in hfi1_dev_affinity_init() 686 ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); in hfi1_dev_affinity_init() 690 if (new_entry) in hfi1_dev_affinity_init() 699 if (new_entry) in hfi1_dev_affinity_init()
|
| /drivers/scsi/ |
| A D | nsp32.c | 1723 int new_entry; in nsp32_adjust_busfree() local 1739 for (new_entry = old_entry; new_entry < sg_num; new_entry++) { in nsp32_adjust_busfree() 1740 sentlen += (le32_to_cpu(sgt[new_entry].len) & ~SGTEND); in nsp32_adjust_busfree() 1747 if (new_entry == sg_num) { in nsp32_adjust_busfree() 1763 len = le32_to_cpu(sgt[new_entry].len); in nsp32_adjust_busfree() 1764 addr = le32_to_cpu(sgt[new_entry].addr); in nsp32_adjust_busfree() 1766 sgt[new_entry].addr = cpu_to_le32(addr); in nsp32_adjust_busfree() 1767 sgt[new_entry].len = cpu_to_le32(restlen); in nsp32_adjust_busfree() 1770 data->cur_entry = new_entry; in nsp32_adjust_busfree()
|
| A D | hpsa.c | 1356 int entry, struct hpsa_scsi_dev_t *new_entry) in hpsa_scsi_update_entry() argument 1370 if (new_entry->offload_config && new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1379 h->dev[entry]->raid_map = new_entry->raid_map; in hpsa_scsi_update_entry() 1382 if (new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1401 if (!new_entry->offload_to_be_enabled) in hpsa_scsi_update_entry() 1409 int entry, struct hpsa_scsi_dev_t *new_entry, in hpsa_scsi_replace_entry() argument 1422 if (new_entry->target == -1) { in hpsa_scsi_replace_entry() 1423 new_entry->target = h->dev[entry]->target; in hpsa_scsi_replace_entry() 1424 new_entry->lun = h->dev[entry]->lun; in hpsa_scsi_replace_entry() 1427 h->dev[entry] = new_entry; in hpsa_scsi_replace_entry() [all …]
|
| A D | pmcraid.c | 1517 u32 new_entry = 1; in pmcraid_handle_config_change() local 1578 new_entry = 0; in pmcraid_handle_config_change() 1583 if (new_entry) { in pmcraid_handle_config_change()
|
| /drivers/net/ethernet/chelsio/cxgb4vf/ |
| A D | cxgb4vf_main.c | 275 struct hash_mac_addr *new_entry, *entry; in cxgb4vf_change_mac() local 291 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in cxgb4vf_change_mac() 292 if (!new_entry) in cxgb4vf_change_mac() 294 ether_addr_copy(new_entry->addr, addr); in cxgb4vf_change_mac() 295 new_entry->iface_mac = true; in cxgb4vf_change_mac() 296 list_add_tail(&new_entry->list, &adapter->mac_hlist); in cxgb4vf_change_mac() 945 struct hash_mac_addr *new_entry; in cxgb4vf_mac_sync() local 956 new_entry = kzalloc(sizeof(*new_entry), GFP_ATOMIC); in cxgb4vf_mac_sync() 957 if (!new_entry) in cxgb4vf_mac_sync() 959 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4vf_mac_sync() [all …]
|
| /drivers/gpu/nova-core/ |
| A D | firmware.rs | 159 .new_entry() in make_entry_file()
|
| /drivers/iommu/ |
| A D | virtio-iommu.c | 970 struct iommu_resv_region *entry, *new_entry, *msi = NULL; in viommu_get_resv_regions() local 978 new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); in viommu_get_resv_regions() 979 if (!new_entry) in viommu_get_resv_regions() 981 list_add_tail(&new_entry->list, head); in viommu_get_resv_regions()
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_acl_tcam.c | 1193 struct mlxsw_sp_acl_tcam_entry *new_entry; in mlxsw_sp_acl_tcam_ventry_migrate() local 1202 new_entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, chunk); in mlxsw_sp_acl_tcam_ventry_migrate() 1203 if (IS_ERR(new_entry)) in mlxsw_sp_acl_tcam_ventry_migrate() 1204 return PTR_ERR(new_entry); in mlxsw_sp_acl_tcam_ventry_migrate() 1206 ventry->entry = new_entry; in mlxsw_sp_acl_tcam_ventry_migrate()
|
| /drivers/net/ethernet/marvell/octeontx2/af/ |
| A D | rvu_npc.c | 2946 u16 old_entry, new_entry; in rvu_mbox_handler_npc_mcam_shift_entry() local 2960 new_entry = req->new_entry[index]; in rvu_mbox_handler_npc_mcam_shift_entry() 2969 rc = npc_mcam_verify_entry(mcam, pcifunc, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry() 2974 if (mcam->entry2cntr_map[new_entry] != NPC_MCAM_INVALID_MAP) { in rvu_mbox_handler_npc_mcam_shift_entry() 2980 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, false); in rvu_mbox_handler_npc_mcam_shift_entry() 2983 npc_copy_mcam_entry(rvu, mcam, blkaddr, old_entry, new_entry); in rvu_mbox_handler_npc_mcam_shift_entry() 2991 new_entry, cntr); in rvu_mbox_handler_npc_mcam_shift_entry() 2995 npc_enable_mcam_entry(rvu, mcam, blkaddr, new_entry, true); in rvu_mbox_handler_npc_mcam_shift_entry()
|
| A D | mbox.h | 1506 u16 new_entry[NPC_MCAM_MAX_SHIFTS]; member
|
| /drivers/md/dm-vdo/ |
| A D | recovery-journal.c | 1335 struct recovery_journal_entry new_entry; in add_queued_recovery_entries() local 1343 new_entry = (struct recovery_journal_entry) { in add_queued_recovery_entries() 1355 *packed_entry = vdo_pack_recovery_journal_entry(&new_entry); in add_queued_recovery_entries()
|
| /drivers/net/ethernet/chelsio/cxgb4/ |
| A D | cxgb4_main.c | 380 struct hash_mac_addr *new_entry; in cxgb4_mac_sync() local 391 new_entry = kzalloc(sizeof(*new_entry), GFP_ATOMIC); in cxgb4_mac_sync() 392 if (!new_entry) in cxgb4_mac_sync() 394 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4_mac_sync() 395 list_add_tail(&new_entry->list, &adap->mac_hlist); in cxgb4_mac_sync() 465 struct hash_mac_addr *entry, *new_entry; in cxgb4_change_mac() local 481 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in cxgb4_change_mac() 482 if (!new_entry) in cxgb4_change_mac() 484 ether_addr_copy(new_entry->addr, addr); in cxgb4_change_mac() 485 new_entry->iface_mac = true; in cxgb4_change_mac() [all …]
|
| /drivers/net/wireless/mediatek/mt76/mt7615/ |
| A D | mcu.c | 849 bool new_entry = true; in mt7615_mcu_wtbl_sta_add() local 863 new_entry = false; in mt7615_mcu_wtbl_sta_add() 869 link_sta, conn_state, new_entry); in mt7615_mcu_wtbl_sta_add()
|
| /drivers/s390/net/ |
| A D | qeth_core_main.c | 6246 struct qeth_dbf_entry *new_entry; in qeth_add_dbf_entry() local 6255 new_entry = kzalloc(sizeof(struct qeth_dbf_entry), GFP_KERNEL); in qeth_add_dbf_entry() 6256 if (!new_entry) in qeth_add_dbf_entry() 6258 strscpy(new_entry->dbf_name, name, sizeof(new_entry->dbf_name)); in qeth_add_dbf_entry() 6259 new_entry->dbf_info = card->debug; in qeth_add_dbf_entry() 6261 list_add(&new_entry->dbf_list, &qeth_dbf_list); in qeth_add_dbf_entry()
|
| /drivers/net/dsa/ |
| A D | mt7530.c | 1952 struct mt7530_hw_vlan_entry new_entry; in mt7530_port_vlan_add() local 1957 mt7530_hw_vlan_entry_init(&new_entry, port, untagged); in mt7530_port_vlan_add() 1958 mt7530_hw_vlan_update(priv, vlan->vid, &new_entry, mt7530_hw_vlan_add); in mt7530_port_vlan_add()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | gfx_v8_0.c | 3875 bool new_entry = true; in gfx_v8_0_parse_ind_reg_list() local 3879 if (new_entry) { in gfx_v8_0_parse_ind_reg_list() 3880 new_entry = false; in gfx_v8_0_parse_ind_reg_list() 3887 new_entry = true; in gfx_v8_0_parse_ind_reg_list()
|