| /linux/drivers/hwmon/ |
| A D | dme1737.c | 74 #define DME1737_REG_IN(ix) ((ix) < 5 ? 0x20 + (ix) : \ argument 115 #define DME1737_REG_PWM(ix) ((ix) < 3 ? 0x30 + (ix) \ argument 617 for (ix = 0; ix < ARRAY_SIZE(data->in); ix++) { in dme1737_update_device() 634 for (ix = 0; ix < ARRAY_SIZE(data->temp); ix++) { in dme1737_update_device() 660 for (ix = 0; ix < ARRAY_SIZE(lsb); ix++) { in dme1737_update_device() 666 for (ix = 0; ix < ARRAY_SIZE(data->in); ix++) { in dme1737_update_device() 672 for (ix = 0; ix < ARRAY_SIZE(data->temp); ix++) { in dme1737_update_device() 678 for (ix = 0; ix < ARRAY_SIZE(data->fan); ix++) { in dme1737_update_device() 703 for (ix = 0; ix < ARRAY_SIZE(data->pwm); ix++) { in dme1737_update_device() 2214 for (ix = 0; ix < 3; ix++) { in dme1737_create_files() [all …]
|
| A D | vt1211.c | 62 #define VT1211_REG_IN(ix) (0x21 + (ix)) argument 63 #define VT1211_REG_IN_MIN(ix) ((ix) == 0 ? 0x3e : 0x2a + 2 * (ix)) argument 64 #define VT1211_REG_IN_MAX(ix) ((ix) == 0 ? 0x3d : 0x29 + 2 * (ix)) argument 72 #define VT1211_REG_FAN(ix) (0x29 + (ix)) argument 78 #define VT1211_REG_PWM(ix) (0x60 + (ix)) argument 254 for (ix = 0; ix < ARRAY_SIZE(data->in); ix++) { in vt1211_update_device() 266 for (ix = 0; ix < ARRAY_SIZE(data->temp); ix++) { in vt1211_update_device() 278 for (ix = 0; ix < ARRAY_SIZE(data->fan); ix++) { in vt1211_update_device() 306 for (ix = 0; ix < ARRAY_SIZE(data->pwm_auto_temp); ix++) { in vt1211_update_device() 346 res = IN_FROM_REG(ix, data->in[ix]); in show_in() [all …]
|
| /linux/arch/sh/kernel/cpu/sh2a/ |
| A D | fpu.c | 120 ix = 0; in denormal_mulf() 262 ix = denormal_subf1(ix, iy); in denormal_addf() 264 ix = -ix; in denormal_addf() 268 ix = denormal_subf1(iy, ix); in denormal_addf() 276 ix = denormal_addf1(ix, iy); in denormal_addf() 278 ix = denormal_addf1(iy, ix); in denormal_addf() 347 ix = denormal_subd1(ix, iy); in denormal_addd() 349 ix = -ix; in denormal_addd() 353 ix = denormal_subd1(iy, ix); in denormal_addd() 361 ix = denormal_addd1(ix, iy); in denormal_addd() [all …]
|
| /linux/io_uring/ |
| A D | xattr.c | 29 if (ix->filename) in io_xattr_cleanup() 30 putname(ix->filename); in io_xattr_cleanup() 32 kfree(ix->ctx.kname); in io_xattr_cleanup() 33 kvfree(ix->ctx.kvalue); in io_xattr_cleanup() 54 ix->filename = NULL; in __io_getxattr_prep() 61 if (ix->ctx.flags) in __io_getxattr_prep() 64 ix->ctx.kname = kmalloc(sizeof(*ix->ctx.kname), GFP_KERNEL); in __io_getxattr_prep() 65 if (!ix->ctx.kname) in __io_getxattr_prep() 117 &ix->ctx); in io_fgetxattr() 165 ix->ctx.kname = kmalloc(sizeof(*ix->ctx.kname), GFP_KERNEL); in __io_setxattr_prep() [all …]
|
| /linux/arch/mips/math-emu/ |
| A D | sp_sqrt.c | 56 ix = x.bits; in ieee754sp_sqrt() 59 m = (ix >> 23); in ieee754sp_sqrt() 62 ix <<= 1; in ieee754sp_sqrt() 66 ix = (ix & 0x007fffff) | 0x00800000; in ieee754sp_sqrt() 68 ix += ix; in ieee754sp_sqrt() 72 ix += ix; in ieee754sp_sqrt() 79 if (t <= ix) { in ieee754sp_sqrt() 81 ix -= t; in ieee754sp_sqrt() 84 ix += ix; in ieee754sp_sqrt() 88 if (ix != 0) { in ieee754sp_sqrt() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | rx_res.c | 335 int ix; in mlx5e_rx_res_channels_init() local 347 for (ix = 0; ix < res->max_nch; ix++) { in mlx5e_rx_res_channels_init() 358 for (ix = 0; ix < res->max_nch; ix++) { in mlx5e_rx_res_channels_init() 431 unsigned int ix; in mlx5e_rx_res_channels_destroy() local 433 for (ix = 0; ix < res->max_nch; ix++) { in mlx5e_rx_res_channels_destroy() 564 for (ix = 0; ix < chs->num; ix++) { in mlx5e_rx_res_channels_activate() 576 for (ix = 0; ix < nch; ix++) in mlx5e_rx_res_channels_activate() 578 for (ix = nch; ix < res->max_nch; ix++) in mlx5e_rx_res_channels_activate() 602 for (ix = 0; ix < res->max_nch; ix++) in mlx5e_rx_res_channels_deactivate() 647 for (ix = 0; ix < MLX5E_MAX_NUM_RSS; ix++) { in mlx5e_rx_res_packet_merge_set_param() [all …]
|
| A D | channels.c | 14 static struct mlx5e_channel *mlx5e_channels_get(struct mlx5e_channels *chs, unsigned int ix) in mlx5e_channels_get() argument 16 WARN_ON_ONCE(ix >= mlx5e_channels_get_num(chs)); in mlx5e_channels_get() 17 return chs->c[ix]; in mlx5e_channels_get() 20 bool mlx5e_channels_is_xsk(struct mlx5e_channels *chs, unsigned int ix) in mlx5e_channels_is_xsk() argument 22 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_is_xsk() 27 void mlx5e_channels_get_regular_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn, in mlx5e_channels_get_regular_rqn() argument 30 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_get_regular_rqn() 37 void mlx5e_channels_get_xsk_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn, in mlx5e_channels_get_xsk_rqn() argument 40 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_get_xsk_rqn()
|
| A D | channels.h | 12 bool mlx5e_channels_is_xsk(struct mlx5e_channels *chs, unsigned int ix); 13 void mlx5e_channels_get_regular_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn, 15 void mlx5e_channels_get_xsk_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn,
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
| A D | pool.c | 54 xsk->pools[ix] = pool; in mlx5e_xsk_add_pool() 60 xsk->pools[ix] = NULL; in mlx5e_xsk_remove_pool() 79 struct xsk_buff_pool *pool, u16 ix) in mlx5e_xsk_enable_locked() argument 96 err = mlx5e_xsk_add_pool(&priv->xsk, pool, ix); in mlx5e_xsk_enable_locked() 123 c = priv->channels.c[ix]; in mlx5e_xsk_enable_locked() 144 mlx5e_xsk_remove_pool(&priv->xsk, ix); in mlx5e_xsk_enable_locked() 166 &priv->xsk, ix); in mlx5e_xsk_disable_locked() 179 c = priv->channels.c[ix]; in mlx5e_xsk_disable_locked() 191 mlx5e_xsk_remove_pool(&priv->xsk, ix); in mlx5e_xsk_disable_locked() 198 u16 ix) in mlx5e_xsk_enable_pool() argument [all …]
|
| A D | pool.h | 10 struct mlx5e_xsk *xsk, u16 ix) in mlx5e_xsk_get_pool() argument 15 if (unlikely(ix >= params->num_channels)) in mlx5e_xsk_get_pool() 18 return xsk->pools[ix]; in mlx5e_xsk_get_pool()
|
| A D | rx.c | 19 int mlx5e_xsk_alloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix) in mlx5e_xsk_alloc_rx_mpwqe() argument 21 struct mlx5e_mpw_info *wi = mlx5e_get_mpw_info(rq, ix); in mlx5e_xsk_alloc_rx_mpwqe() 130 offset = ix * rq->mpwqe.mtts_per_wqe; in mlx5e_xsk_alloc_rx_mpwqe() 160 int mlx5e_xsk_alloc_rx_wqes_batched(struct mlx5e_rq *rq, u16 ix, int wqe_bulk) in mlx5e_xsk_alloc_rx_wqes_batched() argument 171 contig = mlx5_wq_cyc_get_size(wq) - ix; in mlx5e_xsk_alloc_rx_wqes_batched() 173 alloc = xsk_buff_alloc_batch(rq->xsk_pool, buffs + ix, wqe_bulk); in mlx5e_xsk_alloc_rx_wqes_batched() 175 alloc = xsk_buff_alloc_batch(rq->xsk_pool, buffs + ix, contig); in mlx5e_xsk_alloc_rx_wqes_batched() 181 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes_batched() 198 int mlx5e_xsk_alloc_rx_wqes(struct mlx5e_rq *rq, u16 ix, int wqe_bulk) in mlx5e_xsk_alloc_rx_wqes() argument 204 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes()
|
| A D | rx.h | 11 int mlx5e_xsk_alloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix); 12 int mlx5e_xsk_alloc_rx_wqes_batched(struct mlx5e_rq *rq, u16 ix, int wqe_bulk); 13 int mlx5e_xsk_alloc_rx_wqes(struct mlx5e_rq *rq, u16 ix, int wqe_bulk);
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | wq.h | 82 void mlx5_wq_cyc_wqe_dump(struct mlx5_wq_cyc *wq, u16 ix, u8 nstrides); 157 static inline void *mlx5_wq_cyc_get_wqe(struct mlx5_wq_cyc *wq, u16 ix) in mlx5_wq_cyc_get_wqe() argument 159 return mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_wq_cyc_get_wqe() 164 return mlx5_frag_buf_get_idx_last_contig_stride(&wq->fbc, ix) - ix + 1; in mlx5_wq_cyc_get_contig_wqebbs() 202 struct mlx5_cqe64 *cqe = mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_cqwq_get_wqe() 283 static inline void *mlx5_wq_ll_get_wqe(struct mlx5_wq_ll *wq, u16 ix) in mlx5_wq_ll_get_wqe() argument 285 return mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_wq_ll_get_wqe() 288 static inline u16 mlx5_wq_ll_get_wqe_next_ix(struct mlx5_wq_ll *wq, u16 ix) in mlx5_wq_ll_get_wqe_next_ix() argument 290 struct mlx5_wqe_srq_next_seg *wqe = mlx5_wq_ll_get_wqe(wq, ix); in mlx5_wq_ll_get_wqe_next_ix() 302 static inline void mlx5_wq_ll_pop(struct mlx5_wq_ll *wq, __be16 ix, in mlx5_wq_ll_pop() argument [all …]
|
| /linux/fs/qnx4/ |
| A D | dir.c | 24 int ix, ino; in qnx4_readdir() local 37 ix = (ctx->pos >> QNX4_DIR_ENTRY_SIZE_BITS) % QNX4_INODES_PER_BLOCK; in qnx4_readdir() 38 for (; ix < QNX4_INODES_PER_BLOCK; ix++, ctx->pos += QNX4_DIR_ENTRY_SIZE) { in qnx4_readdir() 42 offset = ix * QNX4_DIR_ENTRY_SIZE; in qnx4_readdir() 50 ino = blknum * QNX4_INODES_PER_BLOCK + ix - 1; in qnx4_readdir()
|
| /linux/arch/s390/lib/ |
| A D | spinlock.c | 63 int ix; in arch_spin_lock_setup() local 66 for (ix = 0; ix < 4; ix++, node++) { in arch_spin_lock_setup() 69 (ix << _Q_TAIL_IDX_OFFSET); in arch_spin_lock_setup() 99 int ix, cpu; in arch_spin_decode_tail() local 101 ix = (lock & _Q_TAIL_IDX_MASK) >> _Q_TAIL_IDX_OFFSET; in arch_spin_decode_tail() 103 return per_cpu_ptr(&spin_wait[ix], cpu - 1); in arch_spin_decode_tail() 120 int lockval, ix, node_id, tail_id, old, new, owner, count; in arch_spin_lock_queued() local 122 ix = get_lowcore()->spinlock_index++; in arch_spin_lock_queued() 125 node = this_cpu_ptr(&spin_wait[ix]); in arch_spin_lock_queued()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| A D | fs_ttc.c | 374 int ix = 0; in mlx5_create_ttc_table_groups() local 399 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_groups() 412 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_groups() 414 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5_create_ttc_table_groups() 422 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_groups() 432 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_groups() 538 int ix = 0; in mlx5_create_inner_ttc_table_groups() local 560 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_inner_ttc_table_groups() 573 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_inner_ttc_table_groups() 583 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_inner_ttc_table_groups() [all …]
|
| A D | mpfs.h | 53 int ix = MLX5_L2_ADDR_HASH(mac); \ 57 hlist_for_each_entry(ptr, &(hash)[ix], node.hlist) \ 68 int ix = MLX5_L2_ADDR_HASH(mac); \ 74 hlist_add_head(&ptr->node.hlist, &(hash)[ix]);\
|
| A D | mpfs.c | 79 static int alloc_l2table_index(struct mlx5_mpfs *l2table, u32 *ix) in alloc_l2table_index() argument 83 *ix = find_first_zero_bit(l2table->bitmap, l2table->size); in alloc_l2table_index() 84 if (*ix >= l2table->size) in alloc_l2table_index() 87 __set_bit(*ix, l2table->bitmap); in alloc_l2table_index() 92 static void free_l2table_index(struct mlx5_mpfs *l2table, u32 ix) in free_l2table_index() argument 94 __clear_bit(ix, l2table->bitmap); in free_l2table_index()
|
| A D | ipsec_fs_roce.c | 343 int ix = 0; in ipsec_fs_roce_tx_mpv_create_group_rules() local 351 MLX5_SET_CFG(in, start_flow_index, ix); in ipsec_fs_roce_tx_mpv_create_group_rules() 352 ix += MLX5_TX_ROCE_GROUP_SIZE; in ipsec_fs_roce_tx_mpv_create_group_rules() 448 int ix = 0; in ipsec_fs_roce_rx_mpv_create() local 506 MLX5_SET_CFG(in, start_flow_index, ix); in ipsec_fs_roce_rx_mpv_create() 507 ix += 1; in ipsec_fs_roce_rx_mpv_create() 605 int ix = 0; in mlx5_ipsec_fs_roce_tx_create() local 636 ix += MLX5_TX_ROCE_GROUP_SIZE; in mlx5_ipsec_fs_roce_tx_create() 716 int ix = 0; in mlx5_ipsec_fs_roce_rx_create() local 752 ix += MLX5_RX_ROCE_GROUP_SIZE; in mlx5_ipsec_fs_roce_rx_create() [all …]
|
| /linux/drivers/media/dvb-frontends/ |
| A D | mxl692.c | 203 for (ix = 0; ix < div_size; ix++) in mxl692_checksum() 215 u32 ix, temp; in mxl692_validate_fw_header() local 235 for (ix = 16; ix < buf_len; ix++) in mxl692_validate_fw_header() 255 ix = *index; in mxl692_write_fw_block() 258 total_len = buffer[ix + 1] << 16 | buffer[ix + 2] << 8 | buffer[ix + 3]; in mxl692_write_fw_block() 260 addr = buffer[ix + 4] << 24 | buffer[ix + 5] << 16 | in mxl692_write_fw_block() 261 buffer[ix + 6] << 8 | buffer[ix + 7]; in mxl692_write_fw_block() 289 *index = ix; in mxl692_write_fw_block() 416 u32 ix = 0; in mxl692_opread() local 426 for (ix = 0; ix < size; ix += 4) { in mxl692_opread() [all …]
|
| /linux/drivers/input/misc/ |
| A D | yealink.c | 284 int ix, len; in yealink_set_ringtone() local 300 ix = 0; in yealink_set_ringtone() 301 while (size != ix) { in yealink_set_ringtone() 302 len = size - ix; in yealink_set_ringtone() 309 ix += len; in yealink_set_ringtone() 319 int i, ix, len; in yealink_do_idle_tasks() local 321 ix = yld->stat_ix; in yealink_do_idle_tasks() 351 yld->copy.b[ix] = val; in yealink_do_idle_tasks() 354 switch(ix) { in yealink_do_idle_tasks() 387 ix++; in yealink_do_idle_tasks() [all …]
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
| A D | smu_helper.h | 155 PHM_GET_FIELD(cgs_read_ind_register(device, port, ix##reg), \ 159 PHM_GET_FIELD(cgs_read_ind_register(device, port, ix##reg), \ 167 cgs_write_ind_register(device, port, ix##reg, \ 168 PHM_SET_FIELD(cgs_read_ind_register(device, port, ix##reg), \ 172 cgs_write_ind_register(device, port, ix##reg, \ 173 PHM_SET_FIELD(cgs_read_ind_register(device, port, ix##reg), \ 181 PHM_WAIT_INDIRECT_REGISTER_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask) 192 PHM_WAIT_INDIRECT_REGISTER_UNEQUAL_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask) 206 PHM_WAIT_VFPF_INDIRECT_REGISTER_UNEQUAL_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask) 220 PHM_WAIT_VFPF_INDIRECT_REGISTER_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask)
|
| /linux/fs/ext4/ |
| A D | ext4_extents.h | 239 static inline ext4_fsblk_t ext4_idx_pblock(struct ext4_extent_idx *ix) in ext4_idx_pblock() argument 243 block = le32_to_cpu(ix->ei_leaf_lo); in ext4_idx_pblock() 244 block |= ((ext4_fsblk_t) le16_to_cpu(ix->ei_leaf_hi) << 31) << 1; in ext4_idx_pblock() 266 static inline void ext4_idx_store_pblock(struct ext4_extent_idx *ix, in ext4_idx_store_pblock() argument 269 ix->ei_leaf_lo = cpu_to_le32((unsigned long) (pb & 0xffffffff)); in ext4_idx_store_pblock() 270 ix->ei_leaf_hi = cpu_to_le16((unsigned long) ((pb >> 31) >> 1) & in ext4_idx_store_pblock()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
| A D | conn.c | 103 unsigned int ix; in mlx5_fpga_conn_post_recv() local 122 conn->qp.rq.bufs[ix] = buf; in mlx5_fpga_conn_post_recv() 146 unsigned int ix, sgi; in mlx5_fpga_conn_post_send() local 171 conn->qp.sq.bufs[ix] = buf; in mlx5_fpga_conn_post_send() 254 int ix, err; in mlx5_fpga_conn_rq_cqe() local 257 buf = conn->qp.rq.bufs[ix]; in mlx5_fpga_conn_rq_cqe() 295 int ix; in mlx5_fpga_conn_sq_cqe() local 607 int ix; in mlx5_fpga_conn_free_recv_bufs() local 609 for (ix = 0; ix < conn->qp.rq.size; ix++) { in mlx5_fpga_conn_free_recv_bufs() 621 int ix; in mlx5_fpga_conn_flush_send_bufs() local [all …]
|
| /linux/fs/netfs/ |
| A D | iterator.c | 113 unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0; in netfs_limit_bvec() local 122 while (n && ix < nbv && skip) { in netfs_limit_bvec() 123 len = bvecs[ix].bv_len; in netfs_limit_bvec() 128 ix++; in netfs_limit_bvec() 131 while (n && ix < nbv) { in netfs_limit_bvec() 132 len = min3(n, bvecs[ix].bv_len - skip, max_size); in netfs_limit_bvec() 135 ix++; in netfs_limit_bvec()
|