| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_mr.c | 676 if (mtr->umem) { in mtr_free_bufs() 678 mtr->umem = NULL; in mtr_free_bufs() 682 if (mtr->kmem) { in mtr_free_bufs() 684 mtr->kmem = NULL; in mtr_free_bufs() 698 mtr->kmem = NULL; in mtr_alloc_bufs() 707 mtr->umem = NULL; in mtr_alloc_bufs() 743 return mtr->hem_cfg.is_direct && mtr->hem_cfg.region_count > 1; in need_split_huge_page() 762 if (mtr->umem) in mtr_map_bufs() 972 if (mtr->umem) in get_best_hop_num() 1043 pgoff = mtr->umem ? mtr->umem->address & ~PAGE_MASK : 0; in mtr_init_buf_cfg() [all …]
|
| A D | hns_roce_device.h | 427 struct hns_roce_mtr mtr; member 448 struct hns_roce_mtr mtr; member 622 struct hns_roce_mtr mtr; member 718 struct hns_roce_mtr mtr; member 1180 static inline dma_addr_t hns_roce_get_mtr_ba(struct hns_roce_mtr *mtr) in hns_roce_get_mtr_ba() argument 1182 return mtr->hem_cfg.root_ba; in hns_roce_get_mtr_ba() 1185 int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr, 1187 int hns_roce_mtr_create(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr, 1192 struct hns_roce_mtr *mtr); 1193 int hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
|
| A D | hns_roce_cq.c | 138 ret = hns_roce_mtr_find(hr_dev, &hr_cq->mtr, 0, mtts, ARRAY_SIZE(mtts)); in alloc_cqc() 159 hns_roce_get_mtr_ba(&hr_cq->mtr)); in alloc_cqc() 210 ret = hns_roce_mtr_create(hr_dev, &hr_cq->mtr, &buf_attr, in alloc_cq_buf() 221 hns_roce_mtr_destroy(hr_dev, &hr_cq->mtr); in free_cq_buf()
|
| A D | hns_roce_srq.c | 182 ret = hns_roce_mtr_create(hr_dev, &idx_que->mtr, &buf_attr, in alloc_srq_idx() 205 hns_roce_mtr_destroy(hr_dev, &idx_que->mtr); in alloc_srq_idx() 216 hns_roce_mtr_destroy(hr_dev, &idx_que->mtr); in free_srq_idx()
|
| A D | hns_roce_hw_v2.c | 875 return hns_roce_buf_offset(idx_que->mtr.kmem, in get_idx_buf() 4384 wqe_sge_ba = hns_roce_get_mtr_ba(&hr_qp->mtr); in config_qp_rq_buf() 4460 ret = hns_roce_mtr_find(hr_dev, &hr_qp->mtr, in config_qp_sq_buf() 5694 dma_handle_idx = hns_roce_get_mtr_ba(&idx_que->mtr); in hns_roce_v2_write_srqc_index_queue() 6079 aeqe = hns_roce_buf_offset(eq->mtr.kmem, in next_aeqe_sw_v2() 6164 ceqe = hns_roce_buf_offset(eq->mtr.kmem, in next_ceqe_sw_v2() 6404 hns_roce_mtr_destroy(hr_dev, &eq->mtr); in free_eq_buf() 6451 ret = hns_roce_mtr_find(hr_dev, &eq->mtr, 0, eqe_ba, in config_eqc() 6458 bt_ba = hns_roce_get_mtr_ba(&eq->mtr); in config_eqc() 6468 to_hr_hw_page_shift(eq->mtr.hem_cfg.ba_pg_shift)); in config_eqc() [all …]
|
| A D | hns_roce_qp.c | 760 ret = hns_roce_mtr_create(hr_dev, &hr_qp->mtr, &buf_attr, in alloc_qp_buf() 780 hns_roce_mtr_destroy(hr_dev, &hr_qp->mtr); in free_qp_buf() 1507 return hns_roce_buf_offset(hr_qp->mtr.kmem, offset); in get_wqe()
|
| /linux/drivers/edac/ |
| A D | i5400_edac.c | 286 #define MTR_DIMMS_PRESENT(mtr) ((mtr) & (1 << 10)) argument 287 #define MTR_DIMMS_ETHROTTLE(mtr) ((mtr) & (1 << 9)) argument 288 #define MTR_DRAM_WIDTH(mtr) (((mtr) & (1 << 8)) ? 8 : 4) argument 289 #define MTR_DRAM_BANKS(mtr) (((mtr) & (1 << 6)) ? 8 : 4) argument 291 #define MTR_DIMM_RANK(mtr) (((mtr) >> 5) & 0x1) argument 293 #define MTR_DIMM_ROWS(mtr) (((mtr) >> 2) & 0x3) argument 295 #define MTR_DIMM_COLS(mtr) ((mtr) & 0x3) argument 864 int mtr; in determine_mtr() local 883 return mtr; in determine_mtr() 922 int mtr; in handle_channel() local [all …]
|
| A D | i7300_edac.c | 172 #define MTR_DIMMS_PRESENT(mtr) ((mtr) & (1 << 8)) argument 173 #define MTR_DIMMS_ETHROTTLE(mtr) ((mtr) & (1 << 7)) argument 174 #define MTR_DRAM_WIDTH(mtr) (((mtr) & (1 << 6)) ? 8 : 4) argument 175 #define MTR_DRAM_BANKS(mtr) (((mtr) & (1 << 5)) ? 8 : 4) argument 176 #define MTR_DIMM_RANKS(mtr) (((mtr) & (1 << 4)) ? 1 : 0) argument 177 #define MTR_DIMM_ROWS(mtr) (((mtr) >> 2) & 0x3) argument 179 #define MTR_DIMM_ROWS_ADDR_BITS(mtr) (MTR_DIMM_ROWS(mtr) + 13) argument 180 #define MTR_DIMM_COLS(mtr) ((mtr) & 0x3) argument 594 mtr = pvt->mtr[slot][branch]; in decode_mtr() 673 return mtr; in decode_mtr() [all …]
|
| A D | i5000_edac.c | 279 #define MTR_DIMMS_PRESENT(mtr) ((mtr) & (0x1 << 8)) argument 280 #define MTR_DRAM_WIDTH(mtr) ((((mtr) >> 6) & 0x1) ? 8 : 4) argument 281 #define MTR_DRAM_BANKS(mtr) ((((mtr) >> 5) & 0x1) ? 8 : 4) argument 283 #define MTR_DIMM_RANK(mtr) (((mtr) >> 4) & 0x1) argument 285 #define MTR_DIMM_ROWS(mtr) (((mtr) >> 2) & 0x3) argument 286 #define MTR_DIMM_ROWS_ADDR_BITS(mtr) (MTR_DIMM_ROWS(mtr) + 13) argument 287 #define MTR_DIMM_COLS(mtr) ((mtr) & 0x3) argument 953 int mtr; in determine_mtr() local 960 return mtr; in determine_mtr() 995 int mtr; in handle_channel() local [all …]
|
| A D | sb_edac.c | 215 #define RANK_DISABLE(mtr) GET_BITFIELD(mtr, 16, 19) argument 216 #define IS_DIMM_PRESENT(mtr) GET_BITFIELD(mtr, 14, 14) argument 217 #define RANK_CNT_BITS(mtr) GET_BITFIELD(mtr, 12, 13) argument 218 #define RANK_WIDTH_BITS(mtr) GET_BITFIELD(mtr, 2, 4) argument 219 #define COL_WIDTH_BITS(mtr) GET_BITFIELD(mtr, 0, 1) argument 709 ranks, max, (unsigned int)RANK_CNT_BITS(mtr), mtr); in numrank() 722 rows, (unsigned int)RANK_WIDTH_BITS(mtr), mtr); in numrow() 735 cols, (unsigned int)COL_WIDTH_BITS(mtr), mtr); in numcol() 952 switch (mtr) { in __ibridge_get_width() 1614 u32 mtr, amap = 0; in __populate_dimms() local [all …]
|
| A D | i5100_edac.c | 329 } mtr[I5100_CHANNELS][I5100_MAX_RANKS_PER_CHAN]; member 654 if (!priv->mtr[chan][chan_rank].present) in i5100_npages() 659 priv->mtr[chan][chan_rank].numcol + in i5100_npages() 660 priv->mtr[chan][chan_rank].numrow + in i5100_npages() 661 priv->mtr[chan][chan_rank].numbank; in i5100_npages() 685 priv->mtr[i][j].present = i5100_mtr_present(w); in i5100_init_mtr() 686 priv->mtr[i][j].ethrottle = i5100_mtr_ethrottle(w); in i5100_init_mtr() 687 priv->mtr[i][j].width = 4 + 4 * i5100_mtr_width(w); in i5100_init_mtr() 689 priv->mtr[i][j].numrow = 13 + i5100_mtr_numrow(w); in i5100_init_mtr() 690 priv->mtr[i][j].numcol = 10 + i5100_mtr_numcol(w); in i5100_init_mtr() [all …]
|
| A D | skx_common.c | 229 static int get_width(u32 mtr) in get_width() argument 231 switch (GET_BITFIELD(mtr, 8, 9)) { in get_width() 351 int skx_get_dimm_info(u32 mtr, u32 mcmtr, u32 amap, struct dimm_info *dimm, in skx_get_dimm_info() argument 359 ranks = numrank(mtr); in skx_get_dimm_info() 360 rows = numrow(mtr); in skx_get_dimm_info() 361 cols = imc->hbm_mc ? 6 : numcol(mtr); in skx_get_dimm_info() 392 dimm->dtype = get_width(mtr); in skx_get_dimm_info()
|
| A D | skx_base.c | 180 u32 mtr, mcmtr, amap, mcddrtcfg; in skx_get_dimm_config() local 196 0x80 + 4 * j, &mtr); in skx_get_dimm_config() 197 if (IS_DIMM_PRESENT(mtr)) { in skx_get_dimm_config() 198 ndimms += skx_get_dimm_info(mtr, mcmtr, amap, dimm, imc, i, j, cfg); in skx_get_dimm_config()
|
| A D | i10nm_base.c | 970 u32 mtr, mcddrtcfg = 0; in i10nm_get_dimm_config() local 985 mtr = I10NM_GET_DIMMMTR(imc, i, j); in i10nm_get_dimm_config() 987 mtr, mcddrtcfg, imc->mc, i, j); in i10nm_get_dimm_config() 989 if (IS_DIMM_PRESENT(mtr)) in i10nm_get_dimm_config() 990 ndimms += skx_get_dimm_info(mtr, 0, 0, dimm, in i10nm_get_dimm_config()
|
| A D | skx_common.h | 245 int skx_get_dimm_info(u32 mtr, u32 mcmtr, u32 amap, struct dimm_info *dimm,
|
| /linux/arch/alpha/include/asm/ |
| A D | core_tsunami.h | 36 tsunami_64 mtr; member
|
| A D | core_titan.h | 37 titan_64 mtr; member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | hwmon.c | 60 MLX5_SET(mtmp_reg, mtmp_in, mtr, 1); in mlx5_hwmon_reset_max_temp()
|
| /linux/arch/alpha/kernel/ |
| A D | core_tsunami.c | 397 printk("%s: CSR_MTR 0x%lx\n", __func__, TSUNAMI_cchip.mtr.csr); in tsunami_init_arch()
|
| A D | core_titan.c | 373 printk("%s: CSR_MTR 0x%lx\n", __func__, TITAN_cchip->mtr.csr); in titan_init_arch()
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| A D | reg.h | 9517 MLXSW_ITEM32(reg, mtmp, mtr, 0x08, 30, 1);
|
| /linux/include/linux/mlx5/ |
| A D | mlx5_ifc.h | 11390 u8 mtr[0x1]; member
|