| /drivers/infiniband/hw/mlx5/ |
| A D | wr.c | 85 *seg += stride; in set_eth_seg() 195 memset(seg, 0, sizeof(*seg)); in set_reg_mkey_seg() 213 memset(seg, 0, sizeof(*seg)); in set_linv_mkey_seg() 268 seg = *wqe; in set_data_inl_seg() 516 bsf = *seg; in set_sig_data_segment() 535 memset(seg, 0, sizeof(*seg)); in set_sig_mkey_segment() 734 *ctrl = *seg; in mlx5r_begin_wqe() 780 seg = PTR_ALIGN(seg, MLX5_SEND_WQE_BB); in mlx5r_finish_wqe() 1016 pad = *seg; in handle_qpt_ud() 1065 void *seg; in mlx5_ib_post_send() local [all …]
|
| A D | wr.h | 51 static inline void handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg, in handle_post_send_edge() argument 56 if (likely(*seg != *cur_edge)) in handle_post_send_edge() 62 *seg = mlx5_frag_buf_get_wqe(&sq->fbc, idx); in handle_post_send_edge() 75 void **seg, u32 *wqe_sz, in mlx5r_memcpy_send_wqe() argument 79 size_t leftlen = *cur_edge - *seg; in mlx5r_memcpy_send_wqe() 83 memcpy(*seg, src, copysz); in mlx5r_memcpy_send_wqe() 88 *seg += stride; in mlx5r_memcpy_send_wqe() 90 handle_post_send_edge(sq, seg, *wqe_sz, cur_edge); in mlx5r_memcpy_send_wqe() 95 int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg, 100 void *seg, u8 size, void *cur_edge, unsigned int idx,
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| A D | dr_buddy.c | 77 unsigned int seg, order_iter, m; in dr_buddy_find_free_seg() local 87 if (WARN(seg >= m, in dr_buddy_find_free_seg() 98 *segment = seg; in dr_buddy_find_free_seg() 123 unsigned int seg, order_iter; in mlx5dr_buddy_alloc_mem() local 139 seg <<= 1; in mlx5dr_buddy_alloc_mem() 144 seg <<= order; in mlx5dr_buddy_alloc_mem() 145 *segment = seg; in mlx5dr_buddy_alloc_mem() 151 unsigned int seg, unsigned int order) in mlx5dr_buddy_free_mem() argument 153 seg >>= order; in mlx5dr_buddy_free_mem() 161 seg >>= 1; in mlx5dr_buddy_free_mem() [all …]
|
| A D | dr_icm_pool.c | 13 unsigned int seg; member 84 return (u64)offset * chunk->seg; in mlx5dr_icm_pool_get_chunk_mr_addr() 325 unsigned int seg) in dr_icm_chunk_init() argument 329 chunk->seg = seg; in dr_icm_chunk_init() 355 hot_chunk->seg, ilog2(num_entries)); in dr_icm_pool_clear_hot_chunks_arr() 389 unsigned int *seg) in dr_icm_handle_buddies_get_mem() argument 399 chunk_size, seg); in dr_icm_handle_buddies_get_mem() 440 unsigned int seg; in mlx5dr_icm_alloc_chunk() local 456 dr_icm_chunk_init(chunk, pool, chunk_size, buddy, seg); in mlx5dr_icm_alloc_chunk() 461 mlx5dr_buddy_free_mem(buddy, seg, chunk_size); in mlx5dr_icm_alloc_chunk() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | buddy.c | 87 unsigned int seg, order_iter, m; in hws_buddy_find_free_seg() local 97 if (WARN(seg >= m, in hws_buddy_find_free_seg() 108 *segment = seg; in hws_buddy_find_free_seg() 115 u32 seg, order_iter, err; in mlx5hws_buddy_alloc_mem() local 121 bitmap_clear(buddy->bitmap[order_iter], seg, 1); in mlx5hws_buddy_alloc_mem() 126 seg <<= 1; in mlx5hws_buddy_alloc_mem() 131 seg <<= order; in mlx5hws_buddy_alloc_mem() 133 return seg; in mlx5hws_buddy_alloc_mem() 138 seg >>= order; in mlx5hws_buddy_free_mem() 143 seg >>= 1; in mlx5hws_buddy_free_mem() [all …]
|
| /drivers/firmware/google/ |
| A D | memconsole-coreboot.c | 45 struct seg { /* describes ring buffer segments in logical order */ in memconsole_coreboot_read() struct 48 } seg[2] = { {0}, {0} }; in memconsole_coreboot_read() local 55 seg[0] = (struct seg){.phys = cursor, .len = size - cursor}; in memconsole_coreboot_read() 56 seg[1] = (struct seg){.phys = 0, .len = cursor}; in memconsole_coreboot_read() 58 seg[0] = (struct seg){.phys = 0, .len = min(cursor, size)}; in memconsole_coreboot_read() 61 for (i = 0; i < ARRAY_SIZE(seg) && count > done; i++) { in memconsole_coreboot_read() 63 cbmem_console->body + seg[i].phys, seg[i].len); in memconsole_coreboot_read() 64 pos -= seg[i].len; in memconsole_coreboot_read()
|
| /drivers/acpi/ |
| A D | pci_mcfg.c | 46 #define AL_ECAM(table_id, rev, seg, ops) \ argument 47 { "AMAZON", table_id, rev, seg, MCFG_BUS_ANY, ops } 58 #define QCOM_ECAM32(seg) \ argument 70 #define HISI_QUAD_DOM(table_id, seg, ops) \ argument 100 #define THUNDER_ECAM_QUIRK(rev, seg) \ argument 128 #define XGENE_V1_ECAM_MCFG(rev, seg) \ argument 129 {"APM ", "XGENE ", rev, seg, MCFG_BUS_ANY, \ 132 #define XGENE_V2_ECAM_MCFG(rev, seg) \ argument 155 #define ALTRA_ECAM_QUIRK(rev, seg) \ argument 177 #define LOONGSON_ECAM_MCFG(table_id, seg) \ argument [all …]
|
| /drivers/usb/cdns3/ |
| A D | cdnsp-mem.c | 38 seg = kzalloc(sizeof(*seg), flags); in cdnsp_segment_alloc() 39 if (!seg) in cdnsp_segment_alloc() 44 kfree(seg); in cdnsp_segment_alloc() 62 return seg; in cdnsp_segment_alloc() 66 kfree(seg); in cdnsp_segment_alloc() 75 dma_pool_free(pdev->segment_pool, seg->trbs, seg->dma); in cdnsp_segment_free() 78 kfree(seg); in cdnsp_segment_free() 245 seg = seg->next; in cdnsp_update_stream_segment_mapping() 257 seg = seg->next; in cdnsp_update_stream_segment_mapping() 270 seg = seg->next; in cdnsp_remove_stream_mapping() [all …]
|
| A D | cdnsp-ring.c | 101 return cdnsp_last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg); in cdnsp_last_trb_on_ring() 135 *seg = (*seg)->next; in cdnsp_next_trb() 136 *trb = ((*seg)->trbs); in cdnsp_next_trb() 606 if (!seg) in cdnsp_unmap_td_bounce_buffer() 611 trace_cdnsp_bounce_unmap(td->preq, seg->bounce_len, seg->bounce_offs, in cdnsp_unmap_td_bounce_buffer() 612 seg->bounce_dma, 0); in cdnsp_unmap_td_bounce_buffer() 625 seg->bounce_buf, seg->bounce_len, in cdnsp_unmap_td_bounce_buffer() 631 seg->bounce_len = 0; in cdnsp_unmap_td_bounce_buffer() 632 seg->bounce_offs = 0; in cdnsp_unmap_td_bounce_buffer() 1836 seg->bounce_dma = dma_map_single(dev, seg->bounce_buf, in cdnsp_align_td() [all …]
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_ethtool_fdir.c | 666 if (!memcmp(old_seg, seg, sizeof(*seg))) in ice_fdir_set_hw_fltr_rule() 697 seg, &entry1_h); in ice_fdir_set_hw_fltr_rule() 702 seg, &entry2_h); in ice_fdir_set_hw_fltr_rule() 797 if (!seg) in ice_set_init_fdir_seg() 861 seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL); in ice_create_init_fdir_rule() 862 if (!seg) in ice_create_init_fdir_rule() 896 memcpy(&tun_seg[1], seg, sizeof(*seg)); in ice_create_init_fdir_rule() 1339 seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL); in ice_cfg_fdir_xtrct_seq() 1340 if (!seg) in ice_cfg_fdir_xtrct_seq() 1406 memcpy(&tun_seg[1], seg, sizeof(*seg)); in ice_cfg_fdir_xtrct_seq() [all …]
|
| A D | ice_ddp.c | 90 hw->seg = NULL; in ice_free_seg() 381 if (!hw->seg) in ice_get_prof_index_max() 1365 if (!seg) in ice_is_signing_seg_at_idx() 1391 if (seg && le32_to_cpu(seg->seg_id) == seg_id && in ice_is_signing_seg_type_at_idx() 1430 if (!seg) in ice_download_pkg_config_seg() 1465 if (!seg) { in ice_dwnld_sign_and_cfg_segs() 2061 if (!*seg) { in ice_chk_pkg_compat() 2155 struct ice_seg *seg = hw->seg; in ice_fill_hw_ptype() local 2159 if (!seg) in ice_fill_hw_ptype() 2256 hw->seg = seg; in ice_init_pkg() [all …]
|
| A D | ice_flow.c | 1124 u8 seg) in ice_flow_xtract_raws() argument 1836 seg->match |= bit; in ice_flow_set_fld_ext() 1838 seg->range |= bit; in ice_flow_set_fld_ext() 1900 seg->raws[seg->raws_cnt].off = off; in ice_flow_add_fld_raw() 1901 seg->raws[seg->raws_cnt].info.type = ICE_FLOW_FLD_TYPE_SIZE; in ice_flow_add_fld_raw() 1902 seg->raws[seg->raws_cnt].info.src.val = val_loc; in ice_flow_add_fld_raw() 1903 seg->raws[seg->raws_cnt].info.src.mask = mask_loc; in ice_flow_add_fld_raw() 1905 seg->raws[seg->raws_cnt].info.src.last = len; in ice_flow_add_fld_raw() 1911 seg->raws_cnt++; in ice_flow_add_fld_raw() 1993 struct ice_flow_seg_info *seg; in ice_flow_set_rss_seg_info() local [all …]
|
| /drivers/isdn/mISDN/ |
| A D | dsp_audio.c | 48 int seg; in linear2alaw() local 65 for (seg = 0; seg < 8; seg++) { in linear2alaw() 66 if (pcm_val <= seg_end[seg]) in linear2alaw() 70 return ((seg << 4) | in linear2alaw() 71 ((pcm_val >> ((seg) ? (seg + 3) : 4)) & 0x0F)) ^ mask; in linear2alaw() 78 int seg; in alaw2linear() local 82 seg = (((int) alaw & 0x70) >> 4); in alaw2linear() 83 if (seg) in alaw2linear() 84 i = (i + 0x100) << (seg - 1); in alaw2linear()
|
| /drivers/usb/host/ |
| A D | xhci-mem.c | 38 seg = kzalloc_node(sizeof(*seg), flags, dev_to_node(dev)); in xhci_segment_alloc() 39 if (!seg) in xhci_segment_alloc() 44 kfree(seg); in xhci_segment_alloc() 61 return seg; in xhci_segment_alloc() 67 dma_pool_free(xhci->segment_pool, seg->trbs, seg->dma); in xhci_segment_free() 71 kfree(seg); in xhci_segment_free() 84 seg = next; in xhci_ring_segments_free() 101 if (!seg || !seg->next) in xhci_set_link_trb() 171 for (seg = dst->enq_seg; seg != dst->last_seg; seg = seg->next) in xhci_link_rings() 172 seg->next->num = seg->num + 1; in xhci_link_rings() [all …]
|
| /drivers/media/test-drivers/visl/ |
| A D | visl-trace-vp9.h | 65 __entry->f.seg.feature_data, 66 sizeof(__entry->f.seg.feature_data), 69 ARRAY_SIZE(__entry->f.seg.feature_enabled), 70 sizeof(__entry->f.seg.feature_enabled[0])), 71 __print_array(__entry->f.seg.tree_probs, 72 ARRAY_SIZE(__entry->f.seg.tree_probs), 73 sizeof(__entry->f.seg.tree_probs[0])), 74 __print_array(__entry->f.seg.pred_probs, 75 ARRAY_SIZE(__entry->f.seg.pred_probs), 76 sizeof(__entry->f.seg.pred_probs[0])), [all …]
|
| /drivers/infiniband/hw/mthca/ |
| A D | mthca_mr.c | 88 u32 seg; in mthca_buddy_alloc() local 96 if (seg < m) in mthca_buddy_alloc() 104 __clear_bit(seg, buddy->bits[o]); in mthca_buddy_alloc() 109 seg <<= 1; in mthca_buddy_alloc() 116 seg <<= order; in mthca_buddy_alloc() 118 return seg; in mthca_buddy_alloc() 123 seg >>= order; in mthca_buddy_free() 130 seg >>= 1; in mthca_buddy_free() 193 if (seg == -1) in mthca_alloc_mtt_range() 200 seg = -1; in mthca_alloc_mtt_range() [all …]
|
| /drivers/peci/ |
| A D | internal.h | 47 struct peci_request *peci_xfer_ep_pci_cfg_local_readb(struct peci_device *device, u8 seg, 49 struct peci_request *peci_xfer_ep_pci_cfg_local_readw(struct peci_device *device, u8 seg, 51 struct peci_request *peci_xfer_ep_pci_cfg_local_readl(struct peci_device *device, u8 seg, 54 struct peci_request *peci_xfer_ep_pci_cfg_readb(struct peci_device *device, u8 seg, 56 struct peci_request *peci_xfer_ep_pci_cfg_readw(struct peci_device *device, u8 seg, 58 struct peci_request *peci_xfer_ep_pci_cfg_readl(struct peci_device *device, u8 seg, 61 struct peci_request *peci_xfer_ep_mmio32_readl(struct peci_device *device, u8 bar, u8 seg, 64 struct peci_request *peci_xfer_ep_mmio64_readl(struct peci_device *device, u8 bar, u8 seg,
|
| /drivers/iommu/amd/ |
| A D | debugfs.c | 142 int seg, bus, slot, func; in devid_write() local 164 seg = 0; in devid_write() 171 if (pci_seg->id != seg) in devid_write() 185 if (pci_seg->id != seg) { in devid_write() 190 sbdf = PCI_SEG_DEVID_TO_SBDF(seg, devid); in devid_write() 239 u16 seg, devid; in iommu_devtbl_show() local 245 seg = PCI_SBDF_TO_SEGID(sbdf); in iommu_devtbl_show() 249 if (pci_seg->id != seg) in iommu_devtbl_show() 338 u16 devid, seg; in iommu_irqtbl_show() local 350 seg = PCI_SBDF_TO_SEGID(sbdf); in iommu_irqtbl_show() [all …]
|
| /drivers/media/platform/verisilicon/ |
| A D | hantro_g1_vp8_dec.c | 137 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_lf() local 143 if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) { in cfg_lf() 147 u32 lf_level = clamp(lf->level + seg->lf_update[i], in cfg_lf() 155 seg->lf_update[i]); in cfg_lf() 180 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_qp() local 184 if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) { in cfg_qp() 188 u32 quant = clamp(q->y_ac_qi + seg->quant_update[i], in cfg_qp() 196 seg->quant_update[i]); in cfg_qp() 410 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_buffers() local 421 if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) { in cfg_buffers() [all …]
|
| A D | rockchip_vpu2_hw_vp8_dec.c | 278 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_lf() local 284 if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) { in cfg_lf() 288 u32 lf_level = clamp(lf->level + seg->lf_update[i], in cfg_lf() 296 seg->lf_update[i]); in cfg_lf() 318 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_qp() local 322 if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) { in cfg_qp() 326 u32 quant = clamp(q->y_ac_qi + seg->quant_update[i], in cfg_qp() 334 seg->quant_update[i]); in cfg_qp() 486 const struct v4l2_vp8_segment *seg = &hdr->segment; in cfg_buffers() local 497 if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) { in cfg_buffers() [all …]
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_txrx.c | 252 int seg; in otx2_free_rcv_seg() local 259 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 343 int seg; in otx2_rcv_pkt_handler() local 368 for (seg = 0; seg < sg->segs; seg++, seg_addr++) { in otx2_rcv_pkt_handler() 668 int seg, len; in otx2_sqe_add_sg() local 672 for (seg = 0; seg < num_segs; seg++) { in otx2_sqe_add_sg() 860 for (seg = 0; seg < num_segs; seg++) { in otx2_dma_map_tso_skb() 887 if (seg < 0) in otx2_tso_frag_dma_addr() 893 seg++; in otx2_tso_frag_dma_addr() 903 int seg; in otx2_sqe_tso_add_sg() local [all …]
|
| /drivers/thermal/ |
| A D | k3_j72xx_bandgap.c | 82 static void create_table_segments(struct err_values *err_vals, int seg, in create_table_segments() argument 87 if (seg == 0) in create_table_segments() 90 idx1 = err_vals->refs[seg]; in create_table_segments() 92 idx2 = err_vals->refs[seg + 1]; in create_table_segments() 93 err1 = err_vals->errs[seg]; in create_table_segments() 94 err2 = err_vals->errs[seg + 1]; in create_table_segments() 95 ref1 = err_vals->refs[seg]; in create_table_segments() 96 ref2 = err_vals->refs[seg + 1]; in create_table_segments() 130 int inc, i, seg; in prep_lookup_table() local 138 for (seg = 0; seg < 3; seg++) in prep_lookup_table() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | soc15.h | 56 uint32_t seg; member 63 uint32_t seg; member 73 uint32_t seg; member 82 uint32_t seg; member 94 #define SOC15_REG_ENTRY_OFFSET(entry) (adev->reg_offset[entry.hwip][entry.inst][entry.seg] + entry.… 98 (adev->reg_offset[entry.hwip][inst][entry.seg] + entry.reg_offset)
|
| /drivers/media/platform/rockchip/rkvdec/ |
| A D | rkvdec-vp9.c | 153 struct v4l2_vp9_segmentation seg; member 317 const struct v4l2_vp9_segmentation *seg; in init_probs() local 323 seg = &dec_params->seg; in init_probs() 451 const struct v4l2_vp9_segmentation *seg; in config_seg_registers() local 457 seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg; in config_seg_registers() 460 feature_val = seg->feature_data[segid][feature_id]; in config_seg_registers() 467 feature_val = seg->feature_data[segid][feature_id]; in config_seg_registers() 474 feature_val = seg->feature_data[segid][feature_id]; in config_seg_registers() 507 vp9_ctx->cur.seg = dec_params->seg; in update_ctx_cur_info() 525 const struct v4l2_vp9_segmentation *seg; in config_registers() local [all …]
|
| /drivers/input/ |
| A D | touch-overlay.c | 124 static bool touch_overlay_segment_event(struct touch_overlay_segment *seg, in touch_overlay_segment_event() argument 127 if (pos->x >= seg->x_origin && pos->x < (seg->x_origin + seg->x_size) && in touch_overlay_segment_event() 128 pos->y >= seg->y_origin && pos->y < (seg->y_origin + seg->y_size)) in touch_overlay_segment_event()
|