| /linux/rust/macros/ |
| A D | quote.rs | 42 ($span:expr => $($tt:tt)*) => {{ 47 let span = $span; 52 (@proc $v:ident $span:ident) => {}; 57 quote_spanned!(@proc $v $span $($tt)*); 65 quote_spanned!(@proc $v $span $($tt)*); 74 quote_spanned!(@proc $v $span $($tt)*); 83 quote_spanned!(@proc $v $span $($tt)*); 92 quote_spanned!(@proc $v $span $($tt)*); 101 quote_spanned!(@proc $v $span $($tt)*); 107 quote_spanned!(@proc $v $span $($tt)*); [all …]
|
| A D | paste.rs | 8 let mut span = None; in concat() variables 19 segments.push((value, lit.span())); in concat() 26 segments.push((value, ident.span())); in concat() 38 span.is_none(), in concat() 41 span = Some(sp); in concat() 54 TokenTree::Ident(Ident::new(&pasted, span.unwrap_or(group_span))) in concat() 61 let span = group.span(); in expand() 70 *token = concat(&stream[1..stream.len() - 1], span); in expand() 75 group.set_span(span); in expand()
|
| /linux/drivers/net/ethernet/marvell/prestera/ |
| A D | prestera_span.c | 53 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_id() 67 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_port() 171 struct prestera_span *span; in prestera_span_init() local 173 span = kzalloc(sizeof(*span), GFP_KERNEL); in prestera_span_init() 174 if (!span) in prestera_span_init() 177 INIT_LIST_HEAD(&span->entries); in prestera_span_init() 179 sw->span = span; in prestera_span_init() 180 span->sw = sw; in prestera_span_init() 187 struct prestera_span *span = sw->span; in prestera_span_fini() local 189 WARN_ON(!list_empty(&span->entries)); in prestera_span_fini() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_span.c | 46 struct mlxsw_sp_span *span; member 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 91 if (!span) in mlxsw_sp_span_init() 99 span->mlxsw_sp = mlxsw_sp; in mlxsw_sp_span_init() 100 mlxsw_sp->span = span; in mlxsw_sp_span_init() 117 kfree(mlxsw_sp->span); in mlxsw_sp_span_init() 131 kfree(mlxsw_sp->span); in mlxsw_sp_span_fini() 825 policer_id >= span->policer_id_base + span->entries_count) in mlxsw_sp_span_policer_id_base_set() 1020 struct mlxsw_sp_span *span = mlxsw_sp->span; in mlxsw_sp_span_entry_ops() local 1478 struct mlxsw_sp_span *span = trigger_entry->span; in mlxsw_sp_span_trigger_ops_set() local [all …]
|
| /linux/drivers/scsi/megaraid/ |
| A D | megaraid_sas_fp.c | 151 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 381 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 398 return span; in MR_GetSpanBlock() 442 for (span = 0; span < raid->spanDepth; span++) in mr_spanset_get_span_block() 463 return span; in mr_spanset_get_span_block() 509 for (span = 0, span_offset = 0; span < raid->spanDepth; span++) in get_row_from_strip() 562 for (span = 0; span < raid->spanDepth; span++) in get_strip_from_row() 625 for (span = 0, span_offset = 0; span < raid->spanDepth; span++) in get_arm_from_strip() 832 span = 0; in MR_GetPhyParams() 927 span = 0; in mr_get_phy_params_r56_rmw() [all …]
|
| /linux/drivers/iio/dac/ |
| A D | ltc2664.c | 93 u8 span; member 136 int span, fs; in ltc2664_scale_get() local 138 span = chan->span; in ltc2664_scale_get() 139 if (span < 0) in ltc2664_scale_get() 140 return span; in ltc2664_scale_get() 142 fs = span_helper[span][1] - span_helper[span][0]; in ltc2664_scale_get() 150 int span, fs; in ltc2672_scale_get() local 152 span = chan->span - 1; in ltc2672_scale_get() 167 int span; in ltc2664_offset_get() local 169 span = chan->span; in ltc2664_offset_get() [all …]
|
| A D | ltc2688.c | 150 return span; in ltc2688_span_get() 160 int span, fs; in ltc2688_scale_get() local 163 if (span < 0) in ltc2688_scale_get() 164 return span; in ltc2688_scale_get() 177 int span; in ltc2688_offset_get() local 180 if (span < 0) in ltc2688_offset_get() 181 return span; in ltc2688_offset_get() 735 u32 span; in ltc2688_span_lookup() local 737 for (span = 0; span < ARRAY_SIZE(ltc2688_span_helper); span++) { in ltc2688_span_lookup() 740 return span; in ltc2688_span_lookup() [all …]
|
| A D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 210 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 212 if (span <= 10000) in ad5791_get_lin_comp() 214 else if (span <= 12000) in ad5791_get_lin_comp() 216 else if (span <= 16000) in ad5791_get_lin_comp() 218 else if (span <= 19000) in ad5791_get_lin_comp() 224 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 226 if (span <= 10000) in ad5780_get_lin_comp()
|
| /linux/fs/netfs/ |
| A D | iterator.c | 114 size_t len, span = 0, n = iter->count; in netfs_limit_bvec() local 133 span += len; in netfs_limit_bvec() 136 if (span >= max_size || nsegs >= max_segs) in netfs_limit_bvec() 142 return min(span, max_size); in netfs_limit_bvec() 158 size_t span = 0, n = iter->count; in netfs_limit_xarray() local 181 span += len; in netfs_limit_xarray() 183 if (span >= max_size || nsegs >= max_segs) in netfs_limit_xarray() 188 return min(span, max_size); in netfs_limit_xarray() 202 size_t span = 0, n = iter->count; in netfs_limit_folioq() local 220 span += flen - start_offset; in netfs_limit_folioq() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| A D | pages.c | 968 if (span.is_used) in pfn_reader_unpin() 972 span.last_hole - span.start_hole + 1); in pfn_reader_unpin() 979 struct interval_tree_double_span_iter *span = &pfns->span; in pfn_reader_fill_span() local 1013 span->last_hole); in pfn_reader_fill_span() 1289 if (span.is_used) { in __iopt_area_unfill_domain() 1291 span.last_used - span.start_used + 1); in __iopt_area_unfill_domain() 1295 span.start_hole, span.last_hole, in __iopt_area_unfill_domain() 1557 if (!span.is_used) { in iopt_pages_unfill_xarray() 1565 span.last_hole); in iopt_pages_unfill_xarray() 1569 span.last_used); in iopt_pages_unfill_xarray() [all …]
|
| A D | double_span.h | 46 #define interval_tree_for_each_double_span(span, itree1, itree2, first_index, \ argument 48 for (interval_tree_double_span_iter_first(span, itree1, itree2, \ 50 !interval_tree_double_span_iter_done(span); \ 51 interval_tree_double_span_iter_next(span))
|
| A D | io_pagetable.c | 73 static bool __alloc_iova_check_hole(struct interval_tree_double_span_iter *span, in __alloc_iova_check_hole() argument 78 if (span->is_used || span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 81 span->start_hole = ALIGN(span->start_hole, iova_alignment) | in __alloc_iova_check_hole() 83 if (span->start_hole > span->last_hole || in __alloc_iova_check_hole() 84 span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 89 static bool __alloc_iova_check_used(struct interval_tree_span_iter *span, in __alloc_iova_check_used() argument 94 if (span->is_hole || span->last_used - span->start_used < length - 1) in __alloc_iova_check_used() 97 span->start_used = ALIGN(span->start_used, iova_alignment) | in __alloc_iova_check_used() 99 if (span->start_used > span->last_used || in __alloc_iova_check_used() 100 span->last_used - span->start_used < length - 1) in __alloc_iova_check_used()
|
| A D | ioas.c | 67 struct interval_tree_span_iter span; in iommufd_ioas_iova_ranges() local 83 interval_tree_for_each_span(&span, &ioas->iopt.reserved_itree, 0, in iommufd_ioas_iova_ranges() 85 if (!span.is_hole) in iommufd_ioas_iova_ranges() 89 .start = span.start_hole, in iommufd_ioas_iova_ranges() 90 .last = span.last_hole, in iommufd_ioas_iova_ranges()
|
| A D | vfio_compat.c | 391 struct interval_tree_span_iter span; in iommufd_fill_cap_iova() local 393 interval_tree_for_each_span(&span, &ioas->iopt.reserved_itree, 0, in iommufd_fill_cap_iova() 397 if (!span.is_hole) in iommufd_fill_cap_iova() 399 range.start = span.start_hole; in iommufd_fill_cap_iova() 400 range.end = span.last_hole; in iommufd_fill_cap_iova()
|
| /linux/Documentation/translations/zh_CN/scheduler/ |
| A D | sched-domains.rst | 22 每个调度域管辖数个CPU(存储在->span字段中)。一个调度域的span必须是它的子调度域span的 25 CPU出现永远不会被指定任务运行的情况,直到允许的CPU掩码被显式设定。调度域的span字段意味 29 组织,存储在->groups指针中。这些组的CPU掩码的并集必须和调度域span字段一致。->groups
|
| /linux/include/linux/ |
| A D | interval_tree.h | 82 #define interval_tree_for_each_span(span, itree, first_index, last_index) \ argument 83 for (interval_tree_span_iter_first(span, itree, \ 85 !interval_tree_span_iter_done(span); \ 86 interval_tree_span_iter_next(span))
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | test_access_variable_array.c | 8 unsigned long span = 0; variable 14 span = sd->span[0]; in BPF_PROG()
|
| /linux/tools/testing/selftests/resctrl/ |
| A D | cmt_test.c | 77 static int check_results(struct resctrl_val_param *param, size_t span, int no_of_bits) in check_results() argument 108 return show_results_info(sum_llc_occu_resc, no_of_bits, span, in check_results() 126 size_t span; in cmt_run_test() local 156 span = cache_portion_size(cache_total_size, param.mask, long_mask); in cmt_run_test() 164 ret = asprintf(&span_str, "%zu", span); in cmt_run_test() 177 ret = check_results(¶m, span, n); in cmt_run_test()
|
| A D | cat_test.c | 159 size_t span, unsigned long current_mask) in cat_test() argument 192 buf = alloc_buffer(span, 1); in cat_test() 209 mem_flush(buf, span); in cat_test() 210 fill_cache_read(buf, span, true); in cat_test() 216 fill_cache_read(buf, span, true); in cat_test() 242 size_t span; in cat_run_test() local 278 span = cache_portion_size(cache_total_size, start_mask, full_cache_mask); in cat_run_test() 282 ret = cat_test(test, uparams, ¶m, span, start_mask); in cat_run_test()
|
| A D | mbm_test.c | 18 show_bw_info(unsigned long *bw_imc, unsigned long *bw_resc, size_t span) in show_bw_info() argument 43 ksft_print_msg("Span (MB): %zu\n", span / MB); in show_bw_info() 50 static int check_results(size_t span) in check_results() argument 82 ret = show_bw_info(bw_imc, bw_resc, span); in check_results()
|
| /linux/Documentation/scheduler/ |
| A D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The group pointed to by the ->groups pointer MUST contain the CPU 58 The "base" domain will "span" the first level of the hierarchy. In the case 59 of SMT, you'll span all siblings of the physical CPU, with each group being 62 In SMP, the parent of the base domain will span all physical CPUs in the 64 of the SMP domain will span the entire machine, with each group having the
|
| /linux/drivers/platform/surface/aggregator/ |
| A D | ssh_parser.h | 138 struct ssam_span *span) in sshp_buf_span_from() argument 140 span->ptr = buf->ptr + offset; in sshp_buf_span_from() 141 span->len = buf->len - offset; in sshp_buf_span_from()
|
| /linux/drivers/net/wireless/realtek/rtw89/ |
| A D | sar.c | 58 #define RTW89_SAR_SPAN_VALID(span) ((span)->subband_high) argument 92 const struct rtw89_sar_span *span = NULL; in rtw89_query_sar_config_common() local 104 span = &rtw89_sar_overlapping_6ghz[idx]; in rtw89_query_sar_config_common() 107 if (span && RTW89_SAR_SPAN_VALID(span)) { in rtw89_query_sar_config_common() 108 subband_l = span->subband_low; in rtw89_query_sar_config_common() 109 subband_h = span->subband_high; in rtw89_query_sar_config_common()
|
| /linux/arch/sh/mm/ |
| A D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() 668 newsize = span; in pmb_merge()
|
| /linux/Documentation/sphinx-static/ |
| A D | theme_rtd_colors.css | 29 span.menuselection { 33 code.kbd, code.kbd span {
|