| /linux/block/ |
| A D | badblocks.c | 500 p = bb->page; in prev_badblocks() 893 p = bb->page; in _badblocks_set() 902 bb->count++; in _badblocks_set() 942 bb->count--; in _badblocks_set() 1007 bb->count++; in _badblocks_set() 1465 if (bb->page == NULL || bb->changed) in ack_all_badblocks() 1470 if (bb->changed == 0 && bb->unacked_exist) { in ack_all_badblocks() 1613 if (!bb) in devm_init_badblocks() 1625 if (!bb) in badblocks_exit() 1627 if (bb->dev) in badblocks_exit() [all …]
|
| /linux/tools/bpf/bpftool/ |
| A D | cfg.c | 50 #define bb_prev(bb) list_prev_entry(bb, l) argument 51 #define bb_next(bb) list_next_entry(bb, l) argument 100 bb = bb_prev(bb); in func_append_bb() 118 bb = calloc(1, sizeof(*bb)); in func_insert_dummy_bb() 119 if (!bb) { in func_insert_dummy_bb() 128 return bb; in func_insert_dummy_bb() 174 if (!bb) in func_partition_bb_head() 208 bb->tail = bb_next(bb)->head - 1; in func_partition_bb_tail() 221 if (!bb) in func_add_special_bb() 226 if (!bb) in func_add_special_bb() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_bb.c | 35 struct xe_bb *bb = kmalloc(sizeof(*bb), GFP_KERNEL); in xe_bb_new() local 38 if (!bb) in xe_bb_new() 49 if (IS_ERR(bb->bo)) { in xe_bb_new() 54 bb->cs = xe_sa_bo_cpu_addr(bb->bo); in xe_bb_new() 55 bb->len = 0; in xe_bb_new() 57 return bb; in xe_bb_new() 59 kfree(bb); in xe_bb_new() 68 if (bb->len == 0 || bb->cs[bb->len - 1] != MI_BATCH_BUFFER_END) in __xe_bb_create_job() 69 bb->cs[bb->len++] = MI_BATCH_BUFFER_END; in __xe_bb_create_job() 108 if (!bb) in xe_bb_free() [all …]
|
| A D | xe_gsc_submit.c | 177 struct xe_bb *bb; in xe_gsc_pkt_submit_kernel() local 189 if (IS_ERR(bb)) in xe_gsc_pkt_submit_kernel() 190 return PTR_ERR(bb); in xe_gsc_pkt_submit_kernel() 192 bb->cs[bb->len++] = GSC_HECI_CMD_PKT; in xe_gsc_pkt_submit_kernel() 193 bb->cs[bb->len++] = lower_32_bits(addr_in); in xe_gsc_pkt_submit_kernel() 194 bb->cs[bb->len++] = upper_32_bits(addr_in); in xe_gsc_pkt_submit_kernel() 195 bb->cs[bb->len++] = size_in; in xe_gsc_pkt_submit_kernel() 196 bb->cs[bb->len++] = lower_32_bits(addr_out); in xe_gsc_pkt_submit_kernel() 197 bb->cs[bb->len++] = upper_32_bits(addr_out); in xe_gsc_pkt_submit_kernel() 198 bb->cs[bb->len++] = size_out; in xe_gsc_pkt_submit_kernel() [all …]
|
| A D | xe_migrate.c | 584 bb->cs[bb->len++] = ofs; in emit_pte() 585 bb->cs[bb->len++] = 0; in emit_pte() 629 u32 *cs = bb->cs + bb->len; in emit_copy_ccs() 660 bb->len = cs - bb->cs; in emit_copy_ccs() 684 bb->cs[bb->len++] = 0; in emit_copy() 688 bb->cs[bb->len++] = 0; in emit_copy() 689 bb->cs[bb->len++] = pitch | mocs; in emit_copy() 943 u32 *cs = bb->cs + bb->len; in emit_clear_link_copy() 966 u32 *cs = bb->cs + bb->len; in emit_clear_main_copy() 1223 bb->cs[bb->len++] = MI_NOOP; in write_pgtable() [all …]
|
| A D | xe_gt.c | 167 struct xe_bb *bb; in emit_nop_job() local 172 if (IS_ERR(bb)) in emit_nop_job() 173 return PTR_ERR(bb); in emit_nop_job() 187 xe_bb_free(bb, NULL); in emit_nop_job() 211 struct xe_bb *bb; in emit_wa_job() local 223 if (IS_ERR(bb)) in emit_wa_job() 224 return PTR_ERR(bb); in emit_wa_job() 232 bb->cs[bb->len++] = MI_LOAD_REGISTER_IMM | MI_LRI_NUM_REGS(count); in emit_wa_job() 253 bb->cs[bb->len++] = reg.addr; in emit_wa_job() 254 bb->cs[bb->len++] = val; in emit_wa_job() [all …]
|
| A D | xe_gsc.c | 72 struct xe_bb *bb; in emit_gsc_upload() local 77 bb = xe_bb_new(gt, 4, false); in emit_gsc_upload() 78 if (IS_ERR(bb)) in emit_gsc_upload() 79 return PTR_ERR(bb); in emit_gsc_upload() 81 bb->cs[bb->len++] = GSC_FW_LOAD; in emit_gsc_upload() 82 bb->cs[bb->len++] = lower_32_bits(offset); in emit_gsc_upload() 83 bb->cs[bb->len++] = upper_32_bits(offset); in emit_gsc_upload() 84 bb->cs[bb->len++] = (gsc->private->size / SZ_4K) | GSC_FW_LOAD_LIMIT_VALID; in emit_gsc_upload() 86 job = xe_bb_create_job(gsc->q, bb); in emit_gsc_upload() 88 xe_bb_free(bb, NULL); in emit_gsc_upload() [all …]
|
| A D | xe_bb.h | 19 struct xe_bb *bb); 21 struct xe_bb *bb, u64 batch_ofs, 23 void xe_bb_free(struct xe_bb *bb, struct dma_fence *fence);
|
| /linux/include/linux/ |
| A D | badblocks.h | 56 void ack_all_badblocks(struct badblocks *bb); 61 void badblocks_exit(struct badblocks *bb); 66 if (bb->dev != dev) { in devm_exit_badblocks() 71 badblocks_exit(bb); in devm_exit_badblocks() 76 return (bb->count >= MAX_BADBLOCKS); in badblocks_full() 81 return (bb->count == 0); in badblocks_empty() 84 static inline void set_changed(struct badblocks *bb) in set_changed() argument 86 if (bb->changed != 1) in set_changed() 87 bb->changed = 1; in set_changed() 92 if (bb->changed != 0) in clear_changed() [all …]
|
| /linux/net/tipc/ |
| A D | bcast.c | 101 bb->bc_threshold = 1 + (cluster_size * bb->rc_ratio / 100); in tipc_bcbase_calc_bc_threshold() 122 if (!bb->dests[i]) in tipc_bcbase_select_primary() 619 if (!bb->bcast_support || !bb->rcast_support) in tipc_bc_link_set_broadcast_mode() 636 if (!bb->bcast_support || !bb->rcast_support) in tipc_bc_link_set_broadcast_ratio() 695 bb = kzalloc(sizeof(*bb), GFP_KERNEL); in tipc_bcast_init() 696 if (!bb) in tipc_bcast_init() 698 tn->bcbase = bb; in tipc_bcast_init() 706 &bb->inputq, in tipc_bcast_init() 711 bb->link = l; in tipc_bcast_init() 717 kfree(bb); in tipc_bcast_init() [all …]
|
| /linux/crypto/ |
| A D | rmd160.c | 55 bb = state[1]; in rmd160_transform() 68 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd160_transform() 69 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd160_transform() 70 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd160_transform() 71 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd160_transform() 72 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd160_transform() 73 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd160_transform() 74 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd160_transform() 75 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd160_transform() 76 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd160_transform() [all …]
|
| /linux/arch/alpha/boot/tools/ |
| A D | objstrip.c | 109 unsigned long bb[64], sum = 0; in main() local 124 memset(bb, 0, sizeof(bb)); in main() 125 strcpy((char *) bb, "Linux SRM bootblock"); in main() 126 bb[60] = size / BLOCK_SIZE; /* count */ in main() 127 bb[61] = 1; /* starting sector # */ in main() 128 bb[62] = 0; /* flags---must be 0 */ in main() 130 sum += bb[i]; in main() 132 bb[63] = sum; in main() 133 if (write(ofd, bb, sizeof(bb)) != sizeof(bb)) { in main()
|
| /linux/arch/mips/lib/ |
| A D | multi3.c | 35 TWunion res, aa, bb; in __multi3() local 38 bb.ti = b; in __multi3() 45 res.s.low = dmulu(aa.s.low, bb.s.low); in __multi3() 46 res.s.high = dmuhu(aa.s.low, bb.s.low); in __multi3() 47 res.s.high += dmulu(aa.s.high, bb.s.low); in __multi3() 48 res.s.high += dmulu(aa.s.low, bb.s.high); in __multi3()
|
| /linux/drivers/nvdimm/ |
| A D | badrange.c | 165 static void set_badblock(struct badblocks *bb, sector_t s, int num) in set_badblock() argument 167 dev_dbg(bb->dev, "Found a bad range (0x%llx, 0x%llx)\n", in set_badblock() 170 if (badblocks_set(bb, s, num, 1)) in set_badblock() 171 dev_info_once(bb->dev, "%s: failed for sector %llx\n", in set_badblock() 205 set_badblock(bb, s, done); in __add_badblock_range() 210 set_badblock(bb, start_sector, num_sectors); in __add_badblock_range() 214 struct badblocks *bb, const struct range *range) in badblocks_populate() argument 239 __add_badblock_range(bb, start - range->start, len); in badblocks_populate() 253 __add_badblock_range(bb, 0, len); in badblocks_populate() 270 struct badblocks *bb, const struct range *range) in nvdimm_badblocks_populate() argument [all …]
|
| A D | pmem.c | 41 return pmem->bb.dev; in to_dev() 93 badblocks_clear(&pmem->bb, sector, blks); in pmem_clear_bb() 250 struct badblocks *bb = &pmem->bb; in __pmem_direct_access() local 259 if (bb->count && in __pmem_direct_access() 284 if (bb->count) in __pmem_direct_access() 332 struct device *dev = pmem->bb.dev; in pmem_recovery_write() 555 if (devm_init_badblocks(dev, &pmem->bb)) in pmem_attach_disk() 558 disk->bb = &pmem->bb; in pmem_attach_disk() 680 struct badblocks *bb; in pmem_revalidate_poison() local 690 bb = &nsio->bb; in pmem_revalidate_poison() [all …]
|
| /linux/drivers/spi/ |
| A D | spi-gpio.c | 377 struct spi_bitbang *bb; in spi_gpio_probe() local 416 bb = &spi_gpio->bitbang; in spi_gpio_probe() 417 bb->ctlr = host; in spi_gpio_probe() 424 bb->chipselect = spi_gpio_chipselect; in spi_gpio_probe() 425 bb->set_line_direction = spi_gpio_set_direction; in spi_gpio_probe() 426 bb->set_mosi_idle = spi_gpio_set_mosi_idle; in spi_gpio_probe() 434 bb->txrx_word[SPI_MODE_0] = spi_gpio_txrx_word_mode0; in spi_gpio_probe() 435 bb->txrx_word[SPI_MODE_1] = spi_gpio_txrx_word_mode1; in spi_gpio_probe() 436 bb->txrx_word[SPI_MODE_2] = spi_gpio_txrx_word_mode2; in spi_gpio_probe() 437 bb->txrx_word[SPI_MODE_3] = spi_gpio_txrx_word_mode3; in spi_gpio_probe() [all …]
|
| /linux/scripts/gcc-plugins/ |
| A D | latent_entropy_plugin.c | 349 gsi = gsi_after_labels(bb); in perturb_local_entropy() 437 gimple_stmt_iterator gsi = gsi_after_labels(bb); in init_local_entropy() 503 basic_block bb; in latent_entropy_execute() local 511 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in latent_entropy_execute() 512 if (!single_pred_p(bb)) { in latent_entropy_execute() 515 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in latent_entropy_execute() 522 init_local_entropy(bb, local_entropy); in latent_entropy_execute() 524 bb = bb->next_bb; in latent_entropy_execute() 530 while (bb != EXIT_BLOCK_PTR_FOR_FN(cfun)) { in latent_entropy_execute() 531 perturb_local_entropy(bb, local_entropy); in latent_entropy_execute() [all …]
|
| A D | stackleak_plugin.c | 59 basic_block bb; in add_stack_tracking_gcall() local 70 bb = gimple_bb(gimple_call); in add_stack_tracking_gcall() 74 gimple_call, bb->count, in add_stack_tracking_gcall() 75 compute_call_stmt_bb_frequency(current_function_decl, bb)); in add_stack_tracking_gcall() 173 basic_block bb, entry_bb; in stackleak_instrument_execute() local 190 FOR_EACH_BB_FN(bb, cfun) { in stackleak_instrument_execute() 210 if (bb == entry_bb) in stackleak_instrument_execute() 245 bb = entry_bb; in stackleak_instrument_execute() 246 if (!single_pred_p(bb)) { in stackleak_instrument_execute() 251 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in stackleak_instrument_execute() [all …]
|
| A D | structleak_plugin.c | 118 basic_block bb; in initialize() local 125 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in initialize() 128 for (gsi = gsi_start_bb(bb); !gsi_end_p(gsi); gsi_next(&gsi)) { in initialize() 170 basic_block bb; in structleak_execute() local 176 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in structleak_execute() 177 if (!single_pred_p(bb)) { in structleak_execute()
|
| /linux/drivers/gpu/drm/xe/tests/ |
| A D | xe_migrate.c | 191 struct xe_bb *bb; in xe_migrate_sanity_test() local 233 if (IS_ERR(bb)) { in xe_migrate_sanity_test() 235 PTR_ERR(bb)); in xe_migrate_sanity_test() 257 run_sanity_job(m, xe, bb, bb->len, "Writing PTE for our fake PT", test); in xe_migrate_sanity_test() 264 bb->len = 0; in xe_migrate_sanity_test() 265 bb->cs[bb->len++] = MI_BATCH_BUFFER_END; in xe_migrate_sanity_test() 324 xe_bb_free(bb, NULL); in xe_migrate_sanity_test() 398 struct xe_bb *bb; in blt_copy() local 426 if (IS_ERR(bb)) { in blt_copy() 427 err = PTR_ERR(bb); in blt_copy() [all …]
|
| /linux/net/netfilter/ |
| A D | nft_set_pipapo.c | 446 if (likely(f->bb == 8)) in nft_pipapo_lookup() 554 if (f->bb == 8) in pipapo_get() 556 else if (f->bb == 4) in pipapo_get() 908 unsigned int groups, bb; in pipapo_lt_bits_adjust() local 944 if (f->bb == 4 && bb == 8) { in pipapo_lt_bits_adjust() 948 } else if (f->bb == 8 && bb == 4) { in pipapo_lt_bits_adjust() 957 f->bb = bb; in pipapo_lt_bits_adjust() 992 bit_offset += f->bb; in pipapo_insert() 1004 mask = GENMASK(f->bb - 1, 0); in pipapo_insert() 1989 bit_offset += f->bb; [all …]
|
| /linux/fs/xfs/libxfs/ |
| A D | xfs_bmap_btree.h | 151 struct xfs_btree_block *bb, in xfs_bmap_broot_ptr_addr() argument 155 return xfs_bmbt_ptr_addr(mp, bb, i, xfs_bmbt_maxrecs(mp, sz, false)); in xfs_bmap_broot_ptr_addr() 178 struct xfs_bmdr_block *bb) in xfs_bmap_broot_space() argument 180 return xfs_bmap_broot_space_calc(mp, be16_to_cpu(bb->bb_numrecs)); in xfs_bmap_broot_space() 196 xfs_bmap_bmdr_space(struct xfs_btree_block *bb) in xfs_bmap_bmdr_space() argument 198 return xfs_bmdr_space_calc(be16_to_cpu(bb->bb_numrecs)); in xfs_bmap_bmdr_space()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| A D | selftest_ring_submission.c | 207 struct i915_vma *bb; in __live_ctx_switch_wa() local 211 bb = create_wally(engine); in __live_ctx_switch_wa() 212 if (IS_ERR(bb)) in __live_ctx_switch_wa() 213 return PTR_ERR(bb); in __live_ctx_switch_wa() 215 result = i915_gem_object_pin_map_unlocked(bb->obj, I915_MAP_WC); in __live_ctx_switch_wa() 217 intel_context_put(bb->private); in __live_ctx_switch_wa() 218 i915_vma_unpin_and_release(&bb, 0); in __live_ctx_switch_wa() 223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa()
|
| /linux/drivers/gpu/drm/amd/display/dc/dml/dcn20/ |
| A D | dcn20_fpu.c | 1863 memset(bb->clock_limits, 0, sizeof(bb->clock_limits)); in dcn20_update_bounding_box() 1911 bb->clock_limits[num_calculated_states].state = bb->num_states; in dcn20_update_bounding_box() 1960 if (bb->clock_limits[i-1].dcfclk_mhz != bb->clock_limits[i].dcfclk_mhz) in dcn20_cap_soc_clocks() 1962 if (bb->clock_limits[i-1].dispclk_mhz != bb->clock_limits[i].dispclk_mhz) in dcn20_cap_soc_clocks() 1964 if (bb->clock_limits[i-1].dppclk_mhz != bb->clock_limits[i].dppclk_mhz) in dcn20_cap_soc_clocks() 1966 if (bb->clock_limits[i-1].dram_speed_mts != bb->clock_limits[i].dram_speed_mts) in dcn20_cap_soc_clocks() 1968 if (bb->clock_limits[i-1].dscclk_mhz != bb->clock_limits[i].dscclk_mhz) in dcn20_cap_soc_clocks() 1970 if (bb->clock_limits[i-1].fabricclk_mhz != bb->clock_limits[i].fabricclk_mhz) in dcn20_cap_soc_clocks() 1972 if (bb->clock_limits[i-1].phyclk_mhz != bb->clock_limits[i].phyclk_mhz) in dcn20_cap_soc_clocks() 1974 if (bb->clock_limits[i-1].socclk_mhz != bb->clock_limits[i].socclk_mhz) in dcn20_cap_soc_clocks() [all …]
|
| /linux/tools/testing/selftests/net/forwarding/ |
| A D | tc_tunnel_key.sh | 117 $MZ h1-et -c 1 -p 930 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q 121 $MZ h1-et -c 1 -p 931 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q 131 $MZ h1-et -c 1 -p 931 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q
|