| /drivers/md/dm-vdo/ |
| A D | repair.c | 107 physical_block_number_t pbn; member 147 if (mapping1->block_map_slot.pbn != mapping2->block_map_slot.pbn) in mapping_is_less_than() 148 return mapping1->block_map_slot.pbn < mapping2->block_map_slot.pbn; in mapping_is_less_than() 528 return pbn; in get_pbn_to_fetch() 619 if ((pbn == VDO_ZERO_BLOCK) || !vdo_is_physical_data_block(depot, pbn)) { in process_entry() 822 pbn = entry.mapping.pbn; in add_slab_journal_entries() 824 pbn = entry.unmapping.pbn; in add_slab_journal_entries() 1035 pbn = repair->current_unfetched_entry->block_map_slot.pbn; in fetch_block_map_page() 1063 if (repair->pbn != page_completion->pbn) in recover_ready_pages() 1089 repair->pbn = repair->current_entry->block_map_slot.pbn; in recover_ready_pages() [all …]
|
| A D | block-map.c | 387 info->pbn = pbn; in set_info_pbn() 602 result = VDO_ASSERT(completion->info->pbn == completion->pbn, in validate_completed_page() 877 return (page_completion_from_waiter(waiter)->pbn == *pbn); in completion_needs_page() 909 pbn = page_completion_from_waiter(oldest_waiter)->pbn; in allocate_free_page() 1222 .pbn = pbn, in vdo_get_page() 2078 pbn = tree_lock->tree_slots[height - 1].block_map_slot.pbn; in finish_block_map_allocation() 2172 pbn = data_vio->allocation.pbn; in allocate_block() 2178 .pbn = pbn, in allocate_block() 2238 .pbn = 0, in vdo_find_block_map_slot() 2250 lock->tree_slots[lock->height].block_map_slot.pbn = pbn; in vdo_find_block_map_slot() [all …]
|
| A D | vio.c | 47 return ((pbn == VDO_GEOMETRY_BLOCK_LOCATION) ? pbn : pbn + vdo->geometry.bio_offset); in pbn_from_vio_bio() 174 blk_opf_t bi_opf, physical_block_number_t pbn) in vdo_set_bio_properties() argument 179 pbn -= vdo->geometry.bio_offset; in vdo_set_bio_properties() 180 vio->bio_zone = ((pbn / config->thread_counts.bio_rotation_interval) % in vdo_set_bio_properties() 186 bio->bi_iter.bi_sector = pbn * VDO_SECTORS_PER_BLOCK; in vdo_set_bio_properties() 195 blk_opf_t bi_opf, physical_block_number_t pbn) in vio_reset_bio() argument 198 callback, bi_opf, pbn); in vio_reset_bio() 202 blk_opf_t bi_opf, physical_block_number_t pbn) in vio_reset_bio_with_size() argument 210 vdo_set_bio_properties(bio, vio, callback, bi_opf, pbn); in vio_reset_bio_with_size() 289 physical_block_number_t pbn = pbn_from_vio_bio(vio->bio); in vio_record_metadata_io_error() local [all …]
|
| A D | dump.c | 163 wait_on, data_vio, data_vio->allocation.pbn, data_vio->logical.lbn, in dump_vio_waiters() 164 data_vio->duplicate.pbn, get_data_vio_operation_name(data_vio)); in dump_vio_waiters() 169 data_vio, data_vio->allocation.pbn, data_vio->logical.lbn, in dump_vio_waiters() 170 data_vio->duplicate.pbn, in dump_vio_waiters() 239 data_vio->allocation.pbn, data_vio->logical.lbn, in dump_data_vio() 240 data_vio->duplicate.pbn); in dump_data_vio() 244 data_vio->allocation.pbn, data_vio->logical.lbn); in dump_data_vio()
|
| A D | block-map.h | 139 physical_block_number_t pbn; member 173 physical_block_number_t pbn; member 290 typedef int (*vdo_entry_callback_fn)(physical_block_number_t pbn, 302 struct block_map_zone *zone, physical_block_number_t pbn, 320 physical_block_number_t pbn, 364 physical_block_number_t pbn,
|
| A D | slab-depot.h | 518 physical_block_number_t pbn, 525 physical_block_number_t pbn, 535 physical_block_number_t pbn, 548 physical_block_number_t pbn); 566 physical_block_number_t pbn); 569 physical_block_number_t pbn); 572 physical_block_number_t pbn);
|
| A D | encodings.h | 155 __le64 pbn; member 844 .pbn_high_nibble = ((pbn >> 32) & 0x0F), in vdo_pack_block_map_entry() 845 .pbn_low_word = __cpu_to_le32(pbn & UINT_MAX), in vdo_pack_block_map_entry() 855 .pbn = ((high4 << 32) | low32), in vdo_unpack_block_map_entry() 867 if (location->pbn == VDO_ZERO_BLOCK) in vdo_is_valid_location() 876 return __le64_to_cpu(page->header.pbn); in vdo_get_block_map_page_pbn() 880 physical_block_number_t pbn, 885 physical_block_number_t pbn); 911 .pbn_high_nibble = (entry->slot.pbn >> 32) & 0x0F, in vdo_pack_recovery_journal_entry() 913 .mapping = vdo_pack_block_map_entry(entry->mapping.pbn, in vdo_pack_recovery_journal_entry() [all …]
|
| A D | physical-zone.c | 416 physical_block_number_t pbn) in vdo_get_physical_zone_pbn_lock() argument 418 return ((zone == NULL) ? NULL : vdo_int_map_get(zone->pbn_operations, pbn)); in vdo_get_physical_zone_pbn_lock() 437 physical_block_number_t pbn, in vdo_attempt_physical_zone_pbn_lock() argument 454 result = vdo_int_map_put(zone->pbn_operations, pbn, new_lock, false, in vdo_attempt_physical_zone_pbn_lock() 465 (unsigned long long) pbn); in vdo_attempt_physical_zone_pbn_lock() 491 result = vdo_allocate_block(allocation->zone->allocator, &allocation->pbn); in allocate_and_lock_block() 495 result = vdo_attempt_physical_zone_pbn_lock(allocation->zone, allocation->pbn, in allocate_and_lock_block() 504 (unsigned long long) allocation->pbn, in allocate_and_lock_block()
|
| A D | data-vio.c | 1405 allocation->pbn = VDO_ZERO_BLOCK; in release_data_vio_allocation_lock() 1542 if (data_vio->mapped.pbn == VDO_ZERO_BLOCK) { in read_block() 1557 data_vio->mapped.pbn); in read_block() 1566 data_vio->mapped.pbn); in read_block() 1666 if (data_vio->new_mapped.pbn == VDO_ZERO_BLOCK) { in journal_remapping() 1668 if (data_vio->mapped.pbn == VDO_ZERO_BLOCK) in journal_remapping() 1675 if (data_vio->mapped.pbn == VDO_ZERO_BLOCK) { in journal_remapping() 1886 data_vio->allocation.pbn); in write_data_vio() 1912 if (data_vio->new_mapped.pbn == VDO_ZERO_BLOCK) { in acknowledge_write_callback() 1939 .pbn = data_vio->allocation.pbn, in allocate_block() [all …]
|
| A D | slab-depot.c | 318 physical_block_number_t pbn; in launch_write() local 343 pbn = (depot->summary_origin + in launch_write() 1187 physical_block_number_t pbn; in write_reference_block() local 1300 if (pbn < slab->start) in slab_block_number_from_pbn() 3338 physical_block_number_t pbn) in vdo_release_block_reference() argument 3342 if (pbn == VDO_ZERO_BLOCK) in vdo_release_block_reference() 3349 .pbn = pbn, in vdo_release_block_reference() 3484 physical_block_number_t pbn; in erase_next_slab_journal() local 4398 if (pbn < depot->first_block) in get_slab_number() 4425 if (pbn == VDO_ZERO_BLOCK) in vdo_get_slab() [all …]
|
| A D | data-vio.h | 85 physical_block_number_t pbn; member 149 physical_block_number_t pbn; member 320 return (data_vio->allocation.pbn != VDO_ZERO_BLOCK); in data_vio_has_allocation() 430 (unsigned long long) data_vio->allocation.pbn, thread_id, in assert_data_vio_in_allocated_zone() 460 (unsigned long long) data_vio->duplicate.pbn, thread_id, in assert_data_vio_in_duplicate_zone() 490 (unsigned long long) data_vio->mapped.pbn, thread_id, expected); in assert_data_vio_in_mapped_zone() 507 (unsigned long long) data_vio->new_mapped.pbn, thread_id, in assert_data_vio_in_new_mapped_zone()
|
| A D | vio.h | 124 blk_opf_t bi_opf, physical_block_number_t pbn); 127 blk_opf_t bi_opf, physical_block_number_t pbn); 129 blk_opf_t bi_opf, physical_block_number_t pbn);
|
| A D | dedupe.c | 512 data_vio->is_duplicate = (source.pbn != VDO_ZERO_BLOCK); in set_duplicate_location() 1204 agent->duplicate.pbn); in start_verifying() 1280 struct vdo_slab *slab = vdo_get_slab(depot, agent->duplicate.pbn); in acquire_provisional_reference() 1290 agent->duplicate.pbn, lock); in acquire_provisional_reference() 1618 advice->pbn = get_unaligned_le64(&encoding->data[offset]); in decode_uds_advice() 1625 (unsigned long long) advice->pbn, advice->state, in decode_uds_advice() 1631 result = vdo_get_physical_zone(vdo, advice->pbn, &advice->zone); in decode_uds_advice() 1634 (unsigned long long) advice->pbn, in decode_uds_advice() 1962 VDO_ASSERT_LOG_ONLY(data_vio->new_mapped.pbn == allocation->pbn, in transfer_allocation_lock() 1965 allocation->pbn = VDO_ZERO_BLOCK; in transfer_allocation_lock() [all …]
|
| A D | physical-zone.h | 100 physical_block_number_t pbn); 103 physical_block_number_t pbn,
|
| A D | types.h | 141 physical_block_number_t pbn; member 166 physical_block_number_t pbn; member
|
| A D | packer.c | 260 .pbn = allocation->pbn, in release_compressed_write_waiter() 306 (unsigned long long) allocation->pbn); in handle_compressed_write_error() 468 REQ_OP_WRITE, agent->allocation.pbn); in write_bin()
|
| A D | recovery-journal.c | 1197 has_decrement = (data_vio->decrement_updater.zpbn.pbn != VDO_ZERO_BLOCK); in continue_committed_waiter() 1198 if ((data_vio->increment_updater.zpbn.pbn != VDO_ZERO_BLOCK) || !has_decrement) in continue_committed_waiter() 1345 .pbn = data_vio->increment_updater.zpbn.pbn, in add_queued_recovery_entries() 1349 .pbn = data_vio->decrement_updater.zpbn.pbn, in add_queued_recovery_entries()
|
| A D | encodings.c | 336 physical_block_number_t pbn, in vdo_format_block_map_page() argument 344 page->header.pbn = __cpu_to_le64(pbn); in vdo_format_block_map_page() 351 physical_block_number_t pbn) in vdo_validate_block_map_page() argument 360 if (pbn != vdo_get_block_map_page_pbn(page)) in vdo_validate_block_map_page()
|
| /drivers/bluetooth/ |
| A D | hci_ag6xx.c | 254 struct pbn_entry *pbn = (void *)fw_ptr; in ag6xx_setup() local 257 if (pbn->addr == 0xffffffff) { in ag6xx_setup() 263 addr = le32_to_cpu(pbn->addr); in ag6xx_setup() 264 plen = le32_to_cpu(pbn->plen); in ag6xx_setup() 266 if (fw->data + fw->size <= pbn->data + plen) { in ag6xx_setup() 274 err = intel_mem_write(hdev, addr, plen, pbn->data); in ag6xx_setup() 280 fw_ptr = pbn->data + plen; in ag6xx_setup()
|
| /drivers/gpu/drm/amd/display/amdgpu_dm/ |
| A D | amdgpu_dm_mst_types.c | 922 vars[i + k].pbn); in set_dsc_configs_from_fairness_vars() 1008 vars[next_index].pbn); in increase_dsc_bpp() 1020 vars[next_index].pbn); in increase_dsc_bpp() 1029 vars[next_index].pbn); in increase_dsc_bpp() 1041 vars[next_index].pbn); in increase_dsc_bpp() 1099 var_pbn = vars[next_index].pbn; in try_disable_dsc() 1104 vars[next_index].pbn); in try_disable_dsc() 1108 vars[next_index].pbn = var_pbn; in try_disable_dsc() 1123 vars[next_index].pbn); in try_disable_dsc() 1245 vars[i + k].pbn); in compute_mst_dsc_configs_for_link() [all …]
|
| A D | amdgpu_dm_mst_types.h | 76 int pbn; member
|
| /drivers/gpu/drm/nouveau/nvif/ |
| A D | outp.c | 30 u8 start_slot, u8 num_slots, u16 pbn, u16 aligned_pbn) in nvif_outp_dp_mst_vcpi() argument 39 args.pbn = pbn; in nvif_outp_dp_mst_vcpi() 45 args.head, args.start_slot, args.num_slots, args.pbn, args.aligned_pbn); in nvif_outp_dp_mst_vcpi()
|
| /drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| A D | tu102.c | 34 tu102_sor_dp_vcpi(struct nvkm_ior *sor, int head, u8 slot, u8 slot_nr, u16 pbn, u16 aligned) in tu102_sor_dp_vcpi() argument 39 nvkm_mask(device, 0x61657c + hoff, 0xffffffff, (aligned << 16) | pbn); in tu102_sor_dp_vcpi()
|
| /drivers/gpu/drm/display/ |
| A D | drm_dp_mst_topology.c | 492 a->pbn = buf[++idx] << 8; in drm_dp_decode_sideband_req() 493 a->pbn |= buf[++idx]; in drm_dp_decode_sideband_req() 1208 req.u.allocate_payload.pbn = pbn; in build_allocate_payload() 3122 int pbn) in drm_dp_payload_send_msg() argument 4456 prev_bw = payload->pbn; in drm_dp_atomic_find_time_slots() 4493 payload->pbn = pbn; in drm_dp_atomic_find_time_slots() 4573 payload->pbn = 0; in drm_dp_atomic_release_time_slots() 4952 payload->pbn, in drm_dp_mst_dump_topology() 5272 if (!payload->pbn || in drm_dp_mst_atomic_check_mstb_bw_limit() 5328 pbn_used = payload->pbn; in drm_dp_mst_atomic_check_port_bw_limit() [all …]
|
| /drivers/gpu/drm/amd/display/dc/link/ |
| A D | link_dpms.c | 1435 struct fixed31_32 pbn; in allocate_mst_payload() local 1513 pbn = get_pbn_from_timing(pipe_ctx); in allocate_mst_payload() 1514 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in allocate_mst_payload() 1744 struct fixed31_32 pbn; in link_reduce_mst_payload() local 1753 pbn = get_pbn_from_bw_in_kbps(bw_in_kbps); in link_reduce_mst_payload() 1754 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in link_reduce_mst_payload() 1832 struct fixed31_32 pbn; in link_increase_mst_payload() local 1899 pbn = get_pbn_from_bw_in_kbps(bw_in_kbps); in link_increase_mst_payload() 1901 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in link_increase_mst_payload()
|