| /drivers/mtd/ |
| A D | rfd_ftl.c | 88 struct block *blocks; member 190 if (!part->blocks) in scan_header() 235 kfree(part->blocks); in scan_header() 281 part->blocks[block].free_sectors = 0; in erase_block() 289 part->blocks[block].free_sectors = 0; in erase_block() 298 part->blocks[block].erases++; in erase_block() 345 part->blocks[block_no].offset); in move_block_contents() 535 part->blocks[block].offset); in find_writable_block() 556 addr = part->blocks[block].offset + in mark_sector_deleted() 572 part->blocks[block].used_sectors--; in mark_sector_deleted() [all …]
|
| /drivers/mfd/ |
| A D | stmpe.c | 35 unsigned int blocks; member 411 if (blocks & STMPE_BLOCK_GPIO) in stmpe801_enable() 423 .blocks = stmpe801_blocks, 535 if (blocks & STMPE_BLOCK_ADC) in stmpe811_enable() 587 .blocks = stmpe811_blocks, 602 .blocks = stmpe811_blocks, 658 .blocks = stmpe1600_blocks, 822 .blocks = stmpe1601_blocks, 931 .blocks = stmpe1801_blocks, 1037 .blocks = stmpe24xx_blocks, [all …]
|
| /drivers/media/i2c/adv748x/ |
| A D | adv748x-hdmi.c | 469 if (edid->start_block == 0 && edid->blocks == 0) { in adv748x_hdmi_get_edid() 470 edid->blocks = hdmi->edid.blocks; in adv748x_hdmi_get_edid() 474 if (edid->start_block >= hdmi->edid.blocks) in adv748x_hdmi_get_edid() 477 if (edid->start_block + edid->blocks > hdmi->edid.blocks) in adv748x_hdmi_get_edid() 478 edid->blocks = hdmi->edid.blocks - edid->start_block; in adv748x_hdmi_get_edid() 481 edid->blocks * 128); in adv748x_hdmi_get_edid() 521 if (edid->blocks == 0) { in adv748x_hdmi_set_edid() 522 hdmi->edid.blocks = 0; in adv748x_hdmi_set_edid() 538 if (edid->blocks > 4) { in adv748x_hdmi_set_edid() 539 edid->blocks = 4; in adv748x_hdmi_set_edid() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_vram_mgr.c | 40 struct list_head blocks; member 288 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_bo_visible_size() 322 list_move(&rsv->blocks, &mgr->reserved_pages); in amdgpu_vram_mgr_do_reserve() 345 INIT_LIST_HEAD(&rsv->blocks); in amdgpu_vram_mgr_reserve_range() 500 INIT_LIST_HEAD(&vres->blocks); in amdgpu_vram_mgr_new() 550 &vres->blocks, in amdgpu_vram_mgr_new() 584 &vres->blocks); in amdgpu_vram_mgr_new() 620 drm_buddy_free_list(mm, &vres->blocks, 0); in amdgpu_vram_mgr_new() 648 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_del() 820 list_for_each_entry(block, &mgr->blocks, link) { in amdgpu_vram_mgr_intersects() [all …]
|
| /drivers/iio/buffer/ |
| A D | industrialio-buffer-dma.c | 359 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 376 if (queue->fileio.blocks[i]) { in iio_dma_buffer_request_update() 377 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 395 queue->fileio.blocks[i] = block; in iio_dma_buffer_request_update() 431 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free() 440 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free() 442 iio_buffer_block_put(queue->fileio.blocks[i]); in iio_dma_buffer_fileio_free() 443 queue->fileio.blocks[i] = NULL; in iio_dma_buffer_fileio_free() 557 block = queue->fileio.blocks[idx]; in iio_dma_buffer_dequeue() 560 idx = (idx + 1) % ARRAY_SIZE(queue->fileio.blocks); in iio_dma_buffer_dequeue() [all …]
|
| /drivers/net/ethernet/microchip/sparx5/lan969x/ |
| A D | lan969x_vcap_impl.c | 16 .blocks = 1, 28 .blocks = 1, 40 .blocks = 1, 52 .blocks = 1, 64 .blocks = 1,
|
| /drivers/mmc/core/ |
| A D | sdio_ops.c | 115 unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) in mmc_io_rw_extended() argument 140 if (blocks == 0) in mmc_io_rw_extended() 143 cmd.arg |= 0x08000000 | blocks; /* block mode */ in mmc_io_rw_extended() 148 data.blocks = blocks ? blocks : 1; in mmc_io_rw_extended() 151 left_size = data.blksz * data.blocks; in mmc_io_rw_extended()
|
| A D | mmc_test.c | 81 unsigned int blocks; member 202 mrq->sbc->arg = blocks; in mmc_test_prepare_sbc() 216 if (blocks > 1) { in mmc_test_prepare_mrq() 230 if (blocks == 1) in mmc_test_prepare_mrq() 239 mrq->data->blocks = blocks; in mmc_test_prepare_mrq() 678 if (mrq->data->blocks > 1) { in mmc_test_prepare_broken_mrq() 860 dev_addr += t->blocks; in mmc_test_nonblock_transfer() 887 blocks, blksz, write); in mmc_test_simple_transfer() 949 blocks, blksz, write); in mmc_test_transfer() 1390 t->blocks = sz >> 9; in mmc_test_area_map() [all …]
|
| /drivers/md/ |
| A D | md-bitmap.c | 1519 sector_t blocks; in bitmap_daemon_work() local 1694 sector_t blocks; in bitmap_start_write() local 1731 offset += blocks; in bitmap_start_write() 1748 sector_t blocks; in bitmap_end_write() local 1794 *blocks = 1024; in __bitmap_start_sync() 1831 *blocks = 0; in bitmap_start_sync() 1849 *blocks = 1024; in __bitmap_end_sync() 1886 sector_t blocks; in bitmap_close_sync() local 1902 sector_t blocks; in bitmap_cond_end_sync() local 1923 s += blocks; in bitmap_cond_end_sync() [all …]
|
| A D | dm-ebs-target.c | 145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() local 148 blocks = __nr_blocks(ec, bio); in __ebs_discard_bio() 156 blocks--; in __ebs_discard_bio() 160 if (blocks && __block_mod(bio_end_sector(bio), ec->u_bs)) in __ebs_discard_bio() 161 blocks--; in __ebs_discard_bio() 163 return blocks ? dm_bufio_issue_discard(ec->bufio, block, blocks) : 0; in __ebs_discard_bio() 169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() local 171 blocks = __nr_blocks(ec, bio); in __ebs_forget_bio() 173 dm_bufio_forget_buffers(ec->bufio, __sector_to_block(ec, sector), blocks); in __ebs_forget_bio()
|
| /drivers/usb/storage/ |
| A D | jumpshot.c | 478 unsigned long block, blocks; in jumpshot_transport() local 533 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport() 536 block, blocks); in jumpshot_transport() 537 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport() 550 block, blocks); in jumpshot_transport() 551 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport() 558 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport() 561 block, blocks); in jumpshot_transport() 562 return jumpshot_write_data(us, info, block, blocks); in jumpshot_transport() 575 block, blocks); in jumpshot_transport() [all …]
|
| A D | datafab.c | 551 unsigned long block, blocks; in datafab_transport() local 604 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in datafab_transport() 607 block, blocks); in datafab_transport() 608 return datafab_read_data(us, info, block, blocks); in datafab_transport() 621 block, blocks); in datafab_transport() 622 return datafab_read_data(us, info, block, blocks); in datafab_transport() 629 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in datafab_transport() 632 block, blocks); in datafab_transport() 633 return datafab_write_data(us, info, block, blocks); in datafab_transport() 646 block, blocks); in datafab_transport() [all …]
|
| /drivers/md/dm-vdo/ |
| A D | errors.c | 83 struct error_block blocks[MAX_ERROR_BLOCKS]; member 87 .blocks = { { 106 for (block = registered_errors.blocks; in get_error_info() 107 block < registered_errors.blocks + registered_errors.count; in get_error_info() 294 for (block = registered_errors.blocks; in uds_register_error_block() 295 block < registered_errors.blocks + registered_errors.count; in uds_register_error_block() 305 registered_errors.blocks[registered_errors.count++] = new_block; in uds_register_error_block()
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | core_acl_flex_keys.c | 56 const struct mlxsw_afk_block *blocks; member 66 const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; in mlxsw_afk_blocks_check() 95 mlxsw_afk->blocks = ops->blocks; in mlxsw_afk_create() 117 const struct mlxsw_afk_block *blocks[]; member 154 const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; in mlxsw_afk_picker_count_hits() 219 key_info->blocks[key_info->blocks_count] = in mlxsw_afk_picker_key_info_add() 220 &mlxsw_afk->blocks[block_index]; in mlxsw_afk_picker_key_info_add() 234 if (!mlxsw_afk->blocks[i].high_entropy) in mlxsw_afk_keys_fill() 330 key_info = kzalloc(struct_size(key_info, blocks, mlxsw_afk->max_blocks), in mlxsw_afk_key_info_create() 410 block = key_info->blocks[block_index]; in mlxsw_afk_key_info_elinst_get() [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_buddy.c | 737 struct list_head *blocks, in __alloc_range() argument 833 struct list_head *blocks) in __drm_buddy_alloc_range() argument 848 struct list_head *blocks) in __alloc_contig_try_harder() argument 873 &filled, blocks); in __alloc_contig_try_harder() 886 list_splice(&blocks_lhs, blocks); in __alloc_contig_try_harder() 921 struct list_head *blocks) in drm_buddy_block_trim() argument 930 if (!list_is_singular(blocks)) in drm_buddy_block_trim() 933 block = list_first_entry(blocks, in drm_buddy_block_trim() 983 list_add(&block->link, blocks); in drm_buddy_block_trim() 1029 struct list_head *blocks, in drm_buddy_alloc_blocks() argument [all …]
|
| /drivers/mtd/tests/ |
| A D | speedtest.c | 43 static int multiblock_erase(int ebnum, int blocks) in multiblock_erase() argument 51 ei.len = mtd->erasesize * blocks; in multiblock_erase() 56 err, ebnum, blocks); in multiblock_erase() 175 int err, i, blocks, j, k; in mtd_speedtest_init() local 371 blocks = 1 << k; in mtd_speedtest_init() 373 blocks); in mtd_speedtest_init() 376 for (j = 0; j < blocks && (i + j) < ebcnt; j++) in mtd_speedtest_init() 396 blocks, speed); in mtd_speedtest_init()
|
| /drivers/crypto/ |
| A D | padlock-sha.c | 248 int blocks = len / SHA1_BLOCK_SIZE; in padlock_sha1_update_nano() local 250 len -= blocks * SHA1_BLOCK_SIZE; in padlock_sha1_update_nano() 251 state->count += blocks * SHA1_BLOCK_SIZE; in padlock_sha1_update_nano() 257 "c"((unsigned long)blocks)); in padlock_sha1_update_nano() 266 int blocks = len / SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() local 268 len -= blocks * SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() 269 state->count += blocks * SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() 275 "c"((unsigned long)blocks)); in padlock_sha256_update_nano()
|
| /drivers/gpu/drm/i915/display/ |
| A D | skl_watermark.c | 1864 blocks = max(blocks, in skl_compute_plane_wm() 1889 if (result_prev->blocks > blocks) in skl_compute_plane_wm() 1890 blocks = result_prev->blocks; in skl_compute_plane_wm() 1926 result->blocks = blocks; in skl_compute_plane_wm() 2035 trans_wm->blocks = blocks; in skl_compute_transition_wm() 2394 l1->blocks == l2->blocks && in skl_wm_level_equals() 2676 old_wm->wm[0].blocks, old_wm->wm[1].blocks, in skl_print_plane_changes() 2677 old_wm->wm[2].blocks, old_wm->wm[3].blocks, in skl_print_plane_changes() 2678 old_wm->wm[4].blocks, old_wm->wm[5].blocks, in skl_print_plane_changes() 2679 old_wm->wm[6].blocks, old_wm->wm[7].blocks, in skl_print_plane_changes() [all …]
|
| /drivers/gpu/drm/tests/ |
| A D | drm_buddy_test.c | 507 LIST_HEAD(blocks); in drm_test_buddy_alloc_pathological() 543 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pathological() 575 list_splice_tail(&holes, &blocks); in drm_test_buddy_alloc_pathological() 576 drm_buddy_free_list(&mm, &blocks, 0); in drm_test_buddy_alloc_pathological() 588 LIST_HEAD(blocks); in drm_test_buddy_alloc_pessimistic() 613 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pessimistic() 625 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pessimistic() 671 drm_buddy_free_list(&mm, &blocks, 0); in drm_test_buddy_alloc_pessimistic() 682 LIST_HEAD(blocks); in drm_test_buddy_alloc_optimistic() 708 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_optimistic() [all …]
|
| /drivers/soc/ixp4xx/ |
| A D | ixp4xx-npe.c | 523 DECLARE_FLEX_ARRAY(struct dl_block, blocks); in npe_load_firmware() 533 int i, j, err, data_size, instr_size, blocks, table_end; in npe_load_firmware() local 604 for (blocks = 0; blocks * sizeof(struct dl_block) / 4 < image->size; in npe_load_firmware() 605 blocks++) in npe_load_firmware() 606 if (image->blocks[blocks].type == FW_BLOCK_TYPE_EOF) in npe_load_firmware() 608 if (blocks * sizeof(struct dl_block) / 4 >= image->size) { in npe_load_firmware() 615 print_npe(KERN_DEBUG, npe, "%i firmware blocks found\n", blocks); in npe_load_firmware() 618 table_end = blocks * sizeof(struct dl_block) / 4 + 1 /* EOF marker */; in npe_load_firmware() 619 for (i = 0, blk = image->blocks; i < blocks; i++, blk++) { in npe_load_firmware()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_ttm_vram_mgr.c | 79 INIT_LIST_HEAD(&vres->blocks); in xe_ttm_vram_mgr_new() 123 min_page_size, &vres->blocks, vres->flags); in xe_ttm_vram_mgr_new() 128 if (!drm_buddy_block_trim(mm, NULL, vres->base.size, &vres->blocks)) in xe_ttm_vram_mgr_new() 137 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_new() 153 xe_is_vram_mgr_blocks_contiguous(mm, &vres->blocks)) in xe_ttm_vram_mgr_new() 161 struct drm_buddy_block *block = list_first_entry(&vres->blocks, in xe_ttm_vram_mgr_new() 190 drm_buddy_free_list(mm, &vres->blocks, 0); in xe_ttm_vram_mgr_del() 235 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_intersects() 265 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_compatible()
|
| /drivers/mmc/host/ |
| A D | cb710-mmc.c | 263 return !(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8)); in cb710_is_transfer_size_supported() 269 size_t len, blocks = data->blocks; in cb710_mmc_receive() local 274 if (unlikely(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8))) in cb710_mmc_receive() 284 while (blocks-- > 0) { in cb710_mmc_receive() 313 size_t len, blocks = data->blocks; in cb710_mmc_send() local 318 if (unlikely(data->blocks > 1 && data->blksz & 15)) in cb710_mmc_send() 326 while (blocks-- > 0) { in cb710_mmc_send() 440 data->bytes_xfered = data->blksz * data->blocks; in cb710_mmc_transfer_data() 458 cb710_mmc_set_transfer_size(slot, data->blocks, data->blksz); in cb710_mmc_command()
|
| /drivers/block/ |
| A D | ataflop.c | 1182 if (SUDT[-1].blocks > ReqBlock) { in fd_rwsec_done1() 1186 SUDT->blocks); in fd_rwsec_done1() 1201 SUDT->blocks); in fd_rwsec_done1() 1474 if (default_params[drive].blocks == 0) in floppy_revalidate() 1619 getprm.size = dtp->blocks; in fd_locked_ioctl() 1681 if ( dtp->blocks == setprm.size in fd_locked_ioctl() 1688 set_capacity(disk, UDT->blocks); in fd_locked_ioctl() 1694 default_params[drive].blocks = dtp->blocks; in fd_locked_ioctl() 1714 dtp->blocks = setprm.size; in fd_locked_ioctl() 1734 set_capacity(disk, UDT->blocks); in fd_locked_ioctl() [all …]
|
| /drivers/gpu/drm/i915/ |
| A D | i915_ttm_buddy_manager.c | 56 INIT_LIST_HEAD(&bman_res->blocks); in i915_ttm_buddy_man_alloc() 96 &bman_res->blocks, in i915_ttm_buddy_man_alloc() 106 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_alloc() 129 drm_buddy_free_list(mm, &bman_res->blocks, 0); in i915_ttm_buddy_man_alloc() 144 drm_buddy_free_list(&bman->mm, &bman_res->blocks, 0); in i915_ttm_buddy_man_free() 177 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_intersects() 210 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_compatible()
|
| /drivers/gpu/drm/msm/ |
| A D | NOTES | 4 display controller blocks at play: 12 Plus a handful of blocks around them for HDMI/DSI/etc output. 18 But, HDMI/DSI/etc blocks seem like they can be shared across multiple 19 display controller blocks. And I for sure don't want to have to deal 60 Also unlike MDP4, with MDP5 all the IRQs for other blocks (HDMI, DSI, 64 which blocks need to be allocated to the active pipes based on fetch
|