Home
last modified time | relevance | path

Searched refs:blocks (Results 1 – 25 of 253) sorted by relevance

1234567891011

/drivers/mtd/
A Drfd_ftl.c88 struct block *blocks; member
190 if (!part->blocks) in scan_header()
235 kfree(part->blocks); in scan_header()
281 part->blocks[block].free_sectors = 0; in erase_block()
289 part->blocks[block].free_sectors = 0; in erase_block()
298 part->blocks[block].erases++; in erase_block()
345 part->blocks[block_no].offset); in move_block_contents()
535 part->blocks[block].offset); in find_writable_block()
556 addr = part->blocks[block].offset + in mark_sector_deleted()
572 part->blocks[block].used_sectors--; in mark_sector_deleted()
[all …]
/drivers/mfd/
A Dstmpe.c35 unsigned int blocks; member
411 if (blocks & STMPE_BLOCK_GPIO) in stmpe801_enable()
423 .blocks = stmpe801_blocks,
535 if (blocks & STMPE_BLOCK_ADC) in stmpe811_enable()
587 .blocks = stmpe811_blocks,
602 .blocks = stmpe811_blocks,
658 .blocks = stmpe1600_blocks,
822 .blocks = stmpe1601_blocks,
931 .blocks = stmpe1801_blocks,
1037 .blocks = stmpe24xx_blocks,
[all …]
/drivers/media/i2c/adv748x/
A Dadv748x-hdmi.c469 if (edid->start_block == 0 && edid->blocks == 0) { in adv748x_hdmi_get_edid()
470 edid->blocks = hdmi->edid.blocks; in adv748x_hdmi_get_edid()
474 if (edid->start_block >= hdmi->edid.blocks) in adv748x_hdmi_get_edid()
477 if (edid->start_block + edid->blocks > hdmi->edid.blocks) in adv748x_hdmi_get_edid()
478 edid->blocks = hdmi->edid.blocks - edid->start_block; in adv748x_hdmi_get_edid()
481 edid->blocks * 128); in adv748x_hdmi_get_edid()
521 if (edid->blocks == 0) { in adv748x_hdmi_set_edid()
522 hdmi->edid.blocks = 0; in adv748x_hdmi_set_edid()
538 if (edid->blocks > 4) { in adv748x_hdmi_set_edid()
539 edid->blocks = 4; in adv748x_hdmi_set_edid()
[all …]
/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_vram_mgr.c40 struct list_head blocks; member
288 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_bo_visible_size()
322 list_move(&rsv->blocks, &mgr->reserved_pages); in amdgpu_vram_mgr_do_reserve()
345 INIT_LIST_HEAD(&rsv->blocks); in amdgpu_vram_mgr_reserve_range()
500 INIT_LIST_HEAD(&vres->blocks); in amdgpu_vram_mgr_new()
550 &vres->blocks, in amdgpu_vram_mgr_new()
584 &vres->blocks); in amdgpu_vram_mgr_new()
620 drm_buddy_free_list(mm, &vres->blocks, 0); in amdgpu_vram_mgr_new()
648 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_del()
820 list_for_each_entry(block, &mgr->blocks, link) { in amdgpu_vram_mgr_intersects()
[all …]
/drivers/iio/buffer/
A Dindustrialio-buffer-dma.c359 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update()
376 if (queue->fileio.blocks[i]) { in iio_dma_buffer_request_update()
377 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update()
395 queue->fileio.blocks[i] = block; in iio_dma_buffer_request_update()
431 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free()
440 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free()
442 iio_buffer_block_put(queue->fileio.blocks[i]); in iio_dma_buffer_fileio_free()
443 queue->fileio.blocks[i] = NULL; in iio_dma_buffer_fileio_free()
557 block = queue->fileio.blocks[idx]; in iio_dma_buffer_dequeue()
560 idx = (idx + 1) % ARRAY_SIZE(queue->fileio.blocks); in iio_dma_buffer_dequeue()
[all …]
/drivers/net/ethernet/microchip/sparx5/lan969x/
A Dlan969x_vcap_impl.c16 .blocks = 1,
28 .blocks = 1,
40 .blocks = 1,
52 .blocks = 1,
64 .blocks = 1,
/drivers/mmc/core/
A Dsdio_ops.c115 unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) in mmc_io_rw_extended() argument
140 if (blocks == 0) in mmc_io_rw_extended()
143 cmd.arg |= 0x08000000 | blocks; /* block mode */ in mmc_io_rw_extended()
148 data.blocks = blocks ? blocks : 1; in mmc_io_rw_extended()
151 left_size = data.blksz * data.blocks; in mmc_io_rw_extended()
A Dmmc_test.c81 unsigned int blocks; member
202 mrq->sbc->arg = blocks; in mmc_test_prepare_sbc()
216 if (blocks > 1) { in mmc_test_prepare_mrq()
230 if (blocks == 1) in mmc_test_prepare_mrq()
239 mrq->data->blocks = blocks; in mmc_test_prepare_mrq()
678 if (mrq->data->blocks > 1) { in mmc_test_prepare_broken_mrq()
860 dev_addr += t->blocks; in mmc_test_nonblock_transfer()
887 blocks, blksz, write); in mmc_test_simple_transfer()
949 blocks, blksz, write); in mmc_test_transfer()
1390 t->blocks = sz >> 9; in mmc_test_area_map()
[all …]
/drivers/md/
A Dmd-bitmap.c1519 sector_t blocks; in bitmap_daemon_work() local
1694 sector_t blocks; in bitmap_start_write() local
1731 offset += blocks; in bitmap_start_write()
1748 sector_t blocks; in bitmap_end_write() local
1794 *blocks = 1024; in __bitmap_start_sync()
1831 *blocks = 0; in bitmap_start_sync()
1849 *blocks = 1024; in __bitmap_end_sync()
1886 sector_t blocks; in bitmap_close_sync() local
1902 sector_t blocks; in bitmap_cond_end_sync() local
1923 s += blocks; in bitmap_cond_end_sync()
[all …]
A Ddm-ebs-target.c145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() local
148 blocks = __nr_blocks(ec, bio); in __ebs_discard_bio()
156 blocks--; in __ebs_discard_bio()
160 if (blocks && __block_mod(bio_end_sector(bio), ec->u_bs)) in __ebs_discard_bio()
161 blocks--; in __ebs_discard_bio()
163 return blocks ? dm_bufio_issue_discard(ec->bufio, block, blocks) : 0; in __ebs_discard_bio()
169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() local
171 blocks = __nr_blocks(ec, bio); in __ebs_forget_bio()
173 dm_bufio_forget_buffers(ec->bufio, __sector_to_block(ec, sector), blocks); in __ebs_forget_bio()
/drivers/usb/storage/
A Djumpshot.c478 unsigned long block, blocks; in jumpshot_transport() local
533 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport()
536 block, blocks); in jumpshot_transport()
537 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport()
550 block, blocks); in jumpshot_transport()
551 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport()
558 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport()
561 block, blocks); in jumpshot_transport()
562 return jumpshot_write_data(us, info, block, blocks); in jumpshot_transport()
575 block, blocks); in jumpshot_transport()
[all …]
A Ddatafab.c551 unsigned long block, blocks; in datafab_transport() local
604 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in datafab_transport()
607 block, blocks); in datafab_transport()
608 return datafab_read_data(us, info, block, blocks); in datafab_transport()
621 block, blocks); in datafab_transport()
622 return datafab_read_data(us, info, block, blocks); in datafab_transport()
629 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in datafab_transport()
632 block, blocks); in datafab_transport()
633 return datafab_write_data(us, info, block, blocks); in datafab_transport()
646 block, blocks); in datafab_transport()
[all …]
/drivers/md/dm-vdo/
A Derrors.c83 struct error_block blocks[MAX_ERROR_BLOCKS]; member
87 .blocks = { {
106 for (block = registered_errors.blocks; in get_error_info()
107 block < registered_errors.blocks + registered_errors.count; in get_error_info()
294 for (block = registered_errors.blocks; in uds_register_error_block()
295 block < registered_errors.blocks + registered_errors.count; in uds_register_error_block()
305 registered_errors.blocks[registered_errors.count++] = new_block; in uds_register_error_block()
/drivers/net/ethernet/mellanox/mlxsw/
A Dcore_acl_flex_keys.c56 const struct mlxsw_afk_block *blocks; member
66 const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; in mlxsw_afk_blocks_check()
95 mlxsw_afk->blocks = ops->blocks; in mlxsw_afk_create()
117 const struct mlxsw_afk_block *blocks[]; member
154 const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; in mlxsw_afk_picker_count_hits()
219 key_info->blocks[key_info->blocks_count] = in mlxsw_afk_picker_key_info_add()
220 &mlxsw_afk->blocks[block_index]; in mlxsw_afk_picker_key_info_add()
234 if (!mlxsw_afk->blocks[i].high_entropy) in mlxsw_afk_keys_fill()
330 key_info = kzalloc(struct_size(key_info, blocks, mlxsw_afk->max_blocks), in mlxsw_afk_key_info_create()
410 block = key_info->blocks[block_index]; in mlxsw_afk_key_info_elinst_get()
[all …]
/drivers/gpu/drm/
A Ddrm_buddy.c737 struct list_head *blocks, in __alloc_range() argument
833 struct list_head *blocks) in __drm_buddy_alloc_range() argument
848 struct list_head *blocks) in __alloc_contig_try_harder() argument
873 &filled, blocks); in __alloc_contig_try_harder()
886 list_splice(&blocks_lhs, blocks); in __alloc_contig_try_harder()
921 struct list_head *blocks) in drm_buddy_block_trim() argument
930 if (!list_is_singular(blocks)) in drm_buddy_block_trim()
933 block = list_first_entry(blocks, in drm_buddy_block_trim()
983 list_add(&block->link, blocks); in drm_buddy_block_trim()
1029 struct list_head *blocks, in drm_buddy_alloc_blocks() argument
[all …]
/drivers/mtd/tests/
A Dspeedtest.c43 static int multiblock_erase(int ebnum, int blocks) in multiblock_erase() argument
51 ei.len = mtd->erasesize * blocks; in multiblock_erase()
56 err, ebnum, blocks); in multiblock_erase()
175 int err, i, blocks, j, k; in mtd_speedtest_init() local
371 blocks = 1 << k; in mtd_speedtest_init()
373 blocks); in mtd_speedtest_init()
376 for (j = 0; j < blocks && (i + j) < ebcnt; j++) in mtd_speedtest_init()
396 blocks, speed); in mtd_speedtest_init()
/drivers/crypto/
A Dpadlock-sha.c248 int blocks = len / SHA1_BLOCK_SIZE; in padlock_sha1_update_nano() local
250 len -= blocks * SHA1_BLOCK_SIZE; in padlock_sha1_update_nano()
251 state->count += blocks * SHA1_BLOCK_SIZE; in padlock_sha1_update_nano()
257 "c"((unsigned long)blocks)); in padlock_sha1_update_nano()
266 int blocks = len / SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() local
268 len -= blocks * SHA256_BLOCK_SIZE; in padlock_sha256_update_nano()
269 state->count += blocks * SHA256_BLOCK_SIZE; in padlock_sha256_update_nano()
275 "c"((unsigned long)blocks)); in padlock_sha256_update_nano()
/drivers/gpu/drm/i915/display/
A Dskl_watermark.c1864 blocks = max(blocks, in skl_compute_plane_wm()
1889 if (result_prev->blocks > blocks) in skl_compute_plane_wm()
1890 blocks = result_prev->blocks; in skl_compute_plane_wm()
1926 result->blocks = blocks; in skl_compute_plane_wm()
2035 trans_wm->blocks = blocks; in skl_compute_transition_wm()
2394 l1->blocks == l2->blocks && in skl_wm_level_equals()
2676 old_wm->wm[0].blocks, old_wm->wm[1].blocks, in skl_print_plane_changes()
2677 old_wm->wm[2].blocks, old_wm->wm[3].blocks, in skl_print_plane_changes()
2678 old_wm->wm[4].blocks, old_wm->wm[5].blocks, in skl_print_plane_changes()
2679 old_wm->wm[6].blocks, old_wm->wm[7].blocks, in skl_print_plane_changes()
[all …]
/drivers/gpu/drm/tests/
A Ddrm_buddy_test.c507 LIST_HEAD(blocks); in drm_test_buddy_alloc_pathological()
543 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pathological()
575 list_splice_tail(&holes, &blocks); in drm_test_buddy_alloc_pathological()
576 drm_buddy_free_list(&mm, &blocks, 0); in drm_test_buddy_alloc_pathological()
588 LIST_HEAD(blocks); in drm_test_buddy_alloc_pessimistic()
613 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pessimistic()
625 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_pessimistic()
671 drm_buddy_free_list(&mm, &blocks, 0); in drm_test_buddy_alloc_pessimistic()
682 LIST_HEAD(blocks); in drm_test_buddy_alloc_optimistic()
708 list_move_tail(&block->link, &blocks); in drm_test_buddy_alloc_optimistic()
[all …]
/drivers/soc/ixp4xx/
A Dixp4xx-npe.c523 DECLARE_FLEX_ARRAY(struct dl_block, blocks); in npe_load_firmware()
533 int i, j, err, data_size, instr_size, blocks, table_end; in npe_load_firmware() local
604 for (blocks = 0; blocks * sizeof(struct dl_block) / 4 < image->size; in npe_load_firmware()
605 blocks++) in npe_load_firmware()
606 if (image->blocks[blocks].type == FW_BLOCK_TYPE_EOF) in npe_load_firmware()
608 if (blocks * sizeof(struct dl_block) / 4 >= image->size) { in npe_load_firmware()
615 print_npe(KERN_DEBUG, npe, "%i firmware blocks found\n", blocks); in npe_load_firmware()
618 table_end = blocks * sizeof(struct dl_block) / 4 + 1 /* EOF marker */; in npe_load_firmware()
619 for (i = 0, blk = image->blocks; i < blocks; i++, blk++) { in npe_load_firmware()
/drivers/gpu/drm/xe/
A Dxe_ttm_vram_mgr.c79 INIT_LIST_HEAD(&vres->blocks); in xe_ttm_vram_mgr_new()
123 min_page_size, &vres->blocks, vres->flags); in xe_ttm_vram_mgr_new()
128 if (!drm_buddy_block_trim(mm, NULL, vres->base.size, &vres->blocks)) in xe_ttm_vram_mgr_new()
137 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_new()
153 xe_is_vram_mgr_blocks_contiguous(mm, &vres->blocks)) in xe_ttm_vram_mgr_new()
161 struct drm_buddy_block *block = list_first_entry(&vres->blocks, in xe_ttm_vram_mgr_new()
190 drm_buddy_free_list(mm, &vres->blocks, 0); in xe_ttm_vram_mgr_del()
235 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_intersects()
265 list_for_each_entry(block, &vres->blocks, link) { in xe_ttm_vram_mgr_compatible()
/drivers/mmc/host/
A Dcb710-mmc.c263 return !(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8)); in cb710_is_transfer_size_supported()
269 size_t len, blocks = data->blocks; in cb710_mmc_receive() local
274 if (unlikely(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8))) in cb710_mmc_receive()
284 while (blocks-- > 0) { in cb710_mmc_receive()
313 size_t len, blocks = data->blocks; in cb710_mmc_send() local
318 if (unlikely(data->blocks > 1 && data->blksz & 15)) in cb710_mmc_send()
326 while (blocks-- > 0) { in cb710_mmc_send()
440 data->bytes_xfered = data->blksz * data->blocks; in cb710_mmc_transfer_data()
458 cb710_mmc_set_transfer_size(slot, data->blocks, data->blksz); in cb710_mmc_command()
/drivers/block/
A Dataflop.c1182 if (SUDT[-1].blocks > ReqBlock) { in fd_rwsec_done1()
1186 SUDT->blocks); in fd_rwsec_done1()
1201 SUDT->blocks); in fd_rwsec_done1()
1474 if (default_params[drive].blocks == 0) in floppy_revalidate()
1619 getprm.size = dtp->blocks; in fd_locked_ioctl()
1681 if ( dtp->blocks == setprm.size in fd_locked_ioctl()
1688 set_capacity(disk, UDT->blocks); in fd_locked_ioctl()
1694 default_params[drive].blocks = dtp->blocks; in fd_locked_ioctl()
1714 dtp->blocks = setprm.size; in fd_locked_ioctl()
1734 set_capacity(disk, UDT->blocks); in fd_locked_ioctl()
[all …]
/drivers/gpu/drm/i915/
A Di915_ttm_buddy_manager.c56 INIT_LIST_HEAD(&bman_res->blocks); in i915_ttm_buddy_man_alloc()
96 &bman_res->blocks, in i915_ttm_buddy_man_alloc()
106 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_alloc()
129 drm_buddy_free_list(mm, &bman_res->blocks, 0); in i915_ttm_buddy_man_alloc()
144 drm_buddy_free_list(&bman->mm, &bman_res->blocks, 0); in i915_ttm_buddy_man_free()
177 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_intersects()
210 list_for_each_entry(block, &bman_res->blocks, link) { in i915_ttm_buddy_man_compatible()
/drivers/gpu/drm/msm/
A DNOTES4 display controller blocks at play:
12 Plus a handful of blocks around them for HDMI/DSI/etc output.
18 But, HDMI/DSI/etc blocks seem like they can be shared across multiple
19 display controller blocks. And I for sure don't want to have to deal
60 Also unlike MDP4, with MDP5 all the IRQs for other blocks (HDMI, DSI,
64 which blocks need to be allocated to the active pipes based on fetch

Completed in 104 milliseconds

1234567891011