| /drivers/mtd/ |
| A D | rfd_ftl.c | 53 #define SECTOR_SIZE 512 macro 158 sectors_per_block = part->block_size / SECTOR_SIZE; in scan_header() 167 sizeof(u16) + SECTOR_SIZE - 1) / SECTOR_SIZE; in scan_header() 254 if (!rc && retlen != SECTOR_SIZE) in rfd_ftl_readsect() 263 memset(buf, 0, SECTOR_SIZE); in rfd_ftl_readsect() 328 sector_data = kmalloc(SECTOR_SIZE, GFP_KERNEL); in move_block_contents() 379 if (!rc && retlen != SECTOR_SIZE) in move_block_contents() 633 if (!rc && retlen != SECTOR_SIZE) in do_writesect() 688 for (i=0; i<SECTOR_SIZE; i++) { in rfd_ftl_writesect() 698 if (i == SECTOR_SIZE) in rfd_ftl_writesect() [all …]
|
| A D | ftl.c | 106 #define SECTOR_SIZE 512 macro 431 u_char buf[SECTOR_SIZE]; in copy_erase_unit() 512 src += SECTOR_SIZE; in copy_erase_unit() 513 dest += SECTOR_SIZE; in copy_erase_unit() 771 memset(buffer, 0, SECTOR_SIZE); in ftl_read() 783 buffer += SECTOR_SIZE; in ftl_read() 809 blk = (log_addr % bsize) / SECTOR_SIZE; in set_bam_entry() 910 blk * SECTOR_SIZE); in ftl_write() 936 buffer += SECTOR_SIZE; in ftl_write() 937 virt_addr += SECTOR_SIZE; in ftl_write() [all …]
|
| A D | ssfdc.c | 34 #define SECTOR_SIZE 512 macro 111 sect_buf = kmalloc(SECTOR_SIZE, GFP_KERNEL); in get_valid_cis_sector() 122 ret = mtd_read(mtd, offset, SECTOR_SIZE, &retlen, in get_valid_cis_sector() 126 if (ret < 0 || retlen != SECTOR_SIZE) { in get_valid_cis_sector() 155 ret = mtd_read(mtd, offset, SECTOR_SIZE, &retlen, sect_buf); in read_physical_sector() 156 if (ret < 0 || retlen != SECTOR_SIZE) in read_physical_sector() 400 memset(buf, 0xff, SECTOR_SIZE); in ssfdcr_readsect() 430 .blksize = SECTOR_SIZE,
|
| /drivers/mtd/nand/raw/ |
| A D | diskonchip.c | 116 #define SECTOR_SIZE 512 macro 118 #define NB_DATA (((SECTOR_SIZE + 1) * 8 + 6) / 10) 195 if ((index >= 0 && index < SECTOR_SIZE) || index == (SECTOR_SIZE + 1)) { in doc_ecc_decode() 198 if (index < SECTOR_SIZE) in doc_ecc_decode() 205 if ((index >= 0 && index < SECTOR_SIZE) || index == (SECTOR_SIZE + 1)) { in doc_ecc_decode() 208 if (index < SECTOR_SIZE) in doc_ecc_decode()
|
| /drivers/scsi/ |
| A D | sd_zbc.c | 188 bufsize = roundup((nr_zones + 1) * 64, SECTOR_SIZE); in sd_zbc_alloc_report_buffer() 194 while (bufsize >= SECTOR_SIZE) { in sd_zbc_alloc_report_buffer() 200 bufsize = rounddown(bufsize >> 1, SECTOR_SIZE); in sd_zbc_alloc_report_buffer()
|
| A D | aha1542.c | 1021 SECTOR_SIZE * AHA1542_MAX_SECTORS, in aha1542_init_cmd_priv() 1032 dma_free_coherent(shost->dma_dev, SECTOR_SIZE * AHA1542_MAX_SECTORS, in aha1542_exit_cmd_priv()
|
| /drivers/infiniband/ulp/iser/ |
| A D | iscsi_iser.h | 101 ((ISER_DEF_MAX_SECTORS * SECTOR_SIZE) >> ilog2(SZ_4K)) 103 #define ISCSI_ISER_MAX_SG_TABLESIZE ((32768 * SECTOR_SIZE) >> ilog2(SZ_4K))
|
| A D | iser_verbs.c | 512 sg_tablesize = DIV_ROUND_UP(max_sectors * SECTOR_SIZE, SZ_4K); in iser_calc_scsi_params()
|
| /drivers/vdpa/vdpa_sim/ |
| A D | vdpa_sim_blk.c | 378 blk_config->blk_size = cpu_to_vdpasim32(vdpasim, SECTOR_SIZE); in vdpasim_blk_get_config() 381 cpu_to_vdpasim32(vdpasim, SECTOR_SIZE); in vdpasim_blk_get_config()
|
| /drivers/dax/ |
| A D | super.c | 91 *start_off = get_start_sect(bdev) * SECTOR_SIZE; in fs_dax_get_by_bdev() 92 part_size = bdev_nr_sectors(bdev) * SECTOR_SIZE; in fs_dax_get_by_bdev()
|
| /drivers/block/ |
| A D | loop.c | 465 return SECTOR_SIZE; in loop_query_min_dio_size() 934 return SECTOR_SIZE; in loop_default_blocksize() 1126 lim.logical_block_size = SECTOR_SIZE; in __loop_clr_fd() 1127 lim.physical_block_size = SECTOR_SIZE; in __loop_clr_fd() 1128 lim.io_min = SECTOR_SIZE; in __loop_clr_fd()
|
| A D | zloop.c | 765 zlo->block_size = SECTOR_SIZE; in zloop_get_block_size()
|
| A D | virtio_blk.c | 1435 .logical_block_size = SECTOR_SIZE, in virtblk_probe()
|
| A D | rbd.c | 4939 size = (sector_t)rbd_dev->mapping.size / SECTOR_SIZE; in rbd_dev_update_size() 6278 if (result.uint_32 < SECTOR_SIZE) in rbd_parse_param() 6806 set_capacity(rbd_dev->disk, rbd_dev->mapping.size / SECTOR_SIZE); in rbd_dev_device_setup()
|
| /drivers/block/rnbd/ |
| A D | rnbd-clt.c | 952 size = dev->size * (limit->logical_block_size / SECTOR_SIZE); in rnbd_client_getgeo() 1347 (le16_to_cpu(rsp->logical_block_size) / SECTOR_SIZE)); in rnbd_clt_setup_gen_disk() 1368 .max_hw_sectors = dev->sess->max_io_size / SECTOR_SIZE, in rnbd_client_setup_device() 1621 sess->max_segments, sess->max_io_size / SECTOR_SIZE, in rnbd_clt_map_device()
|
| /drivers/md/ |
| A D | md-bitmap.c | 360 index * (PAGE_SIZE / SECTOR_SIZE); in read_sb_page() 426 start + opt_size / SECTOR_SIZE <= boundary) in bitmap_io_size() 428 if (start + io_size / SECTOR_SIZE <= boundary) in bitmap_io_size() 444 sector_t ps = pg_index * PAGE_SIZE / SECTOR_SIZE; in __write_sb_page() 467 sboff < (doff + mddev->dev_sectors + PAGE_SIZE / SECTOR_SIZE)) in __write_sb_page()
|
| /drivers/mmc/core/ |
| A D | queue.c | 194 lim->discard_granularity = SECTOR_SIZE; in mmc_queue_setup_discard()
|
| /drivers/block/drbd/ |
| A D | drbd_bitmap.c | 1011 if (len_sect < PAGE_SIZE/SECTOR_SIZE) in bm_page_io_async() 1012 len = (unsigned int)len_sect*SECTOR_SIZE; in bm_page_io_async()
|
| /drivers/target/ |
| A D | target_core_iblock.c | 138 SECTOR_SIZE, in iblock_configure_device()
|
| /drivers/vhost/ |
| A D | scsi.c | 835 (offset & (SECTOR_SIZE - 1) || n & (SECTOR_SIZE - 1)) && in vhost_scsi_map_to_sgl()
|
| /drivers/ufs/core/ |
| A D | ufshcd.c | 9149 .max_sectors = SZ_1M / SECTOR_SIZE,
|