| /drivers/block/drbd/ |
| A D | drbd_interval.c | 39 if (this->sector < here->sector) in drbd_insert_interval() 41 else if (this->sector > here->sector) in drbd_insert_interval() 78 if (sector < here->sector) in drbd_contains_interval() 80 else if (sector > here->sector) in drbd_contains_interval() 122 sector_t end = sector + (size >> 9); in drbd_find_overlap() 134 } else if (here->sector < end && in drbd_find_overlap() 135 sector < here->sector + (here->size >> 9)) { in drbd_find_overlap() 138 } else if (sector >= here->sector) { in drbd_find_overlap() 150 sector_t end = sector + (size >> 9); in drbd_next_overlap() 158 if (i->sector >= end) in drbd_next_overlap() [all …]
|
| A D | drbd_actlog.c | 127 sector_t sector, enum req_op op) in _drbd_md_sync_page_io() argument 144 bio->bi_iter.bi_sector = sector; in _drbd_md_sync_page_io() 178 sector_t sector, enum req_op op) in drbd_md_sync_page_io() argument 190 if (sector < drbd_md_first_sector(bdev) || in drbd_md_sync_page_io() 194 (unsigned long long)sector, in drbd_md_sync_page_io() 200 (unsigned long long)sector, in drbd_md_sync_page_io() 319 sector_t sector; in __al_write_transaction() local 863 (unsigned long long)sector, size); in __drbd_change_sync() 871 esector = sector + (size >> 9) - 1; in __drbd_change_sync() 893 sbnr = BM_SECT_TO_BIT(sector); in __drbd_change_sync() [all …]
|
| A D | drbd_worker.c | 356 sector_t sector = peer_req->i.sector; in w_e_send_csum() local 591 sector_t sector; in make_resync_request() local 657 sector = BM_BIT_TO_SECT(bit); in make_resync_request() 714 size = (capacity-sector)<<9; in make_resync_request() 770 sector_t sector; in make_ov_request() local 779 sector = device->ov_position; in make_ov_request() 781 if (sector >= capacity) in make_ov_request() 808 sector += BM_SECT_PER_BIT; in make_ov_request() 810 device->ov_position = sector; in make_ov_request() 1217 sector_t sector = peer_req->i.sector; in w_e_end_ov_req() local [all …]
|
| A D | drbd_receiver.c | 208 peer_req->i.sector = sector; in drbd_alloc_peer_req() 1452 sector_t sector = peer_req->i.sector; in drbd_submit_peer_request() local 1852 sector_t sector = peer_req->i.sector; in e_end_resync_block() local 1942 sector = be64_to_cpu(p->sector); in receive_DataReply() 1973 sector = be64_to_cpu(p->sector); in receive_RSDataReply() 2301 bool superseded = i->sector <= sector && i->sector + in handle_write_conflicts() 2405 sector = be64_to_cpu(p->sector); in receive_Data() 2635 sector = be64_to_cpu(p->sector); in receive_DataRequest() 4774 sector = be64_to_cpu(p->sector); in receive_rs_deallocated() 5610 sector = be64_to_cpu(p->sector); in got_NegRSDReply() [all …]
|
| A D | drbd_interval.h | 10 sector_t sector; /* start sector of the interval */ member 38 #define drbd_for_each_overlap(i, root, sector, size) \ argument 39 for (i = drbd_find_overlap(root, sector, size); \ 41 i = drbd_next_overlap(i, sector, size))
|
| /drivers/block/null_blk/ |
| A D | zoned.c | 56 sector_t sector = 0; in null_init_zoned_dev() local 134 zone->start = sector; in null_init_zoned_dev() 141 sector += dev->zone_size_sects; in null_init_zoned_dev() 148 zone->start = sector; in null_init_zoned_dev() 164 sector += dev->zone_size_sects; in null_init_zoned_dev() 252 if (sector > zone->wp) in null_zone_valid_read_len() 383 sector = zone->wp; in null_zone_write() 387 if (sector != zone->wp || in null_zone_write() 647 sector_t sector) in null_zone_mgmt() argument 781 unsigned long long sector; in zone_cond_store() local [all …]
|
| A D | null_blk.h | 132 blk_status_t null_handle_discard(struct nullb_device *dev, sector_t sector, 135 sector_t sector, unsigned int nr_sectors); 136 blk_status_t null_handle_badblocks(struct nullb_cmd *cmd, sector_t sector, 139 sector_t sector, sector_t nr_sectors); 145 int null_report_zones(struct gendisk *disk, sector_t sector, 148 sector_t sector, sector_t nr_sectors); 150 sector_t sector, unsigned int len); 166 enum req_op op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument 171 sector_t sector, in null_zone_valid_read_len() argument
|
| A D | main.c | 899 idx = sector >> PAGE_SECTORS_SHIFT; in null_free_sector() 970 idx = sector >> PAGE_SECTORS_SHIFT; in __null_lookup_page() 1159 sector += temp >> SECTOR_SHIFT; in copy_to_nullb() 1185 sector += temp >> SECTOR_SHIFT; in copy_from_nullb() 1209 sector += temp >> SECTOR_SHIFT; in null_handle_discard() 1248 sector, len); in null_transfer() 1252 sector, valid_len); in null_transfer() 1279 sector_t sector = blk_rq_pos(rq); in null_handle_data_transfer() local 1295 sector += len >> SECTOR_SHIFT; in null_handle_data_transfer() 1354 if (sector < first_bad) in null_handle_badblocks() [all …]
|
| /drivers/scsi/ |
| A D | sr_vendor.c | 174 unsigned long sector; in sr_cd_check() local 186 sector = 0; /* the multisession sector offset goes here */ in sr_cd_check() 212 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() 216 sector = 0; in sr_cd_check() 272 if (sector) in sr_cd_check() 273 sector -= CD_MSF_OFFSET; in sr_cd_check() 309 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() 318 sector = 0; in sr_cd_check() 322 cd->ms_offset = sector; in sr_cd_check() 334 if (sector) in sr_cd_check() [all …]
|
| /drivers/block/ |
| A D | brd.c | 49 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) in brd_lookup_page() argument 51 return xa_load(&brd->brd_pages, sector >> PAGE_SECTORS_SHIFT); in brd_lookup_page() 57 static struct page *brd_insert_page(struct brd_device *brd, sector_t sector, in brd_insert_page() argument 73 ret = __xa_cmpxchg(&brd->brd_pages, sector >> PAGE_SECTORS_SHIFT, NULL, in brd_insert_page() 112 sector_t sector = bio->bi_iter.bi_sector; in brd_rw_bvec() local 113 u32 offset = (sector & (PAGE_SECTORS - 1)) << SECTOR_SHIFT; in brd_rw_bvec() 121 page = brd_lookup_page(brd, sector); in brd_rw_bvec() 123 page = brd_insert_page(brd, sector, opf); in brd_rw_bvec() 159 static void brd_do_discard(struct brd_device *brd, sector_t sector, u32 size) in brd_do_discard() argument 161 sector_t aligned_sector = round_up(sector, PAGE_SECTORS); in brd_do_discard() [all …]
|
| /drivers/usb/storage/ |
| A D | jumpshot.c | 155 u32 sector, in jumpshot_read_data() argument 171 if (sector > 0x0FFFFFFF) in jumpshot_read_data() 193 command[2] = sector & 0xFF; in jumpshot_read_data() 194 command[3] = (sector >> 8) & 0xFF; in jumpshot_read_data() 195 command[4] = (sector >> 16) & 0xFF; in jumpshot_read_data() 217 sector += thistime; in jumpshot_read_data() 232 u32 sector, in jumpshot_write_data() argument 248 if (sector > 0x0FFFFFFF) in jumpshot_write_data() 275 command[2] = sector & 0xFF; in jumpshot_write_data() 276 command[3] = (sector >> 8) & 0xFF; in jumpshot_write_data() [all …]
|
| A D | datafab.c | 137 u32 sector, in datafab_read_data() argument 182 command[2] = sector & 0xFF; in datafab_read_data() 183 command[3] = (sector >> 8) & 0xFF; in datafab_read_data() 184 command[4] = (sector >> 16) & 0xFF; in datafab_read_data() 187 command[5] |= (sector >> 24) & 0x0F; in datafab_read_data() 205 sector += thistime; in datafab_read_data() 220 u32 sector, in datafab_write_data() argument 270 command[2] = sector & 0xFF; in datafab_write_data() 271 command[3] = (sector >> 8) & 0xFF; in datafab_write_data() 272 command[4] = (sector >> 16) & 0xFF; in datafab_write_data() [all …]
|
| A D | shuttle_usbat.c | 205 buf[2] = sector & 0xFF; in usbat_pack_ata_sector_cmd() 206 buf[3] = (sector >> 8) & 0xFF; in usbat_pack_ata_sector_cmd() 207 buf[4] = (sector >> 16) & 0xFF; in usbat_pack_ata_sector_cmd() 1109 u32 sector, in usbat_flash_read_data() argument 1140 if (sector > 0x0FFFFFFF) in usbat_flash_read_data() 1183 sector += thistime; in usbat_flash_read_data() 1200 u32 sector, in usbat_flash_write_data() argument 1231 if (sector > 0x0FFFFFFF) in usbat_flash_write_data() 1272 sector += thistime; in usbat_flash_write_data() 1296 unsigned int sector; in usbat_hp8200e_handle_read10() local [all …]
|
| /drivers/mtd/ |
| A D | rfd_ftl.c | 247 if (sector >= part->sector_count) in rfd_ftl_readsect() 250 addr = part->sector_map[sector]; in rfd_ftl_readsect() 642 part->sector_map[sector] = addr; in do_writesect() 644 entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector); in do_writesect() 681 if (sector >= part->sector_count) { in rfd_ftl_writesect() 686 old_addr = part->sector_map[sector]; in rfd_ftl_writesect() 699 part->sector_map[sector] = -1; in rfd_ftl_writesect() 716 if (sector >= part->sector_count) in rfd_ftl_discardsect() 719 addr = part->sector_map[sector]; in rfd_ftl_discardsect() 726 part->sector_map[sector] = -1; in rfd_ftl_discardsect() [all …]
|
| /drivers/vdpa/vdpa_sim/ |
| A D | vdpa_sim_blk.c | 118 u64 sector; in vdpasim_blk_handle_req() local 156 sector = vdpasim64_to_cpu(vdpasim, hdr.sector); in vdpasim_blk_handle_req() 157 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req() 161 sector != 0) { in vdpasim_blk_handle_req() 164 type, sector); in vdpasim_blk_handle_req() 171 if (!vdpasim_blk_check_range(vdpasim, sector, in vdpasim_blk_handle_req() 194 if (!vdpasim_blk_check_range(vdpasim, sector, in vdpasim_blk_handle_req() 255 sector = le64_to_cpu(range.sector); in vdpasim_blk_handle_req() 256 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req() 277 if (!vdpasim_blk_check_range(vdpasim, sector, num_sectors, in vdpasim_blk_handle_req()
|
| /drivers/md/ |
| A D | dm-log-writes.c | 97 __le64 sector; member 126 sector_t sector; member 214 sector_t sector) in write_metadata() argument 223 bio->bi_iter.bi_sector = sector; in write_metadata() 260 sector_t sector) in write_inline_data() argument 276 bio->bi_iter.bi_sector = sector; in write_inline_data() 326 entry.sector = cpu_to_le64(block->sector); in log_one_block() 333 metadatalen, sector)) { in log_one_block() 357 bio->bi_iter.bi_sector = sector; in log_one_block() 433 sector_t sector = 0; in log_writes_kthread() local [all …]
|
| A D | raid0.c | 305 sector_t sector = *sectorp; in find_zone() local 308 if (sector < z[i].zone_end) { in find_zone() 310 *sectorp = sector - z[i-1].zone_end; in find_zone() 333 sector >>= chunksect_bits; in map_sector() 559 sector_t sector = bio_sector; in raid0_map_submit_bio() local 563 zone = find_zone(mddev->private, §or); in raid0_map_submit_bio() 569 tmp_dev = map_sector(mddev, zone, sector, §or); in raid0_map_submit_bio() 593 sector_t sector; in raid0_make_request() local 606 sector = bio->bi_iter.bi_sector; in raid0_make_request() 611 ? (sector & (chunk_sects-1)) in raid0_make_request() [all …]
|
| A D | raid1.c | 64 sector_t lo = r1_bio->sector; in check_and_add_serial() 313 sector_t sector = r1_bio->sector; in raid_end_bio_io() local 1357 r1_bio->sector); in raid1_read_request() 2509 sector_t sector; in narrow_write_error() local 2519 sector = r1_bio->sector; in narrow_write_error() 2522 - sector; in narrow_write_error() 2543 bio_trim(wbio, sector - r1_bio->sector, sectors); in narrow_write_error() 2554 sector += sectors; in narrow_write_error() 2592 r1_bio->sector, in handle_write_finished() 2632 sector_t sector; in handle_read_error() local [all …]
|
| A D | raid5-ppl.c | 304 data_sector = dev->sector; in ppl_log_stripe() 328 if ((sh->sector == sh_last->sector + RAID5_STRIPE_SECTORS(conf)) && in ppl_log_stripe() 463 log->next_io_sector = log->rdev->ppl.sector; in ppl_submit_iounit() 857 sector_t sector; in ppl_recover_entry() local 881 (unsigned long long)sector); in ppl_recover_entry() 885 sector >= rdev->recovery_offset)) { in ppl_recover_entry() 894 (unsigned long long)sector); in ppl_recover_entry() 979 sector_t sector = ppl_sector; in ppl_recover() local 1004 sector += s >> 9; in ppl_recover() 1095 rdev->ppl.sector - rdev->data_offset + in ppl_load_distributed() [all …]
|
| A D | raid5.c | 592 sh->sector = sector; in init_stripe() 625 if (sh->sector == sector && sh->generation == generation) in __find_stripe() 1044 if (da->sector > db->sector) in cmp_stripe() 1046 if (da->sector < db->sector) in cmp_stripe() 1121 ent->sector = sector; in defer_issue_bios() 3467 sector_t sector; in stripe_bio_overlaps() local 3476 sector = sh->dev[i].sector; in stripe_bio_overlaps() 3480 last = sector; in stripe_bio_overlaps() 3522 sector_t sector = sh->dev[dd_idx].sector; in __add_stripe_bio() local 3530 if (sector >= sh->dev[dd_idx].sector + RAID5_STRIPE_SECTORS(conf)) in __add_stripe_bio() [all …]
|
| A D | dm-zone.c | 20 sector_t sector, unsigned int nr_zones, in dm_blk_do_report_zones() argument 26 .next_sector = sector, in dm_blk_do_report_zones() 54 int dm_blk_report_zones(struct gendisk *disk, sector_t sector, in dm_blk_report_zones() argument 80 ret = dm_blk_do_report_zones(md, map, sector, nr_zones, cb, in dm_blk_report_zones() 123 int dm_report_zones(struct block_device *bdev, sector_t start, sector_t sector, in dm_report_zones() argument 132 return blkdev_report_zones(bdev, sector, nr_zones, in dm_report_zones() 492 sector_t sector, unsigned int nr_zones, in dm_zone_get_reset_bitmap() argument 497 ret = dm_blk_do_report_zones(md, t, sector, nr_zones, in dm_zone_get_reset_bitmap()
|
| /drivers/char/ |
| A D | ps3flash.c | 98 u64 size, sector, offset; in ps3flash_read() local 118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 128 res = ps3flash_fetch(dev, sector); in ps3flash_read() 151 sector += priv->chunk_sectors; in ps3flash_read() 167 u64 size, sector, offset; in ps3flash_write() local 187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 198 res = ps3flash_fetch(dev, sector); in ps3flash_write() 199 else if (sector != priv->tag) in ps3flash_write() 219 priv->tag = sector; in ps3flash_write() 226 sector += priv->chunk_sectors; in ps3flash_write()
|
| /drivers/mtd/nand/raw/ |
| A D | sh_flctl.c | 490 res = wait_recfifo_ready(flctl , sector); in read_ecfiforeg() 624 int sector, page_sectors; in execmd_read_page_sector() local 640 for (sector = 0; sector < page_sectors; sector++) { in execmd_read_page_sector() 641 read_fiforeg(flctl, 512, 512 * sector); in execmd_read_page_sector() 645 sector); in execmd_read_page_sector() 695 int sector, page_sectors; in execmd_write_page_sector() local 708 for (sector = 0; sector < page_sectors; sector++) { in execmd_write_page_sector() 709 write_fiforeg(flctl, 512, 512 * sector); in execmd_write_page_sector() 721 int sector, page_sectors; in execmd_write_oob() local 728 for (sector = 0; sector < page_sectors; sector++) { in execmd_write_oob() [all …]
|
| /drivers/mtd/devices/ |
| A D | docg3.c | 459 int sector, ret = 0; in doc_read_seek() local 483 doc_setup_addr_sector(docg3, sector); in doc_read_seek() 487 doc_setup_addr_sector(docg3, sector); in doc_read_seek() 508 int ret = 0, sector; in doc_write_seek() local 835 uint sector, pages_biblock; in calc_block_sector() local 841 sector = from / DOC_LAYOUT_PAGE_SIZE; in calc_block_sector() 844 *page = sector % pages_biblock; in calc_block_sector() 848 if (sector % 2) in calc_block_sector() 1144 int ret, sector; in doc_erase_block() local 1156 doc_setup_addr_sector(docg3, sector); in doc_erase_block() [all …]
|
| /drivers/nvdimm/ |
| A D | pmem.c | 60 return (sector << SECTOR_SHIFT) + pmem->data_offset; in to_offset() 92 badblocks_clear(&pmem->bb, sector, blks); in pmem_clear_bb() 166 sector_t sector, unsigned int len) in pmem_do_read() argument 169 phys_addr_t pmem_off = to_offset(pmem, sector); in pmem_do_read() 172 if (unlikely(is_bad_pmem(&pmem->bb, sector, len))) in pmem_do_read() 182 sector_t sector, unsigned int len) in pmem_do_write() argument 184 phys_addr_t pmem_off = to_offset(pmem, sector); in pmem_do_write() 187 if (unlikely(is_bad_pmem(&pmem->bb, sector, len))) { in pmem_do_write() 247 sector_t sector = PFN_PHYS(pgoff) >> SECTOR_SHIFT; in __pmem_direct_access() local 271 PAGE_ALIGN((first_bad - sector) << SECTOR_SHIFT)); in __pmem_direct_access() [all …]
|