| /linux/drivers/block/drbd/ |
| A D | drbd_interval.c | 39 if (this->sector < here->sector) in drbd_insert_interval() 41 else if (this->sector > here->sector) in drbd_insert_interval() 78 if (sector < here->sector) in drbd_contains_interval() 80 else if (sector > here->sector) in drbd_contains_interval() 118 sector_t end = sector + (size >> 9); in drbd_find_overlap() 130 } else if (here->sector < end && in drbd_find_overlap() 131 sector < here->sector + (here->size >> 9)) { in drbd_find_overlap() 134 } else if (sector >= here->sector) { in drbd_find_overlap() 146 sector_t end = sector + (size >> 9); in drbd_next_overlap() 154 if (i->sector >= end) in drbd_next_overlap() [all …]
|
| A D | drbd_actlog.c | 127 sector_t sector, int op) in _drbd_md_sync_page_io() argument 143 bio->bi_iter.bi_sector = sector; in _drbd_md_sync_page_io() 178 sector_t sector, int op) in drbd_md_sync_page_io() argument 194 (unsigned long long)sector, in drbd_md_sync_page_io() 200 (unsigned long long)sector, in drbd_md_sync_page_io() 319 sector_t sector; in __al_write_transaction() local 861 (unsigned long long)sector, size); in __drbd_change_sync() 869 esector = sector + (size >> 9) - 1; in __drbd_change_sync() 871 if (!expect(sector < nr_sectors)) in __drbd_change_sync() 891 sbnr = BM_SECT_TO_BIT(sector); in __drbd_change_sync() [all …]
|
| A D | drbd_worker.c | 361 sector_t sector = peer_req->i.sector; in w_e_send_csum() local 593 sector_t sector; in make_resync_request() local 659 sector = BM_BIT_TO_SECT(bit); in make_resync_request() 716 size = (capacity-sector)<<9; in make_resync_request() 771 sector_t sector; in make_ov_request() local 780 sector = device->ov_position; in make_ov_request() 782 if (sector >= capacity) in make_ov_request() 809 sector += BM_SECT_PER_BIT; in make_ov_request() 811 device->ov_position = sector; in make_ov_request() 1237 sector_t sector = peer_req->i.sector; in w_e_end_ov_req() local [all …]
|
| A D | drbd_interval.h | 10 sector_t sector; /* start sector of the interval */ member 38 #define drbd_for_each_overlap(i, root, sector, size) \ argument 39 for (i = drbd_find_overlap(root, sector, size); \ 41 i = drbd_next_overlap(i, sector, size))
|
| A D | drbd_receiver.c | 390 peer_req->i.sector = sector; in drbd_alloc_peer_req() 1645 sector_t sector = peer_req->i.sector; in drbd_submit_peer_request() local 2067 sector_t sector = peer_req->i.sector; in e_end_resync_block() local 2157 sector = be64_to_cpu(p->sector); in receive_DataReply() 2191 sector = be64_to_cpu(p->sector); in receive_RSDataReply() 2521 bool superseded = i->sector <= sector && i->sector + in handle_write_conflicts() 2622 sector = be64_to_cpu(p->sector); in receive_Data() 2854 sector = be64_to_cpu(p->sector); in receive_DataRequest() 4998 sector = be64_to_cpu(p->sector); in receive_rs_deallocated() 5844 sector = be64_to_cpu(p->sector); in got_NegRSDReply() [all …]
|
| /linux/block/ |
| A D | blk-lib.c | 61 if ((sector | nr_sects) & bs_mask) in __blkdev_issue_discard() 99 bio->bi_iter.bi_sector = sector; in __blkdev_issue_discard() 104 sector += req_sects; in __blkdev_issue_discard() 182 if ((sector | nr_sects) & bs_mask) in __blkdev_issue_write_same() 193 bio->bi_iter.bi_sector = sector; in __blkdev_issue_write_same() 204 sector += max_write_same_sectors; in __blkdev_issue_write_same() 269 bio->bi_iter.bi_sector = sector; in __blkdev_issue_write_zeroes() 321 bio->bi_iter.bi_sector = sector; in __blkdev_issue_zero_pages() 329 sector += bi_size >> 9; in __blkdev_issue_zero_pages() 367 if ((sector | nr_sects) & bs_mask) in __blkdev_issue_zeroout() [all …]
|
| A D | blk-zoned.c | 198 sector_t sector = 0; in blkdev_zone_reset_all_emulated() local 212 while (sector < capacity) { in blkdev_zone_reset_all_emulated() 214 sector += zone_sectors; in blkdev_zone_reset_all_emulated() 221 bio->bi_iter.bi_sector = sector; in blkdev_zone_reset_all_emulated() 222 sector += zone_sectors; in blkdev_zone_reset_all_emulated() 290 if (sector & (zone_sectors - 1)) in blkdev_zone_mgmt() 308 while (sector < end_sector) { in blkdev_zone_mgmt() 313 sector += zone_sectors; in blkdev_zone_mgmt() 387 if (zrange->sector + zrange->nr_sectors <= zrange->sector || in blkdev_truncate_zone_range() 474 sector_t sector; member [all …]
|
| A D | blk-ia-ranges.c | 18 return sprintf(buf, "%llu\n", iar->sector); in blk_ia_range_sector_show() 196 sector_t sector) in disk_find_ia_range() argument 203 if (sector >= iar->sector && in disk_find_ia_range() 204 sector < iar->sector + iar->nr_sectors) in disk_find_ia_range() 216 sector_t sector = 0; in disk_check_ia_ranges() local 225 tmp = disk_find_ia_range(iars, sector); in disk_check_ia_ranges() 226 if (!tmp || tmp->sector != sector) { in disk_check_ia_ranges() 233 swap(iar->sector, tmp->sector); in disk_check_ia_ranges() 237 sector += iar->nr_sectors; in disk_check_ia_ranges() 240 if (sector != capacity) { in disk_check_ia_ranges() [all …]
|
| /linux/drivers/block/ |
| A D | brd.c | 91 page = brd_lookup_page(brd, sector); in brd_insert_page() 110 idx = sector >> PAGE_SECTORS_SHIFT; in brd_insert_page() 179 if (!brd_insert_page(brd, sector)) in copy_to_brd_setup() 182 sector += copy >> SECTOR_SHIFT; in copy_to_brd_setup() 193 sector_t sector, size_t n) in copy_to_brd() argument 210 sector += copy >> SECTOR_SHIFT; in copy_to_brd() 225 sector_t sector, size_t n) in copy_from_brd() argument 243 sector += copy >> SECTOR_SHIFT; in copy_from_brd() 260 sector_t sector) in brd_do_bvec() argument 301 bio_op(bio), sector); in brd_submit_bio() [all …]
|
| /linux/include/trace/events/ |
| A D | block.h | 23 __field( sector_t, sector ) 29 __entry->sector = bh->b_blocknr; 81 __field( sector_t, sector ) 123 __field( sector_t, sector ) 132 __entry->sector = blk_rq_pos(rq); 155 __field( sector_t, sector ) 241 __field( sector_t, sector ) 269 __field( sector_t, sector ) 426 __field( sector_t, sector ) 464 __field( sector_t, sector ) [all …]
|
| A D | bcache.h | 18 __field(sector_t, sector ) 28 __entry->sector = bio->bi_iter.bi_sector; 36 __entry->rwbs, (unsigned long long)__entry->sector, 96 __field(sector_t, sector ) 103 __entry->sector = bio->bi_iter.bi_sector; 129 __field(sector_t, sector ) 138 __entry->sector = bio->bi_iter.bi_sector; 159 __field(sector_t, sector ) 169 __entry->sector = bio->bi_iter.bi_sector; 230 __field(sector_t, sector ) [all …]
|
| /linux/drivers/block/null_blk/ |
| A D | zoned.c | 62 sector_t sector = 0; in null_init_zoned_dev() local 124 zone->start = sector; in null_init_zoned_dev() 131 sector += dev->zone_size_sects; in null_init_zoned_dev() 138 zone->start = zone->wp = sector; in null_init_zoned_dev() 148 sector += dev->zone_size_sects; in null_init_zoned_dev() 242 sector + nr_sectors <= zone->wp) in null_zone_valid_read_len() 245 if (sector > zone->wp) in null_zone_valid_read_len() 397 sector = zone->wp; in null_zone_write() 401 cmd->rq->__sector = sector; in null_zone_write() 402 } else if (sector != zone->wp) { in null_zone_write() [all …]
|
| A D | null_blk.h | 128 blk_status_t null_handle_discard(struct nullb_device *dev, sector_t sector, 131 enum req_opf op, sector_t sector, 138 int null_report_zones(struct gendisk *disk, sector_t sector, 141 enum req_opf op, sector_t sector, 144 sector_t sector, unsigned int len); 158 enum req_opf op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument 163 sector_t sector, in null_zone_valid_read_len() argument
|
| A D | main.c | 825 idx = sector >> PAGE_SECTORS_SHIFT; in null_free_sector() 1090 sector += temp >> SECTOR_SHIFT; in copy_to_nullb() 1122 sector += temp >> SECTOR_SHIFT; in copy_from_nullb() 1150 sector += temp >> SECTOR_SHIFT; in null_handle_discard() 1189 sector, len); in null_transfer() 1193 sector, valid_len); in null_transfer() 1215 sector_t sector = blk_rq_pos(rq); in null_handle_rq() local 1229 sector += len >> SECTOR_SHIFT; in null_handle_rq() 1256 sector += len >> SECTOR_SHIFT; in null_handle_bio() 1300 sector_t sector, in null_handle_badblocks() argument [all …]
|
| /linux/drivers/scsi/ |
| A D | sr_vendor.c | 174 unsigned long sector; in sr_cd_check() local 186 sector = 0; /* the multisession sector offset goes here */ in sr_cd_check() 212 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() 216 sector = 0; in sr_cd_check() 272 if (sector) in sr_cd_check() 273 sector -= CD_MSF_OFFSET; in sr_cd_check() 309 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check() 318 sector = 0; in sr_cd_check() 322 cd->ms_offset = sector; in sr_cd_check() 334 if (sector) in sr_cd_check() [all …]
|
| /linux/drivers/md/ |
| A D | dm-log-writes.c | 96 __le64 sector; member 125 sector_t sector; member 213 sector_t sector) in write_metadata() argument 226 bio->bi_iter.bi_sector = sector; in write_metadata() 265 sector_t sector) in write_inline_data() argument 338 entry.sector = cpu_to_le64(block->sector); in log_one_block() 345 metadatalen, sector)) { in log_one_block() 372 bio->bi_iter.bi_sector = sector; in log_one_block() 455 sector_t sector = 0; in log_writes_kthread() local 471 sector = lc->next_sector; in log_writes_kthread() [all …]
|
| A D | raid0.c | 296 sector_t sector = *sectorp; in find_zone() local 299 if (sector < z[i].zone_end) { in find_zone() 301 *sectorp = sector - z[i-1].zone_end; in find_zone() 324 sector >>= chunksect_bits; in map_sector() 515 sector_t sector; in raid0_make_request() local 530 sector = bio_sector; in raid0_make_request() 535 ? (sector & (chunk_sects-1)) in raid0_make_request() 536 : sector_div(sector, chunk_sects)); in raid0_make_request() 539 sector = bio_sector; in raid0_make_request() 552 orig_sector = sector; in raid0_make_request() [all …]
|
| A D | dm-stripe.c | 206 sector_t chunk = dm_target_offset(sc->ti, sector); in stripe_map_sector() 236 stripe_map_sector(sc, sector, &stripe, result); in stripe_map_range_sector() 241 sector = *result; in stripe_map_range_sector() 243 *result -= sector_div(sector, sc->chunk_size); in stripe_map_range_sector() 245 *result = sector & ~(sector_t)(sc->chunk_size - 1); in stripe_map_range_sector() 307 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in stripe_dax_direct_access() local 314 stripe_map_sector(sc, sector, &stripe, &dev_sector); in stripe_dax_direct_access() 328 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in stripe_dax_copy_from_iter() local 334 stripe_map_sector(sc, sector, &stripe, &dev_sector); in stripe_dax_copy_from_iter() 347 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in stripe_dax_copy_to_iter() local [all …]
|
| A D | dm-linear.c | 173 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in linear_dax_direct_access() local 175 dev_sector = linear_map_sector(ti, sector); in linear_dax_direct_access() 188 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in linear_dax_copy_from_iter() local 190 dev_sector = linear_map_sector(ti, sector); in linear_dax_copy_from_iter() 202 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in linear_dax_copy_to_iter() local 204 dev_sector = linear_map_sector(ti, sector); in linear_dax_copy_to_iter() 217 sector_t dev_sector, sector = pgoff * PAGE_SECTORS; in linear_dax_zero_page_range() local 219 dev_sector = linear_map_sector(ti, sector); in linear_dax_zero_page_range()
|
| A D | raid5-ppl.c | 303 data_sector = dev->sector; in ppl_log_stripe() 327 if ((sh->sector == sh_last->sector + RAID5_STRIPE_SECTORS(conf)) && in ppl_log_stripe() 464 log->next_io_sector = log->rdev->ppl.sector; in ppl_submit_iounit() 869 sector_t sector; in ppl_recover_entry() local 893 (unsigned long long)sector); in ppl_recover_entry() 897 sector >= rdev->recovery_offset)) { in ppl_recover_entry() 906 (unsigned long long)sector); in ppl_recover_entry() 990 sector_t sector = ppl_sector; in ppl_recover() local 1015 sector += s >> 9; in ppl_recover() 1106 rdev->ppl.sector - rdev->data_offset + in ppl_load_distributed() [all …]
|
| /linux/fs/fat/ |
| A D | cache.c | 310 int fat_get_mapped_cluster(struct inode *inode, sector_t sector, in fat_get_mapped_cluster() argument 319 offset = sector & (sbi->sec_per_clus - 1); in fat_get_mapped_cluster() 326 if (*mapped_blocks > last_block - sector) in fat_get_mapped_cluster() 327 *mapped_blocks = last_block - sector; in fat_get_mapped_cluster() 333 static int is_exceed_eof(struct inode *inode, sector_t sector, in is_exceed_eof() argument 341 if (sector >= *last_block) { in is_exceed_eof() 351 if (sector >= *last_block) in is_exceed_eof() 367 if (sector < (sbi->dir_entries >> sbi->dir_per_block_bits)) { in fat_bmap() 368 *phys = sector + sbi->dir_start; in fat_bmap() 375 if (is_exceed_eof(inode, sector, &last_block, create)) in fat_bmap() [all …]
|
| /linux/fs/zonefs/ |
| A D | trace.h | 29 __field(sector_t, sector) 36 __entry->sector = ZONEFS_I(inode)->i_zsector; 42 blk_op_str(__entry->op), __entry->sector, 53 __field(sector_t, sector) 61 __entry->sector = ZONEFS_I(inode)->i_zsector; 68 __entry->sector, __entry->size, __entry->wpoffset,
|
| /linux/drivers/usb/storage/ |
| A D | jumpshot.c | 155 u32 sector, in jumpshot_read_data() argument 171 if (sector > 0x0FFFFFFF) in jumpshot_read_data() 193 command[2] = sector & 0xFF; in jumpshot_read_data() 194 command[3] = (sector >> 8) & 0xFF; in jumpshot_read_data() 195 command[4] = (sector >> 16) & 0xFF; in jumpshot_read_data() 217 sector += thistime; in jumpshot_read_data() 232 u32 sector, in jumpshot_write_data() argument 248 if (sector > 0x0FFFFFFF) in jumpshot_write_data() 275 command[2] = sector & 0xFF; in jumpshot_write_data() 276 command[3] = (sector >> 8) & 0xFF; in jumpshot_write_data() [all …]
|
| /linux/drivers/mtd/ |
| A D | rfd_ftl.c | 247 if (sector >= part->sector_count) in rfd_ftl_readsect() 250 addr = part->sector_map[sector]; in rfd_ftl_readsect() 642 part->sector_map[sector] = addr; in do_writesect() 644 entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector); in do_writesect() 681 if (sector >= part->sector_count) { in rfd_ftl_writesect() 686 old_addr = part->sector_map[sector]; in rfd_ftl_writesect() 699 part->sector_map[sector] = -1; in rfd_ftl_writesect() 716 if (sector >= part->sector_count) in rfd_ftl_discardsect() 719 addr = part->sector_map[sector]; in rfd_ftl_discardsect() 726 part->sector_map[sector] = -1; in rfd_ftl_discardsect() [all …]
|
| /linux/fs/exfat/ |
| A D | nls.c | 651 sector_t sector, unsigned long long num_sectors, in exfat_load_upcase_table() argument 667 num_sectors += sector; in exfat_load_upcase_table() 669 while (sector < num_sectors) { in exfat_load_upcase_table() 672 bh = sb_bread(sb, sector); in exfat_load_upcase_table() 675 (unsigned long long)sector); in exfat_load_upcase_table() 679 sector++; in exfat_load_upcase_table() 751 sector_t sector; in exfat_create_upcase_table() local 782 sector = exfat_cluster_to_sector(sbi, tbl_clu); in exfat_create_upcase_table() 784 ret = exfat_load_upcase_table(sb, sector, num_sectors, in exfat_create_upcase_table()
|