| /linux/block/ |
| A D | blk-settings.c | 37 memset(lim, 0, sizeof(*lim)); in blk_set_stacking_limits() 88 lim->max_open_zones > lim->max_active_zones) in blk_validate_zoned_limits() 91 if (lim->zone_write_granularity < lim->logical_block_size) in blk_validate_zoned_limits() 92 lim->zone_write_granularity = lim->logical_block_size; in blk_validate_zoned_limits() 242 if (lim->physical_block_size < lim->logical_block_size) in blk_validate_limits() 243 lim->physical_block_size = lim->logical_block_size; in blk_validate_limits() 249 if (lim->io_min < lim->physical_block_size) in blk_validate_limits() 250 lim->io_min = lim->physical_block_size; in blk_validate_limits() 268 lim->max_hw_sectors = round_down(lim->max_hw_sectors, in blk_validate_limits() 292 lim->max_sectors = round_down(lim->max_sectors, in blk_validate_limits() [all …]
|
| A D | blk-merge.c | 145 min(lim->max_discard_sectors, bio_allowed_max_sectors(lim)); in bio_split_discard() 173 if (!lim->max_write_zeroes_sectors) in bio_split_write_zeroes() 188 return lim->atomic_write_boundary_sectors; in blk_boundary_sectors() 190 return lim->chunk_sectors; in blk_boundary_sectors() 202 const struct queue_limits *lim) in get_max_io_size() argument 217 max_sectors = lim->max_sectors; in get_max_io_size() 249 min(lim->seg_boundary_mask - (lim->seg_boundary_mask & paddr), in get_max_segment_size() 326 if (nsegs < lim->max_segments && in bio_split_rw_at() 333 lim->max_segments, max_bytes)) in bio_split_rw_at() 377 bio_split_rw_at(bio, lim, nr_segs, in bio_split_rw() [all …]
|
| A D | blk.h | 124 return (offset & lim->virt_boundary_mask) || in __bvec_gap_to_prev() 135 if (!lim->virt_boundary_mask) in bvec_gap_to_prev() 137 return __bvec_gap_to_prev(lim, bprv, offset); in bvec_gap_to_prev() 337 const struct queue_limits *lim, unsigned *nsegs); 352 const struct queue_limits *lim) in bio_may_need_split() argument 354 return lim->chunk_sectors || bio->bi_vcnt != 1 || in bio_may_need_split() 377 if (bio_may_need_split(bio, lim)) in __bio_split_to_limits() 378 return bio_split_rw(bio, lim, nr_segs); in __bio_split_to_limits() 382 return bio_split_zone_append(bio, lim, nr_segs); in __bio_split_to_limits() 385 return bio_split_discard(bio, lim, nr_segs); in __bio_split_to_limits() [all …]
|
| A D | blk-sysfs.c | 159 struct queue_limits lim; local 173 lim = queue_limits_start_update(disk->queue); 197 struct queue_limits lim; in queue_max_sectors_store() local 205 lim = queue_limits_start_update(disk->queue); in queue_max_sectors_store() 206 lim.max_user_sectors = max_sectors_kb << 1; in queue_max_sectors_store() 216 struct queue_limits lim; in queue_feature_store() local 224 lim = queue_limits_start_update(disk->queue); in queue_feature_store() 226 lim.features |= feature; in queue_feature_store() 228 lim.features &= ~feature; in queue_feature_store() 380 struct queue_limits lim; in queue_wc_store() local [all …]
|
| /linux/drivers/media/i2c/ |
| A D | ccs-pll.c | 161 lim_fr = &lim->op_fr; in check_fr_bounds() 164 lim_fr = &lim->vt_fr; in check_fr_bounds() 204 lim_bk = &lim->op_bk; in check_bk_bounds() 207 lim_bk = &lim->vt_bk; in check_bk_bounds() 471 if (lim->min_line_length_pck_bin > lim->min_line_length_pck in ccs_pll_calculate_vt() 512 max_vt_div = lim->vt_bk.max_sys_clk_div * lim->vt_bk.max_pix_clk_div; in ccs_pll_calculate_vt() 725 op_lim_fr = &lim->op_fr; in ccs_pll_calculate() 726 op_lim_bk = &lim->op_bk; in ccs_pll_calculate() 735 op_lim_fr = &lim->vt_fr; in ccs_pll_calculate() 736 op_lim_bk = &lim->vt_bk; in ccs_pll_calculate() [all …]
|
| /linux/drivers/mmc/core/ |
| A D | queue.c | 178 struct queue_limits *lim) in mmc_queue_setup_discard() argument 186 lim->max_hw_discard_sectors = max_discard; in mmc_queue_setup_discard() 188 lim->max_secure_erase_sectors = max_discard; in mmc_queue_setup_discard() 194 lim->discard_granularity = SECTOR_SIZE; in mmc_queue_setup_discard() 350 struct queue_limits lim = { in mmc_alloc_disk() local 356 mmc_queue_setup_discard(card, &lim); in mmc_alloc_disk() 363 lim.logical_block_size = 512; in mmc_alloc_disk() 366 lim.logical_block_size != 4096); in mmc_alloc_disk() 376 lim.max_segment_size = in mmc_alloc_disk() 378 lim.max_segments = host->max_segs; in mmc_alloc_disk() [all …]
|
| /linux/drivers/md/ |
| A D | dm-zone.c | 256 struct queue_limits *lim; member 309 zlim->lim->max_active_zones = in device_get_zone_resource_limits() 315 zlim->lim->max_open_zones = in device_get_zone_resource_limits() 329 struct queue_limits *lim) in dm_set_zones_restrictions() argument 335 .lim = lim, in dm_set_zones_restrictions() 347 lim->max_zone_append_sectors = 0; in dm_set_zones_restrictions() 380 lim->max_open_zones = 0; in dm_set_zones_restrictions() 381 lim->max_active_zones = 0; in dm_set_zones_restrictions() 382 lim->max_zone_append_sectors = 0; in dm_set_zones_restrictions() 383 lim->zone_write_granularity = 0; in dm_set_zones_restrictions() [all …]
|
| /linux/lib/ |
| A D | bitmap.c | 41 for (k = 0; k < lim; ++k) in __bitmap_equal() 61 for (k = 0; k < lim; ++k) { in __bitmap_or_equal() 76 for (k = 0; k < lim; ++k) in __bitmap_complement() 98 for (k = 0; off + k < lim; ++k) { in __bitmap_shift_right() 105 if (!rem || off + k + 1 >= lim) in __bitmap_shift_right() 109 if (off + k + 1 == lim - 1) in __bitmap_shift_right() 114 if (off + k == lim - 1) in __bitmap_shift_right() 236 for (k = 0; k < lim; k++) in __bitmap_and() 274 for (k = 0; k < lim; k++) in __bitmap_andnot() 299 for (k = 0; k < lim; ++k) in __bitmap_intersects() [all …]
|
| /linux/tools/lib/ |
| A D | bitmap.c | 10 unsigned int k, w = 0, lim = bits/BITS_PER_LONG; in __bitmap_weight() local 12 for (k = 0; k < lim; k++) in __bitmap_weight() 64 unsigned int lim = bits/BITS_PER_LONG; in __bitmap_and() local 67 for (k = 0; k < lim; k++) in __bitmap_and() 78 unsigned int k, lim = bits/BITS_PER_LONG; in __bitmap_equal() local 79 for (k = 0; k < lim; ++k) in __bitmap_equal() 93 unsigned int k, lim = bits/BITS_PER_LONG; in __bitmap_intersects() local 94 for (k = 0; k < lim; ++k) in __bitmap_intersects()
|
| /linux/drivers/block/ |
| A D | virtio_blk.c | 724 struct queue_limits *lim) in virtblk_read_zoned_limits() argument 735 lim->max_open_zones = v; in virtblk_read_zoned_limits() 740 lim->max_active_zones = v; in virtblk_read_zoned_limits() 750 lim->io_min = wg; in virtblk_read_zoned_limits() 799 struct queue_limits *lim) in virtblk_read_zoned_limits() argument 1099 struct queue_limits lim; in cache_type_store() local 1246 struct queue_limits *lim) in virtblk_read_limits() argument 1304 lim->alignment_offset = in virtblk_read_limits() 1311 lim->io_min = lim->logical_block_size * min_io_size; in virtblk_read_limits() 1317 lim->io_opt = lim->logical_block_size * opt_io_size; in virtblk_read_limits() [all …]
|
| /linux/tools/include/nolibc/ |
| A D | stdlib.h | 244 unsigned long lim; in utoa_r() local 250 for (dig = 0, lim = 1; dig < pos; dig++) in utoa_r() 251 lim *= 10; in utoa_r() 253 if (digits || in >= lim || !pos) { in utoa_r() 254 for (dig = 0; in >= lim; dig++) in utoa_r() 255 in -= lim; in utoa_r() 380 unsigned long long lim; in u64toa_r() local 387 lim *= 10; in u64toa_r() 389 if (digits || in >= lim || !pos) { in u64toa_r() 390 for (dig = 0; in >= lim; dig++) in u64toa_r() [all …]
|
| /linux/arch/um/os-Linux/ |
| A D | main.c | 29 struct rlimit lim; in set_stklim() local 31 if (getrlimit(RLIMIT_STACK, &lim) < 0) { in set_stklim() 35 if ((lim.rlim_cur == RLIM_INFINITY) || (lim.rlim_cur > STACKSIZE)) { in set_stklim() 36 lim.rlim_cur = STACKSIZE; in set_stklim() 37 if (setrlimit(RLIMIT_STACK, &lim) < 0) { in set_stklim()
|
| A D | start_up.c | 229 struct rlimit lim; in check_coredump_limit() local 230 int err = getrlimit(RLIMIT_CORE, &lim); in check_coredump_limit() 238 if (lim.rlim_cur == RLIM_INFINITY) in check_coredump_limit() 241 os_info("%llu\n", (unsigned long long)lim.rlim_cur); in check_coredump_limit() 244 if (lim.rlim_max == RLIM_INFINITY) in check_coredump_limit() 247 os_info("%llu\n", (unsigned long long)lim.rlim_max); in check_coredump_limit()
|
| /linux/drivers/usb/storage/ |
| A D | scsiglue.c | 91 static int device_configure(struct scsi_device *sdev, struct queue_limits *lim) in device_configure() argument 106 lim->max_hw_sectors = min(lim->max_hw_sectors, max_sectors); in device_configure() 113 lim->max_hw_sectors = 0x7FFFFF; in device_configure() 119 lim->max_hw_sectors = 2048; in device_configure() 126 lim->max_hw_sectors = min_t(size_t, in device_configure() 127 lim->max_hw_sectors, dma_max_mapping_size(dev) >> SECTOR_SHIFT); in device_configure() 588 struct queue_limits lim; in max_sectors_store() local 596 lim = queue_limits_start_update(sdev->request_queue); in max_sectors_store() 597 lim.max_hw_sectors = ms; in max_sectors_store() 598 ret = queue_limits_commit_update(sdev->request_queue, &lim); in max_sectors_store()
|
| /linux/drivers/pps/generators/ |
| A D | pps_gen_parport.c | 60 long lim, delta; in hrtimer_event() local 78 lim = NSEC_PER_SEC - send_delay - dev->port_write_time; in hrtimer_event() 81 if (expire_time.tv_sec != ts1.tv_sec || ts1.tv_nsec > lim) { in hrtimer_event() 91 } while (expire_time.tv_sec == ts2.tv_sec && ts2.tv_nsec < lim); in hrtimer_event() 98 lim = NSEC_PER_SEC - dev->port_write_time; in hrtimer_event() 101 } while (expire_time.tv_sec == ts2.tv_sec && ts2.tv_nsec < lim); in hrtimer_event()
|
| /linux/rust/kernel/block/mq/ |
| A D | gen_disk.rs | 97 let mut lim: bindings::queue_limits = unsafe { core::mem::zeroed() }; in build() variables 99 lim.logical_block_size = self.logical_block_size; in build() 100 lim.physical_block_size = self.physical_block_size; in build() 102 lim.features = bindings::BLK_FEAT_ROTATIONAL; in build() 109 &mut lim, in build()
|
| /linux/drivers/edac/ |
| A D | amd64_edac.h | 136 #define dram_intlv_sel(pvt, i) ((u8)((pvt->ranges[i].lim.lo >> 8) & 0x7)) 137 #define dram_dst_node(pvt, i) ((u8)(pvt->ranges[i].lim.lo & 0x7)) 289 struct reg_pair lim; member 422 u64 lim = (((u64)pvt->ranges[i].lim.lo & 0xffff0000) << 8) | 0x00ffffff; in get_dram_limit() local 425 return lim; in get_dram_limit() 427 return (((u64)pvt->ranges[i].lim.hi & 0x000000ff) << 40) | lim; in get_dram_limit()
|
| /linux/drivers/scsi/ |
| A D | sd.c | 108 struct queue_limits *lim); 125 struct queue_limits *lim) in sd_set_flush_flag() argument 173 struct queue_limits lim; in cache_type_store() local 471 struct queue_limits lim; in provisioning_mode_store() local 573 struct queue_limits lim; in max_write_same_blocks_store() local 1091 struct queue_limits *lim) in sd_config_write_same() argument 2577 struct queue_limits *lim) in sd_config_protection() argument 3336 struct queue_limits *lim) in sd_read_block_limits() argument 3703 struct queue_limits lim; in sd_revalidate_disk() local 3787 lim.io_min = 0; in sd_revalidate_disk() [all …]
|
| A D | sun3x_esp.c | 87 int lim; in sun3x_esp_dma_drain() local 95 lim = 1000; in sun3x_esp_dma_drain() 97 if (--lim == 0) { in sun3x_esp_dma_drain() 109 int lim; in sun3x_esp_dma_invalidate() local 111 lim = 1000; in sun3x_esp_dma_invalidate() 113 if (--lim == 0) { in sun3x_esp_dma_invalidate()
|
| A D | sd_zbc.c | 591 int sd_zbc_read_zones(struct scsi_disk *sdkp, struct queue_limits *lim, in sd_zbc_read_zones() argument 601 lim->features |= BLK_FEAT_ZONED; in sd_zbc_read_zones() 608 lim->zone_write_granularity = sdkp->physical_block_size; in sd_zbc_read_zones() 631 lim->max_open_zones = 0; in sd_zbc_read_zones() 633 lim->max_open_zones = sdkp->zones_max_open; in sd_zbc_read_zones() 634 lim->max_active_zones = 0; in sd_zbc_read_zones() 635 lim->chunk_sectors = logical_to_sectors(sdkp->device, zone_blocks); in sd_zbc_read_zones() 637 lim->max_zone_append_sectors = 0; in sd_zbc_read_zones()
|
| A D | sun_esp.c | 225 int can_do_sbus64, lim; in sbus_esp_reset_dma() local 265 lim = 1000; in sbus_esp_reset_dma() 267 if (--lim == 0) { in sbus_esp_reset_dma() 324 int lim; in sbus_esp_dma_drain() local 336 lim = 1000; in sbus_esp_dma_drain() 338 if (--lim == 0) { in sbus_esp_dma_drain() 366 int lim; in sbus_esp_dma_invalidate() local 368 lim = 1000; in sbus_esp_dma_invalidate() 370 if (--lim == 0) { in sbus_esp_dma_invalidate()
|
| /linux/drivers/regulator/ |
| A D | qcom-labibb-regulator.c | 310 static int qcom_labibb_set_ocp(struct regulator_dev *rdev, int lim, in qcom_labibb_set_ocp() argument 322 if (lim || severity != REGULATOR_SEVERITY_PROT || !enable) in qcom_labibb_set_ocp() 565 struct labibb_current_limits *lim = &vreg->uA_limits; in qcom_labibb_set_current_limit() local 569 if (min_uA < lim->uA_min || max_uA < lim->uA_min) in qcom_labibb_set_current_limit() 573 int uA_limit = (lim->uA_step * i) + lim->uA_min; in qcom_labibb_set_current_limit() 587 mask = desc->csel_mask | lim->ovr_val; in qcom_labibb_set_current_limit() 589 val = (u32)sel | lim->ovr_val; in qcom_labibb_set_current_limit() 599 struct labibb_current_limits *lim = &vreg->uA_limits; in qcom_labibb_get_current_limit() local 608 return (cur_step * lim->uA_step) + lim->uA_min; in qcom_labibb_get_current_limit()
|
| /linux/drivers/nvme/host/ |
| A D | zns.c | 108 void nvme_update_zone_info(struct nvme_ns *ns, struct queue_limits *lim, in nvme_update_zone_info() argument 111 lim->features |= BLK_FEAT_ZONED; in nvme_update_zone_info() 112 lim->max_open_zones = zi->max_open_zones; in nvme_update_zone_info() 113 lim->max_active_zones = zi->max_active_zones; in nvme_update_zone_info() 114 lim->max_zone_append_sectors = ns->ctrl->max_zone_append; in nvme_update_zone_info() 115 lim->chunk_sectors = ns->head->zsze = in nvme_update_zone_info()
|
| A D | core.c | 1852 lim->discard_granularity = lim->logical_block_size; in nvme_config_discard() 2006 struct queue_limits *lim) in nvme_set_ctrl_limits() argument 2014 lim->dma_alignment = 3; in nvme_set_ctrl_limits() 2018 struct queue_limits *lim) in nvme_update_disk_info() argument 2065 lim->io_min = phys_bs; in nvme_update_disk_info() 2066 lim->io_opt = io_opt; in nvme_update_disk_info() 2114 lim->chunk_sectors = iob; in nvme_set_chunk_sectors() 2120 struct queue_limits lim; in nvme_update_ns_info_generic() local 2139 struct queue_limits lim; in nvme_update_ns_info_block() local 2275 struct queue_limits lim; in nvme_update_ns_info() local [all …]
|
| /linux/tools/testing/selftests/mm/ |
| A D | compaction_test.c | 199 struct rlimit lim; in main() local 219 lim.rlim_cur = RLIM_INFINITY; in main() 220 lim.rlim_max = RLIM_INFINITY; in main() 221 if (setrlimit(RLIMIT_MEMLOCK, &lim)) in main()
|