| /linux/arch/x86/hyperv/ |
| A D | mmu.c | 64 struct hv_tlb_flush *flush; in hyperv_flush_tlb_multi() local 78 if (unlikely(!flush)) { in hyperv_flush_tlb_multi() 90 flush->flags = 0; in hyperv_flush_tlb_multi() 92 flush->address_space = 0; in hyperv_flush_tlb_multi() 96 flush->processor_mask = 0; in hyperv_flush_tlb_multi() 143 max_gvas = (PAGE_SIZE - sizeof(*flush)) / sizeof(flush->gva_list[0]); in hyperv_flush_tlb_multi() 148 flush, NULL); in hyperv_flush_tlb_multi() 152 flush, NULL); in hyperv_flush_tlb_multi() 192 flush->flags = 0; in hyperv_flush_tlb_others_ex() 194 flush->address_space = 0; in hyperv_flush_tlb_others_ex() [all …]
|
| A D | nested.c | 22 struct hv_guest_mapping_flush *flush; in hyperv_flush_guest_mapping() local 32 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping() 34 if (unlikely(!flush)) { in hyperv_flush_guest_mapping() 39 flush->address_space = as; in hyperv_flush_guest_mapping() 40 flush->flags = 0; in hyperv_flush_guest_mapping() 43 flush, NULL); in hyperv_flush_guest_mapping() 56 struct hv_guest_mapping_flush_list *flush, in hyperv_fill_flush_guest_mapping_list() argument 102 if (unlikely(!flush)) { in hyperv_flush_guest_mapping_range() 107 flush->address_space = as; in hyperv_flush_guest_mapping_range() 108 flush->flags = 0; in hyperv_flush_guest_mapping_range() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| A D | flush.c | 114 if (flush != NULL) { in allocate_flush() 121 return flush; in allocate_flush() 251 struct vdo_flush *flush = in notify_flush() local 279 vdo_complete_flush(flush); in flush_vdo() 327 struct vdo_flush *flush = in vdo_complete_flushes() local 339 vdo_complete_flush(flush); in vdo_complete_flushes() 371 bio_list_init(&flush->bios); in initialize_flush() 410 if (flush == NULL) { in vdo_launch_flush() 420 launch_flush(flush); in vdo_launch_flush() 448 launch_flush(flush); in release_flush() [all …]
|
| /linux/lib/ |
| A D | decompress_inflate.c | 44 long (*flush)(void*, unsigned long), in __gunzip() 53 if (flush) { in __gunzip() 82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip() 138 if (!flush) { in __gunzip() 159 if (flush && strm->next_out > out_buf) { in __gunzip() 161 if (l != flush(out_buf, l)) { in __gunzip() 193 if (flush) in __gunzip() 202 long (*flush)(void*, unsigned long), in gunzip() 207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip() 212 long (*flush)(void*, unsigned long), in __decompress() [all …]
|
| A D | decompress_unxz.c | 260 long (*flush)(void *src, unsigned long size), in unxz() 276 if (fill == NULL && flush == NULL) in unxz() 284 if (flush == NULL) { in unxz() 306 if (fill == NULL && flush == NULL) { in unxz() 332 if (flush != NULL && (b.out_pos == b.out_size in unxz() 339 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz() 349 if (flush != NULL) in unxz() 389 if (flush != NULL) in unxz() 407 long (*flush)(void *src, unsigned long size), in __decompress() 412 return unxz(in, in_size, fill, flush, out, in_used, error); in __decompress()
|
| A D | decompress_unzstd.c | 167 long (*flush)(void*, unsigned long), in __unzstd() 191 if (fill == NULL && flush == NULL) in __unzstd() 227 if (flush != NULL) { in __unzstd() 308 if (flush != NULL && out.pos > 0) { in __unzstd() 309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd() 335 long (*flush)(void*, unsigned long), in unzstd() 340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd() 345 long (*flush)(void*, unsigned long), in __decompress() 350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
|
| A D | decompress_unlz4.c | 33 long (*flush)(void *, unsigned long), in unlz4() 52 } else if (!flush) { in unlz4() 174 if (flush && flush(outp, dest_len) != dest_len) in unlz4() 209 long (*flush)(void*, unsigned long), in __decompress() 215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
| A D | decompress_unlzo.c | 99 long (*flush)(void *, unsigned long), in unlzo() 112 } else if (!flush) { in unlzo() 243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo() 279 long (*flush)(void*, unsigned long), in __decompress() 284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
|
| /linux/fs/btrfs/ |
| A D | space-info.c | 491 flush)) { in btrfs_try_granting_tickets() 1096 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1108 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1224 flush = FLUSH_DELALLOC; in btrfs_preempt_reclaim_metadata_space() 1229 flush = COMMIT_TRANS; in btrfs_preempt_reclaim_metadata_space() 1313 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1324 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1341 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1374 space_info->flush = 0; in btrfs_async_reclaim_data_space() 1546 switch (flush) { in handle_reserve_ticket() [all …]
|
| A D | delalloc-space.c | 118 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_alloc_data_chunk_ondemand() local 124 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_alloc_data_chunk_ondemand() 126 return btrfs_reserve_data_bytes(fs_info, bytes, flush); in btrfs_alloc_data_chunk_ondemand() 134 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_check_data_free_space() local 143 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_check_data_free_space() 145 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_check_data_free_space() 147 ret = btrfs_reserve_data_bytes(fs_info, len, flush); in btrfs_check_data_free_space() 313 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_ALL; in btrfs_delalloc_reserve_metadata() local 326 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_delalloc_reserve_metadata() 329 flush = BTRFS_RESERVE_FLUSH_LIMIT; in btrfs_delalloc_reserve_metadata() [all …]
|
| /linux/net/ipv4/ |
| A D | tcp_offload.c | 315 int flush = 1; in tcp_gro_receive() local 331 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive() 345 flush |= (len - 1) >= mss; in tcp_gro_receive() 348 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive() 374 flush = len < mss; in tcp_gro_receive() 383 NAPI_GRO_CB(skb)->flush |= (flush != 0); in tcp_gro_receive() 444 if (!NAPI_GRO_CB(skb)->flush && in tcp4_gro_receive() 447 goto flush; in tcp4_gro_receive() 451 goto flush; in tcp4_gro_receive() 457 flush: in tcp4_gro_receive() [all …]
|
| A D | udp_offload.c | 484 int flush; in udp_gro_receive_segment() local 488 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 495 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 514 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 566 int flush = 1; in udp_gro_receive() local 601 flush = 0; in udp_gro_receive() 651 goto flush; in udp4_gro_receive() 654 if (NAPI_GRO_CB(skb)->flush) in udp4_gro_receive() 659 goto flush; in udp4_gro_receive() 672 flush: in udp4_gro_receive() [all …]
|
| /linux/include/net/ |
| A D | gro.h | 37 u16 flush; member 368 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 372 int flush, in skb_gro_flush_final_remcsum() argument 376 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum() 384 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 388 int flush, in skb_gro_flush_final_remcsum() argument 391 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum() 452 int flush; in inet_gro_flush() local 458 return flush; in inet_gro_flush() 498 int flush; in gro_receive_network_flush() local [all …]
|
| /linux/lib/zlib_deflate/ |
| A D | deflate.c | 331 int flush in zlib_deflate() argument 338 flush > Z_FINISH || flush < 0) { in zlib_deflate() 351 s->last_flush = flush; in zlib_deflate() 394 flush != Z_FINISH) { in zlib_deflate() 453 if (flush != Z_FINISH) return Z_OK; in zlib_deflate() 866 int flush in deflate_stored() argument 912 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_stored() 925 int flush in deflate_fast() argument 1008 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_fast() 1019 int flush in deflate_slow() argument [all …]
|
| /linux/tools/testing/selftests/drivers/net/netdevsim/ |
| A D | nexthop.sh | 147 $IP nexthop flush &> /dev/null 182 $IP nexthop flush &> /dev/null 202 $IP nexthop flush &> /dev/null 247 $IP nexthop flush &> /dev/null 267 $IP nexthop flush &> /dev/null 289 $IP nexthop flush &> /dev/null 314 $IP nexthop flush &> /dev/null 343 $IP nexthop flush &> /dev/null 373 $IP nexthop flush &> /dev/null 422 $IP nexthop flush &> /dev/null [all …]
|
| /linux/Documentation/arch/x86/ |
| A D | tlb.rst | 12 from areas other than the one we are trying to flush will be 21 1. The size of the flush being performed. A flush of the entire 25 be no collateral damage caused by doing the global flush, and 26 all of the individual flush will have ended up being wasted 29 damage we do with a full flush. So, the larger the TLB, the 30 more attractive an individual flush looks. Data and 37 especially the contents of the TLB during a given flush. The 38 sizes of the flush will vary greatly depending on the workload as 48 This will cause us to do the global flush for more cases. 53 Despite the fact that a single individual flush on x86 is [all …]
|
| /linux/lib/zlib_dfltcc/ |
| A D | dfltcc_deflate.c | 121 int flush, in dfltcc_deflate() argument 136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate() 159 if (flush == Z_FINISH) in dfltcc_deflate() 162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate() 225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate() 292 if (flush == Z_FINISH) { in dfltcc_deflate() 301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
|
| A D | dfltcc_inflate.h | 16 int flush, int *ret); 20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument 25 action = dfltcc_inflate((strm), (flush), &ret); \
|
| /linux/tools/testing/selftests/kvm/x86_64/ |
| A D | hyperv_tlb_flush.c | 217 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 218 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 229 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 230 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 231 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() 245 flush->processor_mask = 0; in sender_guest_code() 258 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() 373 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 386 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 387 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() [all …]
|
| /linux/drivers/md/ |
| A D | dm-delay.c | 41 struct delay_class flush; member 167 if (dc->flush.dev) in delay_dtr() 168 dm_put_device(ti, dc->flush.dev); in delay_dtr() 244 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr() 256 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr() 262 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr() 265 max_delay = max(max_delay, dc->flush.delay); in delay_ctr() 359 c = &dc->flush; in delay_map() 383 DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops); in delay_status() 394 DMEMIT_DELAY_CLASS(&dc->flush); in delay_status() [all …]
|
| /linux/drivers/gpu/drm/etnaviv/ |
| A D | etnaviv_buffer.c | 93 u32 flush = 0; in etnaviv_cmd_select_pipe() local 104 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_cmd_select_pipe() 106 flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR; in etnaviv_cmd_select_pipe() 108 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_cmd_select_pipe() 237 u32 link_target, flush = 0; in etnaviv_buffer_end() local 244 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_buffer_end() 246 flush = VIVS_GL_FLUSH_CACHE_DEPTH | in etnaviv_buffer_end() 252 if (flush) { in etnaviv_buffer_end() 268 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_buffer_end() 415 u32 flush = VIVS_MMUv2_CONFIGURATION_MODE_MASK | in etnaviv_buffer_queue() local [all …]
|
| /linux/Documentation/block/ |
| A D | stat.rst | 44 flush I/Os requests number of flush I/Os processed 45 flush ticks milliseconds total wait time for flush requests 53 flush I/Os 56 These values increment when an flush I/O request completes. 58 Block layer combines flush requests and executes at most one at a time. 59 This counts flush requests executed by disk. Not tracked for partitions. 75 read ticks, write ticks, discard ticks, flush ticks
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_overlay.c | 63 SVGAEscapeVideoFlush flush; member 77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush() 78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush() 79 cmd->flush.streamId = stream_id; in fill_flush() 93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local 117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put() 125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put() 164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put() 184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member 206 fill_flush(&cmds->flush, stream_id); in vmw_overlay_send_stop()
|
| /linux/Documentation/admin-guide/device-mapper/ |
| A D | delay.rst | 15 3: apply offset and delay to read, write and flush operations on device 18 to write and flush operations on optionally different write_device with 35 # Create mapped device named "delayed" delaying read, write and flush operations for 500ms. 42 # Create mapped device delaying write and flush operations for 400ms and
|
| /linux/arch/arm/mm/ |
| A D | cache-v4.S | 41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache 61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache 123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
|