Home
last modified time | relevance | path

Searched refs:flush (Results 1 – 25 of 692) sorted by relevance

12345678910>>...28

/linux/arch/x86/hyperv/
A Dmmu.c64 struct hv_tlb_flush *flush; in hyperv_flush_tlb_multi() local
78 if (unlikely(!flush)) { in hyperv_flush_tlb_multi()
90 flush->flags = 0; in hyperv_flush_tlb_multi()
92 flush->address_space = 0; in hyperv_flush_tlb_multi()
96 flush->processor_mask = 0; in hyperv_flush_tlb_multi()
143 max_gvas = (PAGE_SIZE - sizeof(*flush)) / sizeof(flush->gva_list[0]); in hyperv_flush_tlb_multi()
148 flush, NULL); in hyperv_flush_tlb_multi()
152 flush, NULL); in hyperv_flush_tlb_multi()
192 flush->flags = 0; in hyperv_flush_tlb_others_ex()
194 flush->address_space = 0; in hyperv_flush_tlb_others_ex()
[all …]
A Dnested.c22 struct hv_guest_mapping_flush *flush; in hyperv_flush_guest_mapping() local
32 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping()
34 if (unlikely(!flush)) { in hyperv_flush_guest_mapping()
39 flush->address_space = as; in hyperv_flush_guest_mapping()
40 flush->flags = 0; in hyperv_flush_guest_mapping()
43 flush, NULL); in hyperv_flush_guest_mapping()
56 struct hv_guest_mapping_flush_list *flush, in hyperv_fill_flush_guest_mapping_list() argument
102 if (unlikely(!flush)) { in hyperv_flush_guest_mapping_range()
107 flush->address_space = as; in hyperv_flush_guest_mapping_range()
108 flush->flags = 0; in hyperv_flush_guest_mapping_range()
[all …]
/linux/drivers/md/dm-vdo/
A Dflush.c114 if (flush != NULL) { in allocate_flush()
121 return flush; in allocate_flush()
251 struct vdo_flush *flush = in notify_flush() local
279 vdo_complete_flush(flush); in flush_vdo()
327 struct vdo_flush *flush = in vdo_complete_flushes() local
339 vdo_complete_flush(flush); in vdo_complete_flushes()
371 bio_list_init(&flush->bios); in initialize_flush()
410 if (flush == NULL) { in vdo_launch_flush()
420 launch_flush(flush); in vdo_launch_flush()
448 launch_flush(flush); in release_flush()
[all …]
/linux/lib/
A Ddecompress_inflate.c44 long (*flush)(void*, unsigned long), in __gunzip()
53 if (flush) { in __gunzip()
82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip()
138 if (!flush) { in __gunzip()
159 if (flush && strm->next_out > out_buf) { in __gunzip()
161 if (l != flush(out_buf, l)) { in __gunzip()
193 if (flush) in __gunzip()
202 long (*flush)(void*, unsigned long), in gunzip()
207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip()
212 long (*flush)(void*, unsigned long), in __decompress()
[all …]
A Ddecompress_unxz.c260 long (*flush)(void *src, unsigned long size), in unxz()
276 if (fill == NULL && flush == NULL) in unxz()
284 if (flush == NULL) { in unxz()
306 if (fill == NULL && flush == NULL) { in unxz()
332 if (flush != NULL && (b.out_pos == b.out_size in unxz()
339 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz()
349 if (flush != NULL) in unxz()
389 if (flush != NULL) in unxz()
407 long (*flush)(void *src, unsigned long size), in __decompress()
412 return unxz(in, in_size, fill, flush, out, in_used, error); in __decompress()
A Ddecompress_unzstd.c167 long (*flush)(void*, unsigned long), in __unzstd()
191 if (fill == NULL && flush == NULL) in __unzstd()
227 if (flush != NULL) { in __unzstd()
308 if (flush != NULL && out.pos > 0) { in __unzstd()
309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd()
335 long (*flush)(void*, unsigned long), in unzstd()
340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd()
345 long (*flush)(void*, unsigned long), in __decompress()
350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
A Ddecompress_unlz4.c33 long (*flush)(void *, unsigned long), in unlz4()
52 } else if (!flush) { in unlz4()
174 if (flush && flush(outp, dest_len) != dest_len) in unlz4()
209 long (*flush)(void*, unsigned long), in __decompress()
215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
A Ddecompress_unlzo.c99 long (*flush)(void *, unsigned long), in unlzo()
112 } else if (!flush) { in unlzo()
243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo()
279 long (*flush)(void*, unsigned long), in __decompress()
284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
/linux/fs/btrfs/
A Dspace-info.c491 flush)) { in btrfs_try_granting_tickets()
1096 space_info->flush = 0; in btrfs_async_reclaim_metadata_space()
1108 space_info->flush = 0; in btrfs_async_reclaim_metadata_space()
1224 flush = FLUSH_DELALLOC; in btrfs_preempt_reclaim_metadata_space()
1229 flush = COMMIT_TRANS; in btrfs_preempt_reclaim_metadata_space()
1313 space_info->flush = 0; in btrfs_async_reclaim_data_space()
1324 space_info->flush = 0; in btrfs_async_reclaim_data_space()
1341 space_info->flush = 0; in btrfs_async_reclaim_data_space()
1374 space_info->flush = 0; in btrfs_async_reclaim_data_space()
1546 switch (flush) { in handle_reserve_ticket()
[all …]
A Ddelalloc-space.c118 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_alloc_data_chunk_ondemand() local
124 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_alloc_data_chunk_ondemand()
126 return btrfs_reserve_data_bytes(fs_info, bytes, flush); in btrfs_alloc_data_chunk_ondemand()
134 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_check_data_free_space() local
143 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_check_data_free_space()
145 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_check_data_free_space()
147 ret = btrfs_reserve_data_bytes(fs_info, len, flush); in btrfs_check_data_free_space()
313 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_ALL; in btrfs_delalloc_reserve_metadata() local
326 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_delalloc_reserve_metadata()
329 flush = BTRFS_RESERVE_FLUSH_LIMIT; in btrfs_delalloc_reserve_metadata()
[all …]
/linux/net/ipv4/
A Dtcp_offload.c315 int flush = 1; in tcp_gro_receive() local
331 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive()
345 flush |= (len - 1) >= mss; in tcp_gro_receive()
348 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive()
374 flush = len < mss; in tcp_gro_receive()
383 NAPI_GRO_CB(skb)->flush |= (flush != 0); in tcp_gro_receive()
444 if (!NAPI_GRO_CB(skb)->flush && in tcp4_gro_receive()
447 goto flush; in tcp4_gro_receive()
451 goto flush; in tcp4_gro_receive()
457 flush: in tcp4_gro_receive()
[all …]
A Dudp_offload.c484 int flush; in udp_gro_receive_segment() local
488 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
495 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
514 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
566 int flush = 1; in udp_gro_receive() local
601 flush = 0; in udp_gro_receive()
651 goto flush; in udp4_gro_receive()
654 if (NAPI_GRO_CB(skb)->flush) in udp4_gro_receive()
659 goto flush; in udp4_gro_receive()
672 flush: in udp4_gro_receive()
[all …]
/linux/include/net/
A Dgro.h37 u16 flush; member
368 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final()
372 int flush, in skb_gro_flush_final_remcsum() argument
376 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum()
384 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final()
388 int flush, in skb_gro_flush_final_remcsum() argument
391 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum()
452 int flush; in inet_gro_flush() local
458 return flush; in inet_gro_flush()
498 int flush; in gro_receive_network_flush() local
[all …]
/linux/lib/zlib_deflate/
A Ddeflate.c331 int flush in zlib_deflate() argument
338 flush > Z_FINISH || flush < 0) { in zlib_deflate()
351 s->last_flush = flush; in zlib_deflate()
394 flush != Z_FINISH) { in zlib_deflate()
453 if (flush != Z_FINISH) return Z_OK; in zlib_deflate()
866 int flush in deflate_stored() argument
912 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_stored()
925 int flush in deflate_fast() argument
1008 FLUSH_BLOCK(s, flush == Z_FINISH); in deflate_fast()
1019 int flush in deflate_slow() argument
[all …]
/linux/tools/testing/selftests/drivers/net/netdevsim/
A Dnexthop.sh147 $IP nexthop flush &> /dev/null
182 $IP nexthop flush &> /dev/null
202 $IP nexthop flush &> /dev/null
247 $IP nexthop flush &> /dev/null
267 $IP nexthop flush &> /dev/null
289 $IP nexthop flush &> /dev/null
314 $IP nexthop flush &> /dev/null
343 $IP nexthop flush &> /dev/null
373 $IP nexthop flush &> /dev/null
422 $IP nexthop flush &> /dev/null
[all …]
/linux/Documentation/arch/x86/
A Dtlb.rst12 from areas other than the one we are trying to flush will be
21 1. The size of the flush being performed. A flush of the entire
25 be no collateral damage caused by doing the global flush, and
26 all of the individual flush will have ended up being wasted
29 damage we do with a full flush. So, the larger the TLB, the
30 more attractive an individual flush looks. Data and
37 especially the contents of the TLB during a given flush. The
38 sizes of the flush will vary greatly depending on the workload as
48 This will cause us to do the global flush for more cases.
53 Despite the fact that a single individual flush on x86 is
[all …]
/linux/lib/zlib_dfltcc/
A Ddfltcc_deflate.c121 int flush, in dfltcc_deflate() argument
136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate()
159 if (flush == Z_FINISH) in dfltcc_deflate()
162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate()
225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate()
292 if (flush == Z_FINISH) { in dfltcc_deflate()
301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
A Ddfltcc_inflate.h16 int flush, int *ret);
20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument
25 action = dfltcc_inflate((strm), (flush), &ret); \
/linux/tools/testing/selftests/kvm/x86_64/
A Dhyperv_tlb_flush.c217 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code()
218 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
229 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code()
230 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
231 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code()
245 flush->processor_mask = 0; in sender_guest_code()
258 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code()
373 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
386 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
387 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code()
[all …]
/linux/drivers/md/
A Ddm-delay.c41 struct delay_class flush; member
167 if (dc->flush.dev) in delay_dtr()
168 dm_put_device(ti, dc->flush.dev); in delay_dtr()
244 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr()
256 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr()
262 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr()
265 max_delay = max(max_delay, dc->flush.delay); in delay_ctr()
359 c = &dc->flush; in delay_map()
383 DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops); in delay_status()
394 DMEMIT_DELAY_CLASS(&dc->flush); in delay_status()
[all …]
/linux/drivers/gpu/drm/etnaviv/
A Detnaviv_buffer.c93 u32 flush = 0; in etnaviv_cmd_select_pipe() local
104 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_cmd_select_pipe()
106 flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR; in etnaviv_cmd_select_pipe()
108 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_cmd_select_pipe()
237 u32 link_target, flush = 0; in etnaviv_buffer_end() local
244 flush = VIVS_GL_FLUSH_CACHE_PE2D; in etnaviv_buffer_end()
246 flush = VIVS_GL_FLUSH_CACHE_DEPTH | in etnaviv_buffer_end()
252 if (flush) { in etnaviv_buffer_end()
268 CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush); in etnaviv_buffer_end()
415 u32 flush = VIVS_MMUv2_CONFIGURATION_MODE_MASK | in etnaviv_buffer_queue() local
[all …]
/linux/Documentation/block/
A Dstat.rst44 flush I/Os requests number of flush I/Os processed
45 flush ticks milliseconds total wait time for flush requests
53 flush I/Os
56 These values increment when an flush I/O request completes.
58 Block layer combines flush requests and executes at most one at a time.
59 This counts flush requests executed by disk. Not tracked for partitions.
75 read ticks, write ticks, discard ticks, flush ticks
/linux/drivers/gpu/drm/vmwgfx/
A Dvmwgfx_overlay.c63 SVGAEscapeVideoFlush flush; member
77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush()
78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush()
79 cmd->flush.streamId = stream_id; in fill_flush()
93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local
117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put()
125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put()
164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put()
184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member
206 fill_flush(&cmds->flush, stream_id); in vmw_overlay_send_stop()
/linux/Documentation/admin-guide/device-mapper/
A Ddelay.rst15 3: apply offset and delay to read, write and flush operations on device
18 to write and flush operations on optionally different write_device with
35 # Create mapped device named "delayed" delaying read, write and flush operations for 500ms.
42 # Create mapped device delaying write and flush operations for 400ms and
/linux/arch/arm/mm/
A Dcache-v4.S41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache
123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache

Completed in 46 milliseconds

12345678910>>...28