| /linux/fs/squashfs/ |
| A D | decompressor_multi_percpu.c | 31 struct squashfs_stream __percpu *percpu; in squashfs_decompressor_create() local 35 if (percpu == NULL) in squashfs_decompressor_create() 39 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 53 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 57 free_percpu(percpu); in squashfs_decompressor_create() 63 struct squashfs_stream __percpu *percpu = in squashfs_decompressor_destroy() local 70 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy() 73 free_percpu(percpu); in squashfs_decompressor_destroy() 85 local_lock(&percpu->lock); in squashfs_decompress() 86 stream = this_cpu_ptr(percpu); in squashfs_decompress() [all …]
|
| /linux/kernel/bpf/ |
| A D | memalloc.c | 255 static void free_one(void *obj, bool percpu) in free_one() argument 257 if (percpu) { in free_one() 272 free_one(pos, percpu); in free_all() 519 if (percpu && size == 0) in bpf_mem_alloc_init() 523 if (percpu) in bpf_mem_alloc_init() 525 ma->percpu = percpu; in bpf_mem_alloc_init() 532 if (!percpu) in bpf_mem_alloc_init() 590 ma->percpu = true; in bpf_mem_alloc_percpu_init() 632 bool percpu = !!c->percpu_size; in drain_mem_cache() local 900 if (!ma->percpu) in bpf_mem_alloc() [all …]
|
| A D | bpf_lru_list.c | 501 if (lru->percpu) in bpf_lru_pop_free() 558 if (lru->percpu) in bpf_lru_push_free() 616 if (lru->percpu) in bpf_lru_populate() 651 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, in bpf_lru_init() argument 656 if (percpu) { in bpf_lru_init() 686 lru->percpu = percpu; in bpf_lru_init() 696 if (lru->percpu) in bpf_lru_destroy()
|
| A D | bpf_lru_list.h | 62 bool percpu; member 71 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
|
| /linux/include/linux/ |
| A D | bpf_mem_alloc.h | 15 bool percpu; member 29 int bpf_mem_alloc_init(struct bpf_mem_alloc *ma, int size, bool percpu); 37 int bpf_mem_alloc_check_size(bool percpu, size_t size);
|
| A D | spinlock_rt.h | 11 struct lock_class_key *key, bool percpu); 14 struct lock_class_key *key, bool percpu) in __rt_spin_lock_init() argument
|
| /linux/arch/alpha/boot/ |
| A D | bootp.c | 71 struct percpu_struct * percpu; in pal_init() local 101 percpu = (struct percpu_struct *) in pal_init() 103 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| A D | main.c | 65 struct percpu_struct * percpu; in pal_init() local 95 percpu = (struct percpu_struct *) in pal_init() 97 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| A D | bootpz.c | 119 struct percpu_struct * percpu; in pal_init() local 149 percpu = (struct percpu_struct *) in pal_init() 151 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| A D | workqueue.rst | 595 events percpu 0 2 4 6 596 events_highpri percpu 1 3 5 7 597 events_long percpu 0 2 4 6 599 events_freezable percpu 0 2 4 6 600 events_power_efficient percpu 0 2 4 6 601 events_freezable_power_ percpu 0 2 4 6 602 rcu_gp percpu 0 2 4 6 603 rcu_par_gp percpu 0 2 4 6 604 slub_flushwq percpu 0 2 4 6
|
| /linux/net/rds/ |
| A D | ib_recv.c | 108 if (!cache->percpu) in rds_ib_recv_alloc_cache() 112 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_recv_alloc_cache() 130 free_percpu(ic->i_cache_incs.percpu); in rds_ib_recv_alloc_caches() 143 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_cache_splice_all_lists() 166 free_percpu(ic->i_cache_incs.percpu); in rds_ib_recv_free_caches() 177 free_percpu(ic->i_cache_frags.percpu); in rds_ib_recv_free_caches() 487 chpfirst = __this_cpu_read(cache->percpu->first); in rds_ib_recv_cache_put() 493 __this_cpu_write(cache->percpu->first, new_item); in rds_ib_recv_cache_put() 494 __this_cpu_inc(cache->percpu->count); in rds_ib_recv_cache_put() 513 __this_cpu_write(cache->percpu->first, NULL); in rds_ib_recv_cache_put() [all …]
|
| /linux/tools/testing/selftests/cgroup/ |
| A D | test_kmem.c | 356 long current, percpu; in test_percpu_basic() local 381 percpu = cg_read_key_long(parent, "memory.stat", "percpu "); in test_percpu_basic() 383 if (current > 0 && percpu > 0 && labs(current - percpu) < in test_percpu_basic() 388 current, percpu); in test_percpu_basic()
|
| /linux/include/asm-generic/ |
| A D | vmlinux.lds.h | 959 *(.data..percpu..decrypted) \ 1037 *(.data..percpu..first) \ 1039 *(.data..percpu..page_aligned) \ 1041 *(.data..percpu..read_mostly) \ 1043 *(.data..percpu) \ 1044 *(.data..percpu..shared_aligned) \ 1074 .data..percpu vaddr : AT(__per_cpu_load - LOAD_OFFSET) { \ 1077 . = __per_cpu_load + SIZEOF(.data..percpu); 1093 .data..percpu : AT(ADDR(.data..percpu) - LOAD_OFFSET) { \
|
| /linux/drivers/md/ |
| A D | raid5.c | 1539 return percpu->scribble + i * percpu->scribble_obj_size; in to_addr_page() 2283 percpu = this_cpu_ptr(conf->percpu); in raid_run_ops() 2475 kvfree(percpu->scribble); in scribble_alloc() 2498 percpu = per_cpu_ptr(conf->percpu, cpu); in resize_chunks() 7300 percpu->spare_page = NULL; in free_scratch_buffer() 7301 kvfree(percpu->scribble); in free_scratch_buffer() 7302 percpu->scribble = NULL; in free_scratch_buffer() 7309 if (!percpu->spare_page) in alloc_scratch_buffer() 7337 if (!conf->percpu) in raid5_free_percpu() 7367 struct raid5_percpu *percpu = per_cpu_ptr(conf->percpu, cpu); in raid456_cpu_up_prepare() local [all …]
|
| /linux/drivers/clocksource/ |
| A D | timer-qcom.c | 153 bool percpu) in msm_timer_init() argument 159 msm_timer_has_ppi = percpu; in msm_timer_init() 167 if (percpu) in msm_timer_init()
|
| /linux/Documentation/trace/coresight/ |
| A D | coresight-trbe.rst | 13 Trace Buffer Extension (TRBE) is a percpu hardware which captures in system 14 memory, CPU traces generated from a corresponding percpu tracing unit. This
|
| /linux/arch/sparc/kernel/ |
| A D | sun4m_irq.c | 107 bool percpu; member 200 if (handler_data->percpu) { in sun4m_mask_irq() 219 if (handler_data->percpu) { in sun4m_unmask_irq() 278 handler_data->percpu = real_irq < OBP_INT_LEVEL_ONBOARD; in sun4m_build_device_irq()
|
| /linux/arch/x86/kernel/ |
| A D | vmlinux.lds.S | 104 percpu PT_LOAD FLAGS(6); /* RW_ */ 231 PERCPU_VADDR(INTERNODE_CACHE_BYTES, 0, :percpu) 232 ASSERT(SIZEOF(.data..percpu) < CONFIG_PHYSICAL_START,
|
| /linux/kernel/sched/ |
| A D | cpuacct.c | 213 u64 percpu; in __cpuacct_percpu_seq_show() local 217 percpu = cpuacct_cpuusage_read(ca, i, index); in __cpuacct_percpu_seq_show() 218 seq_printf(m, "%llu ", (unsigned long long) percpu); in __cpuacct_percpu_seq_show()
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| A D | hyp.lds.S | 25 BEGIN_HYP_SECTION(.data..percpu)
|
| /linux/Documentation/translations/zh_CN/dev-tools/ |
| A D | kmemleak.rst | 140 - ``kmemleak_alloc_percpu`` - 通知一个 percpu 类型的内存分配 144 - ``kmemleak_free_percpu`` - 通知一个 percpu 类型的内存释放
|
| /linux/Documentation/locking/ |
| A D | index.rst | 24 percpu-rw-semaphore
|
| /linux/Documentation/RCU/ |
| A D | rcuref.rst | 8 Please note that the percpu-ref feature is likely your first 10 include/linux/percpu-refcount.h for more information. However, in 11 those unusual cases where percpu-ref would consume too much memory,
|
| /linux/arch/arm64/kernel/ |
| A D | vmlinux.lds.S | 28 HYP_SECTION_NAME(.data..percpu) : { \ 29 *(HYP_SECTION_NAME(.data..percpu)) \
|
| /linux/kernel/locking/ |
| A D | spinlock_rt.c | 136 struct lock_class_key *key, bool percpu) in __rt_spin_lock_init() argument 138 u8 type = percpu ? LD_LOCK_PERCPU : LD_LOCK_NORMAL; in __rt_spin_lock_init()
|