| /include/asm-generic/ |
| A D | rqspinlock.h | 70 int cnt; member 78 int cnt = this_cpu_inc_return(rqspinlock_held_locks.cnt); in grab_held_lock_entry() local 80 if (unlikely(cnt > RES_NR_HELD)) { in grab_held_lock_entry() 95 this_cpu_write(rqspinlock_held_locks.locks[cnt - 1], lock); in grab_held_lock_entry() 114 if (unlikely(rqh->cnt > RES_NR_HELD)) in release_held_lock_entry() 116 WRITE_ONCE(rqh->locks[rqh->cnt - 1], NULL); in release_held_lock_entry() 160 this_cpu_dec(rqspinlock_held_locks.cnt); in release_held_lock_entry() 195 if (unlikely(rqh->cnt > RES_NR_HELD)) in res_spin_unlock() 197 WRITE_ONCE(rqh->locks[rqh->cnt - 1], NULL); in res_spin_unlock() 217 this_cpu_dec(rqspinlock_held_locks.cnt); in res_spin_unlock()
|
| /include/linux/ |
| A D | rcuref.h | 24 static inline void rcuref_init(rcuref_t *ref, unsigned int cnt) in rcuref_init() argument 26 atomic_set(&ref->refcnt, cnt - 1); in rcuref_init() 94 extern __must_check bool rcuref_put_slowpath(rcuref_t *ref, unsigned int cnt); 101 int cnt; in __rcuref_put() local 109 cnt = atomic_sub_return_release(1, &ref->refcnt); in __rcuref_put() 110 if (likely(cnt >= 0)) in __rcuref_put() 117 return rcuref_put_slowpath(ref, cnt); in __rcuref_put()
|
| A D | file_ref.h | 59 static inline void file_ref_init(file_ref_t *ref, unsigned long cnt) in file_ref_init() argument 61 atomic_long_set(&ref->refcnt, cnt - 1); in file_ref_init() 64 bool __file_ref_put(file_ref_t *ref, unsigned long cnt); 141 long cnt; in file_ref_put() local 157 cnt = atomic_long_dec_return(&ref->refcnt); in file_ref_put() 158 if (cnt >= 0) in file_ref_put() 160 return __file_ref_put(ref, cnt); in file_ref_put()
|
| A D | bottom_half.h | 9 extern void __local_bh_disable_ip(unsigned long ip, unsigned int cnt); 11 static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) in __local_bh_disable_ip() argument 13 preempt_count_add(cnt); in __local_bh_disable_ip() 24 extern void __local_bh_enable_ip(unsigned long ip, unsigned int cnt);
|
| A D | eventfd.h | 40 __u64 *cnt); 41 void eventfd_ctx_do_read(struct eventfd_ctx *ctx, __u64 *cnt); 70 wait_queue_entry_t *wait, __u64 *cnt) in eventfd_ctx_remove_wait_queue() argument 80 static inline void eventfd_ctx_do_read(struct eventfd_ctx *ctx, __u64 *cnt) in eventfd_ctx_do_read() argument
|
| A D | mbcache.h | 51 unsigned int cnt = atomic_dec_return(&entry->e_refcnt); in mb_cache_entry_put() local 53 if (cnt > 0) { in mb_cache_entry_put() 54 if (cnt <= 2) in mb_cache_entry_put()
|
| A D | dmar.h | 97 #define for_each_dev_scope(devs, cnt, i, tmp) \ argument 98 for ((i) = 0; ((tmp) = (i) < (cnt) ? \ 99 dmar_rcu_dereference((devs)[(i)].dev) : NULL, (i) < (cnt)); \ 102 #define for_each_active_dev_scope(devs, cnt, i, tmp) \ argument 103 for_each_dev_scope((devs), (cnt), (i), (tmp)) \ 109 extern void *dmar_alloc_dev_scope(void *start, void *end, int *cnt); 110 extern void dmar_free_dev_scope(struct dmar_dev_scope **devices, int *cnt);
|
| A D | unwind_deferred_types.h | 27 u32 cnt; member
|
| A D | trace_seq.h | 95 int cnt); 132 int cnt) in trace_seq_to_user() argument
|
| A D | irq.h | 934 int __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, 939 unsigned int cnt, int node, struct module *owner, 943 #define irq_alloc_descs(irq, from, cnt, node) \ argument 944 __irq_alloc_descs(irq, from, cnt, node, THIS_MODULE, NULL) 955 #define irq_alloc_descs_from(from, cnt, node) \ argument 956 irq_alloc_descs(-1, from, cnt, node) 958 #define devm_irq_alloc_descs(dev, irq, from, cnt, node) \ argument 959 __devm_irq_alloc_descs(dev, irq, from, cnt, node, THIS_MODULE, NULL) 970 #define devm_irq_alloc_descs_from(dev, from, cnt, node) \ argument 971 devm_irq_alloc_descs(dev, -1, from, cnt, node) [all …]
|
| A D | ftrace.h | 498 int ftrace_lookup_symbols(const char **sorted_syms, size_t cnt, unsigned long *addrs); 509 static inline int ftrace_lookup_symbols(const char **sorted_syms, size_t cnt, unsigned long *addrs) in ftrace_lookup_symbols() argument 724 unsigned int cnt, int remove, int reset); 794 size_t cnt, loff_t *ppos); 796 size_t cnt, loff_t *ppos); 1004 #define ftrace_set_filter_ips(ops, ips, cnt, remove, reset) ({ -ENODEV; }) argument 1011 size_t cnt, loff_t *ppos) { return -ENODEV; } in ftrace_filter_write() argument 1013 size_t cnt, loff_t *ppos) { return -ENODEV; } in ftrace_notrace_write() argument
|
| A D | backing-dev-defs.h | 64 atomic_t cnt; member 69 (struct wb_completion){ .cnt = ATOMIC_INIT(1), .waitq = (_waitq) }
|
| A D | posix-clock.h | 65 char __user *buf, size_t cnt);
|
| A D | irqdesc.h | 26 unsigned int cnt; member 134 return desc->kstat_irqs ? per_cpu(desc->kstat_irqs->cnt, cpu) : 0; in irq_desc_kstat_cpu()
|
| A D | dcache.h | 36 #define bytemask_from_count(cnt) (~(~0ul << (cnt)*8)) argument 39 #define bytemask_from_count(cnt) (~(~0ul >> (cnt)*8)) argument
|
| A D | btf_ids.h | 9 u32 cnt; member 17 u32 cnt; member
|
| /include/trace/events/ |
| A D | nilfs2.h | 102 unsigned long cnt), 104 TP_ARGS(sufile, segnum, cnt), 109 __field(unsigned long, cnt) 115 __entry->cnt = cnt; 121 __entry->cnt)
|
| A D | page_pool.h | 26 __field(u64, cnt) 34 __entry->cnt = pool->destroy_cnt; 39 __entry->release, __entry->cnt)
|
| A D | rcu.h | 802 TP_PROTO(const char *rcuname, const char *s, int cpu, int cnt, unsigned long done), 804 TP_ARGS(rcuname, s, cpu, cnt, done), 810 __field(int, cnt) 818 __entry->cnt = cnt; 823 __entry->rcuname, __entry->s, __entry->cpu, __entry->cnt,
|
| /include/linux/platform_data/ |
| A D | sh_mmcif.h | 102 int cnt; in sh_mmcif_boot_cmd_poll() local 104 for (cnt = 0; cnt < 1000000; cnt++) { in sh_mmcif_boot_cmd_poll()
|
| /include/linux/netfilter/ |
| A D | x_tables.h | 420 void xt_percpu_counter_free(struct xt_counters *cnt); 423 xt_get_this_cpu_counter(struct xt_counters *cnt) in xt_get_this_cpu_counter() argument 426 return this_cpu_ptr((void __percpu *) (unsigned long) cnt->pcnt); in xt_get_this_cpu_counter() 428 return cnt; in xt_get_this_cpu_counter() 432 xt_get_per_cpu_counter(struct xt_counters *cnt, unsigned int cpu) in xt_get_per_cpu_counter() argument 435 return per_cpu_ptr((void __percpu *) (unsigned long) cnt->pcnt, cpu); in xt_get_per_cpu_counter() 437 return cnt; in xt_get_per_cpu_counter()
|
| /include/linux/usb/ |
| A D | typec_altmode.h | 64 const u32 *vdo, int cnt); 90 const u32 hdr, const u32 *vdo, int cnt);
|
| A D | pd.h | 94 #define PD_HEADER(type, pwr, data, rev, id, cnt, ext_hdr) \ argument 100 (((cnt) & PD_HEADER_CNT_MASK) << PD_HEADER_CNT_SHIFT) | \ 103 #define PD_HEADER_LE(type, pwr, data, rev, id, cnt) \ argument 104 cpu_to_le16(PD_HEADER((type), (pwr), (data), (rev), (id), (cnt), (0)))
|
| /include/linux/ceph/ |
| A D | osd_client.h | 559 int __ceph_alloc_sparse_ext_map(struct ceph_osd_req_op *op, int cnt); 568 static inline int ceph_alloc_sparse_ext_map(struct ceph_osd_req_op *op, int cnt) in ceph_alloc_sparse_ext_map() argument 570 if (!cnt) in ceph_alloc_sparse_ext_map() 571 cnt = CEPH_SPARSE_EXT_ARRAY_INITIAL; in ceph_alloc_sparse_ext_map() 573 return __ceph_alloc_sparse_ext_map(op, cnt); in ceph_alloc_sparse_ext_map()
|
| /include/uapi/drm/ |
| A D | panthor_drm.h | 183 #define DRM_PANTHOR_OBJ_ARRAY(cnt, ptr) \ argument 184 { .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) }
|