Lines Matching refs:t
163 #define RET_STACK(t, offset) ((struct ftrace_ret_stack *)(&(t)->ret_stack[offset])) argument
249 static inline unsigned long get_fgraph_entry(struct task_struct *t, int offset) in get_fgraph_entry() argument
251 return t->ret_stack[offset]; in get_fgraph_entry()
255 static inline int get_frame_offset(struct task_struct *t, int offset) in get_frame_offset() argument
257 return __get_offset(t->ret_stack[offset]); in get_frame_offset()
262 get_bitmap_bits(struct task_struct *t, int offset) in get_bitmap_bits() argument
264 return (t->ret_stack[offset] >> FGRAPH_INDEX_SHIFT) & FGRAPH_INDEX_MASK; in get_bitmap_bits()
269 set_bitmap(struct task_struct *t, int offset, unsigned long bitmap) in set_bitmap() argument
271 t->ret_stack[offset] = (bitmap << FGRAPH_INDEX_SHIFT) | in set_bitmap()
276 static inline void *get_data_type_data(struct task_struct *t, int offset) in get_data_type_data() argument
278 unsigned long val = t->ret_stack[offset]; in get_data_type_data()
283 return (void *)&t->ret_stack[offset]; in get_data_type_data()
307 static void ret_stack_set_task_var(struct task_struct *t, int idx, long val) in ret_stack_set_task_var() argument
309 unsigned long *gvals = SHADOW_STACK_TASK_VARS(t->ret_stack); in ret_stack_set_task_var()
315 ret_stack_get_task_var(struct task_struct *t, int idx) in ret_stack_get_task_var() argument
317 unsigned long *gvals = SHADOW_STACK_TASK_VARS(t->ret_stack); in ret_stack_get_task_var()
434 get_ret_stack(struct task_struct *t, int offset, int *frame_offset) in get_ret_stack() argument
443 offs = get_frame_offset(t, --offset); in get_ret_stack()
450 return RET_STACK(t, offset); in get_ret_stack()
1036 struct task_struct *g, *t; in alloc_retstack_tasklist() local
1052 for_each_process_thread(g, t) { in alloc_retstack_tasklist()
1058 if (t->ret_stack == NULL) { in alloc_retstack_tasklist()
1059 atomic_set(&t->trace_overrun, 0); in alloc_retstack_tasklist()
1061 t->curr_ret_stack = 0; in alloc_retstack_tasklist()
1062 t->curr_ret_depth = -1; in alloc_retstack_tasklist()
1065 t->ret_stack = ret_stack_list[start++]; in alloc_retstack_tasklist()
1106 graph_init_task(struct task_struct *t, unsigned long *ret_stack) in graph_init_task() argument
1108 atomic_set(&t->trace_overrun, 0); in graph_init_task()
1110 t->ftrace_timestamp = 0; in graph_init_task()
1111 t->curr_ret_stack = 0; in graph_init_task()
1112 t->curr_ret_depth = -1; in graph_init_task()
1115 t->ret_stack = ret_stack; in graph_init_task()
1122 void ftrace_graph_init_idle_task(struct task_struct *t, int cpu) in ftrace_graph_init_idle_task() argument
1124 t->curr_ret_stack = 0; in ftrace_graph_init_idle_task()
1125 t->curr_ret_depth = -1; in ftrace_graph_init_idle_task()
1130 if (t->ret_stack) in ftrace_graph_init_idle_task()
1131 WARN_ON(t->ret_stack != per_cpu(idle_ret_stack, cpu)); in ftrace_graph_init_idle_task()
1146 graph_init_task(t, ret_stack); in ftrace_graph_init_idle_task()
1151 void ftrace_graph_init_task(struct task_struct *t) in ftrace_graph_init_task() argument
1154 t->ret_stack = NULL; in ftrace_graph_init_task()
1155 t->curr_ret_stack = 0; in ftrace_graph_init_task()
1156 t->curr_ret_depth = -1; in ftrace_graph_init_task()
1167 graph_init_task(t, ret_stack); in ftrace_graph_init_task()
1171 void ftrace_graph_exit_task(struct task_struct *t) in ftrace_graph_exit_task() argument
1173 unsigned long *ret_stack = t->ret_stack; in ftrace_graph_exit_task()
1175 t->ret_stack = NULL; in ftrace_graph_exit_task()
1261 struct task_struct *g, *t; in init_task_vars() local
1270 for_each_process_thread(g, t) { in init_task_vars()
1271 if (t->ret_stack) in init_task_vars()
1272 ret_stack_set_task_var(t, idx, 0); in init_task_vars()