Lines Matching refs:u64

104 static u64 cycles_at_suspend;
106 static u64 dummy_clock_read(struct clocksource *cs) in dummy_clock_read()
186 while (tk->tkr_mono.xtime_nsec >= ((u64)NSEC_PER_SEC << tk->tkr_mono.shift)) { in tk_normalize_xtime()
187 tk->tkr_mono.xtime_nsec -= (u64)NSEC_PER_SEC << tk->tkr_mono.shift; in tk_normalize_xtime()
190 while (tk->tkr_raw.xtime_nsec >= ((u64)NSEC_PER_SEC << tk->tkr_raw.shift)) { in tk_normalize_xtime()
191 tk->tkr_raw.xtime_nsec -= (u64)NSEC_PER_SEC << tk->tkr_raw.shift; in tk_normalize_xtime()
231 tk->tkr_mono.xtime_nsec = (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_set_xtime()
238 tk->tkr_mono.xtime_nsec += (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_xtime_add()
285 static inline u64 tk_clock_read(const struct tk_read_base *tkr) in tk_clock_read()
305 u64 interval; in tk_setup_internals()
306 u64 tmp, ntpinterval; in tk_setup_internals()
328 interval = (u64) tmp; in tk_setup_internals()
367 static noinline u64 delta_to_ns_safe(const struct tk_read_base *tkr, u64 delta) in delta_to_ns_safe()
372 static inline u64 timekeeping_cycles_to_ns(const struct tk_read_base *tkr, u64 cycles) in timekeeping_cycles_to_ns()
375 u64 mask = tkr->mask, delta = (cycles - tkr->cycle_last) & mask; in timekeeping_cycles_to_ns()
396 static __always_inline u64 timekeeping_get_ns(const struct tk_read_base *tkr) in timekeeping_get_ns()
436 static __always_inline u64 __ktime_get_fast_ns(struct tk_fast *tkf) in __ktime_get_fast_ns()
440 u64 now; in __ktime_get_fast_ns()
484 u64 notrace ktime_get_mono_fast_ns(void) in ktime_get_mono_fast_ns()
496 u64 notrace ktime_get_raw_fast_ns(void) in ktime_get_raw_fast_ns()
526 u64 notrace ktime_get_boot_fast_ns(void) in ktime_get_boot_fast_ns()
543 u64 notrace ktime_get_tai_fast_ns(void) in ktime_get_tai_fast_ns()
556 u64 ktime_get_real_fast_ns(void) in ktime_get_real_fast_ns()
560 u64 baser, delta; in ktime_get_real_fast_ns()
665 u64 seconds; in tk_update_ktime_data()
675 seconds = (u64)(tk->xtime_sec + tk->wall_to_monotonic.tv_sec); in tk_update_ktime_data()
761 u64 cycle_now, delta; in timekeeping_forward_now()
770 u64 max = tk->tkr_mono.clock->max_cycles; in timekeeping_forward_now()
771 u64 incr = delta < max ? delta : max; in timekeeping_forward_now()
791 u64 nsecs; in ktime_get_real_ts64()
813 u64 nsecs; in ktime_get()
856 u64 nsecs; in ktime_get_with_offset()
877 u64 nsecs; in ktime_get_coarse_with_offset()
928 u64 nsecs; in ktime_get_raw()
954 u64 nsec; in ktime_get_ts64()
1047 u64 nsec_raw; in ktime_get_snapshot()
1048 u64 nsec_real; in ktime_get_snapshot()
1049 u64 now; in ktime_get_snapshot()
1076 static int scale64_check_overflow(u64 mult, u64 div, u64 *base) in scale64_check_overflow()
1078 u64 tmp, rem; in scale64_check_overflow()
1082 if (((int)sizeof(u64)*8 - fls64(mult) < fls64(tmp)) || in scale64_check_overflow()
1083 ((int)sizeof(u64)*8 - fls64(mult) < fls64(rem))) in scale64_check_overflow()
1110 u64 partial_history_cycles, in adjust_historical_crosststamp()
1111 u64 total_history_cycles, in adjust_historical_crosststamp()
1116 u64 corr_raw, corr_real; in adjust_historical_crosststamp()
1133 corr_raw = (u64)ktime_to_ns( in adjust_historical_crosststamp()
1151 corr_real = (u64)ktime_to_ns( in adjust_historical_crosststamp()
1176 static bool timestamp_in_interval(u64 start, u64 end, u64 ts) in timestamp_in_interval()
1185 static bool convert_clock(u64 *val, u32 numerator, u32 denominator) in convert_clock()
1187 u64 rem, res; in convert_clock()
1225 static bool convert_cs_to_base(u64 *cycles, enum clocksource_ids base_id) in convert_cs_to_base()
1244 static bool convert_ns_to_cs(u64 *delta) in convert_ns_to_cs()
1265 bool ktime_real_to_base_clock(ktime_t treal, enum clocksource_ids base_id, u64 *cycles) in ktime_real_to_base_clock()
1269 u64 delta; in ktime_real_to_base_clock()
1273 if ((u64)treal < tk->tkr_mono.base_real) in ktime_real_to_base_clock()
1275 delta = (u64)treal - tk->tkr_mono.base_real; in ktime_real_to_base_clock()
1308 u64 cycles, now, interval_start; in get_device_system_crosststamp()
1311 u64 nsec_real, nsec_raw; in get_device_system_crosststamp()
1368 u64 partial_history_cycles, total_history_cycles; in get_device_system_crosststamp()
1643 u64 nsecs; in ktime_get_raw_ts64()
1713 u64 timekeeping_max_deferment(void) in timekeeping_max_deferment()
1717 u64 ret; in timekeeping_max_deferment()
1937 u64 cycle_now, nsec; in timekeeping_resume()
1998 u64 cycle_now; in timekeeping_suspend()
2154 u64 ntp_tl = ntp_tick_length(tk->id); in timekeeping_adjust()
2200 tk->tkr_mono.xtime_nsec += (u64)NSEC_PER_SEC << in timekeeping_adjust()
2216 u64 nsecps = (u64)NSEC_PER_SEC << tk->tkr_mono.shift; in accumulate_nsecs_to_secs()
2263 static u64 logarithmic_accumulation(struct timekeeper *tk, u64 offset, in logarithmic_accumulation()
2266 u64 interval = tk->cycle_interval << shift; in logarithmic_accumulation()
2267 u64 snsec_per_sec; in logarithmic_accumulation()
2283 snsec_per_sec = (u64)NSEC_PER_SEC << tk->tkr_raw.shift; in logarithmic_accumulation()
2307 u64 offset, orig_offset; in __timekeeping_advance()
2425 u64 floor = atomic64_read(&mg_floor); in ktime_get_coarse_real_ts64_mg()
2466 u64 nsecs; in ktime_get_real_ts64_mg()
2547 u64 nsecs; in ktime_get_update_offsets_now()
2848 u64 nsecs; in ktime_get_aux()