Lines Matching refs:power

82 	if (dev->power.disable_depth > 0)  in update_pm_runtime_accounting()
85 last = dev->power.accounting_timestamp; in update_pm_runtime_accounting()
88 dev->power.accounting_timestamp = now; in update_pm_runtime_accounting()
100 if (dev->power.runtime_status == RPM_SUSPENDED) in update_pm_runtime_accounting()
101 dev->power.suspended_time += delta; in update_pm_runtime_accounting()
103 dev->power.active_time += delta; in update_pm_runtime_accounting()
110 dev->power.runtime_status = status; in __update_runtime_status()
118 spin_lock_irqsave(&dev->power.lock, flags); in rpm_get_accounted_time()
121 time = suspended ? dev->power.suspended_time : dev->power.active_time; in rpm_get_accounted_time()
123 spin_unlock_irqrestore(&dev->power.lock, flags); in rpm_get_accounted_time()
145 if (dev->power.timer_expires > 0) { in pm_runtime_deactivate_timer()
146 hrtimer_try_to_cancel(&dev->power.suspend_timer); in pm_runtime_deactivate_timer()
147 dev->power.timer_expires = 0; in pm_runtime_deactivate_timer()
162 dev->power.request = RPM_REQ_NONE; in pm_runtime_cancel_pending()
182 if (!dev->power.use_autosuspend) in pm_runtime_autosuspend_expiration()
185 autosuspend_delay = READ_ONCE(dev->power.autosuspend_delay); in pm_runtime_autosuspend_expiration()
189 expires = READ_ONCE(dev->power.last_busy); in pm_runtime_autosuspend_expiration()
200 return dev->power.memalloc_noio; in dev_memalloc_noio()
240 spin_lock_irq(&dev->power.lock); in pm_runtime_set_memalloc_noio()
241 enabled = dev->power.memalloc_noio; in pm_runtime_set_memalloc_noio()
242 dev->power.memalloc_noio = enable; in pm_runtime_set_memalloc_noio()
243 spin_unlock_irq(&dev->power.lock); in pm_runtime_set_memalloc_noio()
275 if (dev->power.runtime_error) in rpm_check_suspend_allowed()
277 else if (dev->power.disable_depth > 0) in rpm_check_suspend_allowed()
279 else if (atomic_read(&dev->power.usage_count)) in rpm_check_suspend_allowed()
281 else if (!dev->power.ignore_children && atomic_read(&dev->power.child_count)) in rpm_check_suspend_allowed()
285 else if ((dev->power.deferred_resume && in rpm_check_suspend_allowed()
286 dev->power.runtime_status == RPM_SUSPENDING) || in rpm_check_suspend_allowed()
287 (dev->power.request_pending && dev->power.request == RPM_REQ_RESUME)) in rpm_check_suspend_allowed()
291 else if (dev->power.runtime_status == RPM_SUSPENDED) in rpm_check_suspend_allowed()
336 atomic_read(&supplier->power.usage_count) > 0) in pm_runtime_release_supplier()
375 __releases(&dev->power.lock) __acquires(&dev->power.lock) in __rpm_callback()
378 bool use_links = dev->power.links_count > 0; in __rpm_callback()
380 if (dev->power.irq_safe) { in __rpm_callback()
381 spin_unlock(&dev->power.lock); in __rpm_callback()
383 spin_unlock_irq(&dev->power.lock); in __rpm_callback()
392 if (use_links && dev->power.runtime_status == RPM_RESUMING) { in __rpm_callback()
408 if (dev->power.irq_safe) { in __rpm_callback()
409 spin_lock(&dev->power.lock); in __rpm_callback()
419 ((dev->power.runtime_status == RPM_SUSPENDING && !retval) || in __rpm_callback()
420 (dev->power.runtime_status == RPM_RESUMING && retval))) { in __rpm_callback()
429 spin_lock_irq(&dev->power.lock); in __rpm_callback()
444 if (dev->power.memalloc_noio) { in rpm_callback()
473 dev->power.runtime_error = retval; in rpm_callback()
502 else if (dev->power.runtime_status != RPM_ACTIVE) in rpm_idle()
509 else if (dev->power.request_pending && in rpm_idle()
510 dev->power.request > RPM_REQ_IDLE) in rpm_idle()
514 else if (dev->power.idle_notification) in rpm_idle()
521 dev->power.request = RPM_REQ_NONE; in rpm_idle()
526 if (!callback || dev->power.no_callbacks) in rpm_idle()
531 dev->power.request = RPM_REQ_IDLE; in rpm_idle()
532 if (!dev->power.request_pending) { in rpm_idle()
533 dev->power.request_pending = true; in rpm_idle()
534 queue_work(pm_wq, &dev->power.work); in rpm_idle()
540 dev->power.idle_notification = true; in rpm_idle()
542 if (dev->power.irq_safe) in rpm_idle()
543 spin_unlock(&dev->power.lock); in rpm_idle()
545 spin_unlock_irq(&dev->power.lock); in rpm_idle()
549 if (dev->power.irq_safe) in rpm_idle()
550 spin_lock(&dev->power.lock); in rpm_idle()
552 spin_lock_irq(&dev->power.lock); in rpm_idle()
554 dev->power.idle_notification = false; in rpm_idle()
555 wake_up_all(&dev->power.wait_queue); in rpm_idle()
584 __releases(&dev->power.lock) __acquires(&dev->power.lock) in rpm_suspend()
598 if (dev->power.runtime_status == RPM_RESUMING && !(rpmflags & RPM_ASYNC)) in rpm_suspend()
605 if ((rpmflags & RPM_AUTO) && dev->power.runtime_status != RPM_SUSPENDING) { in rpm_suspend()
610 dev->power.request = RPM_REQ_NONE; in rpm_suspend()
619 if (!(dev->power.timer_expires && in rpm_suspend()
620 dev->power.timer_expires <= expires)) { in rpm_suspend()
625 u64 slack = (u64)READ_ONCE(dev->power.autosuspend_delay) * in rpm_suspend()
628 dev->power.timer_expires = expires; in rpm_suspend()
629 hrtimer_start_range_ns(&dev->power.suspend_timer, in rpm_suspend()
634 dev->power.timer_autosuspends = 1; in rpm_suspend()
642 if (dev->power.runtime_status == RPM_SUSPENDING) { in rpm_suspend()
650 if (dev->power.irq_safe) { in rpm_suspend()
651 spin_unlock(&dev->power.lock); in rpm_suspend()
655 spin_lock(&dev->power.lock); in rpm_suspend()
661 prepare_to_wait(&dev->power.wait_queue, &wait, in rpm_suspend()
663 if (dev->power.runtime_status != RPM_SUSPENDING) in rpm_suspend()
666 spin_unlock_irq(&dev->power.lock); in rpm_suspend()
670 spin_lock_irq(&dev->power.lock); in rpm_suspend()
672 finish_wait(&dev->power.wait_queue, &wait); in rpm_suspend()
676 if (dev->power.no_callbacks) in rpm_suspend()
681 dev->power.request = (rpmflags & RPM_AUTO) ? in rpm_suspend()
683 if (!dev->power.request_pending) { in rpm_suspend()
684 dev->power.request_pending = true; in rpm_suspend()
685 queue_work(pm_wq, &dev->power.work); in rpm_suspend()
707 atomic_add_unless(&parent->power.child_count, -1, 0); in rpm_suspend()
709 wake_up_all(&dev->power.wait_queue); in rpm_suspend()
711 if (dev->power.deferred_resume) { in rpm_suspend()
712 dev->power.deferred_resume = false; in rpm_suspend()
718 if (dev->power.irq_safe) in rpm_suspend()
722 if (parent && !parent->power.ignore_children) { in rpm_suspend()
723 spin_unlock(&dev->power.lock); in rpm_suspend()
725 spin_lock(&parent->power.lock); in rpm_suspend()
727 spin_unlock(&parent->power.lock); in rpm_suspend()
729 spin_lock(&dev->power.lock); in rpm_suspend()
732 if (dev->power.links_count > 0) { in rpm_suspend()
733 spin_unlock_irq(&dev->power.lock); in rpm_suspend()
737 spin_lock_irq(&dev->power.lock); in rpm_suspend()
748 dev->power.deferred_resume = false; in rpm_suspend()
749 wake_up_all(&dev->power.wait_queue); in rpm_suspend()
757 if (!dev->power.runtime_error && (rpmflags & RPM_AUTO) && in rpm_suspend()
784 __releases(&dev->power.lock) __acquires(&dev->power.lock) in rpm_resume()
793 if (dev->power.runtime_error) { in rpm_resume()
795 } else if (dev->power.disable_depth > 0) { in rpm_resume()
796 if (dev->power.runtime_status == RPM_ACTIVE && in rpm_resume()
797 dev->power.last_status == RPM_ACTIVE) in rpm_resume()
811 dev->power.request = RPM_REQ_NONE; in rpm_resume()
812 if (!dev->power.timer_autosuspends) in rpm_resume()
815 if (dev->power.runtime_status == RPM_ACTIVE) { in rpm_resume()
820 if (dev->power.runtime_status == RPM_RESUMING || in rpm_resume()
821 dev->power.runtime_status == RPM_SUSPENDING) { in rpm_resume()
825 if (dev->power.runtime_status == RPM_SUSPENDING) { in rpm_resume()
826 dev->power.deferred_resume = true; in rpm_resume()
835 if (dev->power.irq_safe) { in rpm_resume()
836 spin_unlock(&dev->power.lock); in rpm_resume()
840 spin_lock(&dev->power.lock); in rpm_resume()
846 prepare_to_wait(&dev->power.wait_queue, &wait, in rpm_resume()
848 if (dev->power.runtime_status != RPM_RESUMING && in rpm_resume()
849 dev->power.runtime_status != RPM_SUSPENDING) in rpm_resume()
852 spin_unlock_irq(&dev->power.lock); in rpm_resume()
856 spin_lock_irq(&dev->power.lock); in rpm_resume()
858 finish_wait(&dev->power.wait_queue, &wait); in rpm_resume()
867 if (dev->power.no_callbacks && !parent && dev->parent) { in rpm_resume()
868 spin_lock_nested(&dev->parent->power.lock, SINGLE_DEPTH_NESTING); in rpm_resume()
869 if (dev->parent->power.disable_depth > 0 || in rpm_resume()
870 dev->parent->power.ignore_children || in rpm_resume()
871 dev->parent->power.runtime_status == RPM_ACTIVE) { in rpm_resume()
872 atomic_inc(&dev->parent->power.child_count); in rpm_resume()
873 spin_unlock(&dev->parent->power.lock); in rpm_resume()
877 spin_unlock(&dev->parent->power.lock); in rpm_resume()
882 dev->power.request = RPM_REQ_RESUME; in rpm_resume()
883 if (!dev->power.request_pending) { in rpm_resume()
884 dev->power.request_pending = true; in rpm_resume()
885 queue_work(pm_wq, &dev->power.work); in rpm_resume()
898 if (dev->power.irq_safe) in rpm_resume()
901 spin_unlock(&dev->power.lock); in rpm_resume()
905 spin_lock(&parent->power.lock); in rpm_resume()
910 if (!parent->power.disable_depth && in rpm_resume()
911 !parent->power.ignore_children) { in rpm_resume()
913 if (parent->power.runtime_status != RPM_ACTIVE) in rpm_resume()
916 spin_unlock(&parent->power.lock); in rpm_resume()
918 spin_lock(&dev->power.lock); in rpm_resume()
926 if (dev->power.no_callbacks) in rpm_resume()
944 atomic_inc(&parent->power.child_count); in rpm_resume()
946 wake_up_all(&dev->power.wait_queue); in rpm_resume()
952 if (parent && !dev->power.irq_safe) { in rpm_resume()
953 spin_unlock_irq(&dev->power.lock); in rpm_resume()
957 spin_lock_irq(&dev->power.lock); in rpm_resume()
974 struct device *dev = container_of(work, struct device, power.work); in pm_runtime_work()
977 spin_lock_irq(&dev->power.lock); in pm_runtime_work()
979 if (!dev->power.request_pending) in pm_runtime_work()
982 req = dev->power.request; in pm_runtime_work()
983 dev->power.request = RPM_REQ_NONE; in pm_runtime_work()
984 dev->power.request_pending = false; in pm_runtime_work()
1004 spin_unlock_irq(&dev->power.lock); in pm_runtime_work()
1015 struct device *dev = container_of(timer, struct device, power.suspend_timer); in pm_suspend_timer_fn()
1019 spin_lock_irqsave(&dev->power.lock, flags); in pm_suspend_timer_fn()
1021 expires = dev->power.timer_expires; in pm_suspend_timer_fn()
1027 dev->power.timer_expires = 0; in pm_suspend_timer_fn()
1028 rpm_suspend(dev, dev->power.timer_autosuspends ? in pm_suspend_timer_fn()
1032 spin_unlock_irqrestore(&dev->power.lock, flags); in pm_suspend_timer_fn()
1048 spin_lock_irqsave(&dev->power.lock, flags); in pm_schedule_suspend()
1063 dev->power.timer_expires = expires; in pm_schedule_suspend()
1064 dev->power.timer_autosuspends = 0; in pm_schedule_suspend()
1065 hrtimer_start(&dev->power.suspend_timer, expires, HRTIMER_MODE_ABS); in pm_schedule_suspend()
1068 spin_unlock_irqrestore(&dev->power.lock, flags); in pm_schedule_suspend()
1078 ret = atomic_sub_return(1, &dev->power.usage_count); in rpm_drop_usage_count()
1088 atomic_inc(&dev->power.usage_count); in rpm_drop_usage_count()
1121 might_sleep_if(!(rpmflags & RPM_ASYNC) && !dev->power.irq_safe); in __pm_runtime_idle()
1123 spin_lock_irqsave(&dev->power.lock, flags); in __pm_runtime_idle()
1125 spin_unlock_irqrestore(&dev->power.lock, flags); in __pm_runtime_idle()
1159 might_sleep_if(!(rpmflags & RPM_ASYNC) && !dev->power.irq_safe); in __pm_runtime_suspend()
1161 spin_lock_irqsave(&dev->power.lock, flags); in __pm_runtime_suspend()
1163 spin_unlock_irqrestore(&dev->power.lock, flags); in __pm_runtime_suspend()
1185 might_sleep_if(!(rpmflags & RPM_ASYNC) && !dev->power.irq_safe && in __pm_runtime_resume()
1186 dev->power.runtime_status != RPM_ACTIVE); in __pm_runtime_resume()
1189 atomic_inc(&dev->power.usage_count); in __pm_runtime_resume()
1191 spin_lock_irqsave(&dev->power.lock, flags); in __pm_runtime_resume()
1193 spin_unlock_irqrestore(&dev->power.lock, flags); in __pm_runtime_resume()
1228 spin_lock_irqsave(&dev->power.lock, flags); in pm_runtime_get_conditional()
1229 if (dev->power.disable_depth > 0) { in pm_runtime_get_conditional()
1231 } else if (dev->power.runtime_status != RPM_ACTIVE) { in pm_runtime_get_conditional()
1233 } else if (ign_usage_count || (!dev->power.ignore_children && in pm_runtime_get_conditional()
1234 atomic_read(&dev->power.child_count) > 0)) { in pm_runtime_get_conditional()
1236 atomic_inc(&dev->power.usage_count); in pm_runtime_get_conditional()
1238 retval = atomic_inc_not_zero(&dev->power.usage_count); in pm_runtime_get_conditional()
1241 spin_unlock_irqrestore(&dev->power.lock, flags); in pm_runtime_get_conditional()
1318 spin_lock_irqsave(&dev->power.lock, flags); in __pm_runtime_set_status()
1324 if (dev->power.runtime_error || dev->power.disable_depth) in __pm_runtime_set_status()
1325 dev->power.disable_depth++; in __pm_runtime_set_status()
1329 spin_unlock_irqrestore(&dev->power.lock, flags); in __pm_runtime_set_status()
1350 spin_lock_irqsave(&dev->power.lock, flags); in __pm_runtime_set_status()
1352 if (dev->power.runtime_status == status || !parent) in __pm_runtime_set_status()
1356 atomic_add_unless(&parent->power.child_count, -1, 0); in __pm_runtime_set_status()
1357 notify_parent = !parent->power.ignore_children; in __pm_runtime_set_status()
1359 spin_lock_nested(&parent->power.lock, SINGLE_DEPTH_NESTING); in __pm_runtime_set_status()
1366 if (!parent->power.disable_depth && in __pm_runtime_set_status()
1367 !parent->power.ignore_children && in __pm_runtime_set_status()
1368 parent->power.runtime_status != RPM_ACTIVE) { in __pm_runtime_set_status()
1373 } else if (dev->power.runtime_status == RPM_SUSPENDED) { in __pm_runtime_set_status()
1374 atomic_inc(&parent->power.child_count); in __pm_runtime_set_status()
1377 spin_unlock(&parent->power.lock); in __pm_runtime_set_status()
1388 dev->power.runtime_error = 0; in __pm_runtime_set_status()
1391 spin_unlock_irqrestore(&dev->power.lock, flags); in __pm_runtime_set_status()
1423 if (dev->power.request_pending) { in __pm_runtime_barrier()
1424 dev->power.request = RPM_REQ_NONE; in __pm_runtime_barrier()
1425 spin_unlock_irq(&dev->power.lock); in __pm_runtime_barrier()
1427 cancel_work_sync(&dev->power.work); in __pm_runtime_barrier()
1429 spin_lock_irq(&dev->power.lock); in __pm_runtime_barrier()
1430 dev->power.request_pending = false; in __pm_runtime_barrier()
1433 if (dev->power.runtime_status == RPM_SUSPENDING || in __pm_runtime_barrier()
1434 dev->power.runtime_status == RPM_RESUMING || in __pm_runtime_barrier()
1435 dev->power.idle_notification) { in __pm_runtime_barrier()
1440 prepare_to_wait(&dev->power.wait_queue, &wait, in __pm_runtime_barrier()
1442 if (dev->power.runtime_status != RPM_SUSPENDING in __pm_runtime_barrier()
1443 && dev->power.runtime_status != RPM_RESUMING in __pm_runtime_barrier()
1444 && !dev->power.idle_notification) in __pm_runtime_barrier()
1446 spin_unlock_irq(&dev->power.lock); in __pm_runtime_barrier()
1450 spin_lock_irq(&dev->power.lock); in __pm_runtime_barrier()
1452 finish_wait(&dev->power.wait_queue, &wait); in __pm_runtime_barrier()
1475 spin_lock_irq(&dev->power.lock); in pm_runtime_barrier()
1477 if (dev->power.request_pending in pm_runtime_barrier()
1478 && dev->power.request == RPM_REQ_RESUME) { in pm_runtime_barrier()
1485 spin_unlock_irq(&dev->power.lock); in pm_runtime_barrier()
1496 spin_lock_irq(&dev->power.lock); in pm_runtime_block_if_disabled()
1499 if (ret && dev->power.last_status == RPM_INVALID) in pm_runtime_block_if_disabled()
1500 dev->power.last_status = RPM_BLOCKED; in pm_runtime_block_if_disabled()
1502 spin_unlock_irq(&dev->power.lock); in pm_runtime_block_if_disabled()
1509 spin_lock_irq(&dev->power.lock); in pm_runtime_unblock()
1511 if (dev->power.last_status == RPM_BLOCKED) in pm_runtime_unblock()
1512 dev->power.last_status = RPM_INVALID; in pm_runtime_unblock()
1514 spin_unlock_irq(&dev->power.lock); in pm_runtime_unblock()
1519 spin_lock_irq(&dev->power.lock); in __pm_runtime_disable()
1521 if (dev->power.disable_depth > 0) { in __pm_runtime_disable()
1522 dev->power.disable_depth++; in __pm_runtime_disable()
1531 if (check_resume && dev->power.request_pending && in __pm_runtime_disable()
1532 dev->power.request == RPM_REQ_RESUME) { in __pm_runtime_disable()
1547 if (!dev->power.disable_depth++) { in __pm_runtime_disable()
1549 dev->power.last_status = dev->power.runtime_status; in __pm_runtime_disable()
1553 spin_unlock_irq(&dev->power.lock); in __pm_runtime_disable()
1565 spin_lock_irqsave(&dev->power.lock, flags); in pm_runtime_enable()
1567 if (!dev->power.disable_depth) { in pm_runtime_enable()
1572 if (--dev->power.disable_depth > 0) in pm_runtime_enable()
1575 if (dev->power.last_status == RPM_BLOCKED) { in pm_runtime_enable()
1579 dev->power.last_status = RPM_INVALID; in pm_runtime_enable()
1580 dev->power.accounting_timestamp = ktime_get_mono_fast_ns(); in pm_runtime_enable()
1582 if (dev->power.runtime_status == RPM_SUSPENDED && in pm_runtime_enable()
1583 !dev->power.ignore_children && in pm_runtime_enable()
1584 atomic_read(&dev->power.child_count) > 0) in pm_runtime_enable()
1588 spin_unlock_irqrestore(&dev->power.lock, flags); in pm_runtime_enable()
1668 spin_lock_irq(&dev->power.lock); in pm_runtime_forbid()
1669 if (!dev->power.runtime_auto) in pm_runtime_forbid()
1672 dev->power.runtime_auto = false; in pm_runtime_forbid()
1673 atomic_inc(&dev->power.usage_count); in pm_runtime_forbid()
1677 spin_unlock_irq(&dev->power.lock); in pm_runtime_forbid()
1691 spin_lock_irq(&dev->power.lock); in pm_runtime_allow()
1692 if (dev->power.runtime_auto) in pm_runtime_allow()
1695 dev->power.runtime_auto = true; in pm_runtime_allow()
1703 spin_unlock_irq(&dev->power.lock); in pm_runtime_allow()
1717 spin_lock_irq(&dev->power.lock); in pm_runtime_no_callbacks()
1718 dev->power.no_callbacks = 1; in pm_runtime_no_callbacks()
1719 spin_unlock_irq(&dev->power.lock); in pm_runtime_no_callbacks()
1741 spin_lock_irq(&dev->power.lock); in pm_runtime_irq_safe()
1742 dev->power.irq_safe = 1; in pm_runtime_irq_safe()
1743 spin_unlock_irq(&dev->power.lock); in pm_runtime_irq_safe()
1760 int delay = dev->power.autosuspend_delay; in update_autosuspend()
1763 if (dev->power.use_autosuspend && delay < 0) { in update_autosuspend()
1767 atomic_inc(&dev->power.usage_count); in update_autosuspend()
1779 atomic_dec(&dev->power.usage_count); in update_autosuspend()
1799 spin_lock_irq(&dev->power.lock); in pm_runtime_set_autosuspend_delay()
1800 old_delay = dev->power.autosuspend_delay; in pm_runtime_set_autosuspend_delay()
1801 old_use = dev->power.use_autosuspend; in pm_runtime_set_autosuspend_delay()
1802 dev->power.autosuspend_delay = delay; in pm_runtime_set_autosuspend_delay()
1804 spin_unlock_irq(&dev->power.lock); in pm_runtime_set_autosuspend_delay()
1820 spin_lock_irq(&dev->power.lock); in __pm_runtime_use_autosuspend()
1821 old_delay = dev->power.autosuspend_delay; in __pm_runtime_use_autosuspend()
1822 old_use = dev->power.use_autosuspend; in __pm_runtime_use_autosuspend()
1823 dev->power.use_autosuspend = use; in __pm_runtime_use_autosuspend()
1825 spin_unlock_irq(&dev->power.lock); in __pm_runtime_use_autosuspend()
1835 dev->power.runtime_status = RPM_SUSPENDED; in pm_runtime_init()
1836 dev->power.last_status = RPM_INVALID; in pm_runtime_init()
1837 dev->power.idle_notification = false; in pm_runtime_init()
1839 dev->power.disable_depth = 1; in pm_runtime_init()
1840 atomic_set(&dev->power.usage_count, 0); in pm_runtime_init()
1842 dev->power.runtime_error = 0; in pm_runtime_init()
1844 atomic_set(&dev->power.child_count, 0); in pm_runtime_init()
1846 dev->power.runtime_auto = true; in pm_runtime_init()
1848 dev->power.request_pending = false; in pm_runtime_init()
1849 dev->power.request = RPM_REQ_NONE; in pm_runtime_init()
1850 dev->power.deferred_resume = false; in pm_runtime_init()
1851 dev->power.needs_force_resume = false; in pm_runtime_init()
1852 INIT_WORK(&dev->power.work, pm_runtime_work); in pm_runtime_init()
1854 dev->power.timer_expires = 0; in pm_runtime_init()
1855 hrtimer_setup(&dev->power.suspend_timer, pm_suspend_timer_fn, CLOCK_MONOTONIC, in pm_runtime_init()
1858 init_waitqueue_head(&dev->power.wait_queue); in pm_runtime_init()
1868 if (dev->power.runtime_status == RPM_ACTIVE) in pm_runtime_reinit()
1870 if (dev->power.irq_safe) { in pm_runtime_reinit()
1871 spin_lock_irq(&dev->power.lock); in pm_runtime_reinit()
1872 dev->power.irq_safe = 0; in pm_runtime_reinit()
1873 spin_unlock_irq(&dev->power.lock); in pm_runtime_reinit()
1882 dev->power.needs_force_resume = false; in pm_runtime_reinit()
1939 spin_lock_irq(&dev->power.lock); in pm_runtime_new_link()
1940 dev->power.links_count++; in pm_runtime_new_link()
1941 spin_unlock_irq(&dev->power.lock); in pm_runtime_new_link()
1946 spin_lock_irq(&dev->power.lock); in pm_runtime_drop_link_count()
1947 WARN_ON(dev->power.links_count == 0); in pm_runtime_drop_link_count()
1948 dev->power.links_count--; in pm_runtime_drop_link_count()
1949 spin_unlock_irq(&dev->power.lock); in pm_runtime_drop_link_count()
2010 if (pm_runtime_status_suspended(dev) || dev->power.needs_force_resume) in pm_runtime_force_suspend()
2034 dev->power.needs_force_resume = true; in pm_runtime_force_suspend()
2071 if (!dev->power.needs_force_resume && (!dev_pm_smart_suspend(dev) || in pm_runtime_force_resume()
2093 dev->power.smart_suspend = false; in pm_runtime_force_resume()
2098 dev->power.needs_force_resume = false; in pm_runtime_force_resume()
2107 return atomic_read(&dev->power.usage_count) <= 1 && in pm_runtime_need_not_resume()
2108 (atomic_read(&dev->power.child_count) == 0 || in pm_runtime_need_not_resume()
2109 dev->power.ignore_children); in pm_runtime_need_not_resume()