| /linux/drivers/platform/surface/aggregator/ |
| A D | ssh_packet_layer.c | 741 struct ssh_ptl *ptl = packet->ptl; in __ssh_ptl_queue_push() local 776 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_queue_remove() local 793 struct ssh_ptl *ptl = p->ptl; in ssh_ptl_pending_push() local 833 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_pending_remove() local 853 struct ssh_ptl *ptl = READ_ONCE(p->ptl); in __ssh_ptl_complete() local 885 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_tx_can_process() local 975 struct ssh_ptl *ptl = packet->ptl; in ssh_ptl_tx_compl_success() local 1336 WRITE_ONCE(p->ptl, ptl); in ssh_ptl_submit() 1633 ptl->rx.blocked.offset = (ptl->rx.blocked.offset + 1) in ssh_ptl_rx_retransmit_check() 1646 ptl->ops.data_received(ptl, payload); in ssh_ptl_rx_dataframe() [all …]
|
| A D | ssh_packet_layer.h | 142 void ssh_ptl_destroy(struct ssh_ptl *ptl); 153 return ptl->serdev ? &ptl->serdev->dev : NULL; in ssh_ptl_get_device() 156 int ssh_ptl_tx_start(struct ssh_ptl *ptl); 157 int ssh_ptl_tx_stop(struct ssh_ptl *ptl); 158 int ssh_ptl_rx_start(struct ssh_ptl *ptl); 159 int ssh_ptl_rx_stop(struct ssh_ptl *ptl); 160 void ssh_ptl_shutdown(struct ssh_ptl *ptl); 162 int ssh_ptl_submit(struct ssh_ptl *ptl, struct ssh_packet *p); 176 static inline void ssh_ptl_tx_wakeup_transfer(struct ssh_ptl *ptl) in ssh_ptl_tx_wakeup_transfer() argument 178 if (test_bit(SSH_PTL_SF_SHUTDOWN_BIT, &ptl->state)) in ssh_ptl_tx_wakeup_transfer() [all …]
|
| A D | ssh_request_layer.h | 66 struct ssh_ptl ptl; member 94 #define rtl_dbg(r, fmt, ...) ptl_dbg(&(r)->ptl, fmt, ##__VA_ARGS__) 95 #define rtl_info(p, fmt, ...) ptl_info(&(p)->ptl, fmt, ##__VA_ARGS__) 96 #define rtl_warn(r, fmt, ...) ptl_warn(&(r)->ptl, fmt, ##__VA_ARGS__) 97 #define rtl_err(r, fmt, ...) ptl_err(&(r)->ptl, fmt, ##__VA_ARGS__) 112 return ssh_ptl_get_device(&rtl->ptl); in ssh_rtl_get_device() 123 struct ssh_ptl *ptl; in ssh_request_rtl() local 125 ptl = READ_ONCE(rqst->packet.ptl); in ssh_request_rtl() 126 return likely(ptl) ? to_ssh_rtl(ptl, ptl) : NULL; in ssh_request_rtl()
|
| A D | ssh_request_layer.c | 258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet); in ssh_rtl_tx_try_process_one() 391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) { in ssh_rtl_submit() 619 if (flags == fixed && !READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_nonpending() 674 if (!READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_pending() 911 struct ssh_rtl *rtl = to_ssh_rtl(p, ptl); in ssh_rtl_rx_command() 1033 status = ssh_ptl_init(&rtl->ptl, serdev, &ptl_ops); in ssh_rtl_init() 1067 ssh_ptl_destroy(&rtl->ptl); in ssh_rtl_destroy() 1080 status = ssh_ptl_tx_start(&rtl->ptl); in ssh_rtl_start() 1086 status = ssh_ptl_rx_start(&rtl->ptl); in ssh_rtl_start() 1089 ssh_ptl_tx_stop(&rtl->ptl); in ssh_rtl_start() [all …]
|
| /linux/drivers/media/platform/allegro-dvt/ |
| A D | nal-hevc.c | 115 if (ptl->general_profile_idc == 4 || in nal_hevc_rbsp_profile_tier_level() 117 ptl->general_profile_idc == 5 || in nal_hevc_rbsp_profile_tier_level() 119 ptl->general_profile_idc == 6 || in nal_hevc_rbsp_profile_tier_level() 121 ptl->general_profile_idc == 7 || in nal_hevc_rbsp_profile_tier_level() 123 ptl->general_profile_idc == 8 || in nal_hevc_rbsp_profile_tier_level() 125 ptl->general_profile_idc == 9 || in nal_hevc_rbsp_profile_tier_level() 127 ptl->general_profile_idc == 10 || in nal_hevc_rbsp_profile_tier_level() 138 if (ptl->general_profile_idc == 5 || in nal_hevc_rbsp_profile_tier_level() 140 ptl->general_profile_idc == 9 || in nal_hevc_rbsp_profile_tier_level() 161 if ((ptl->general_profile_idc >= 1 && ptl->general_profile_idc <= 5) || in nal_hevc_rbsp_profile_tier_level() [all …]
|
| /linux/mm/ |
| A D | page_vma_mapped.c | 24 *ptlp = pvmw->ptl; in map_pte() 70 pvmw->ptl = *ptlp; in map_pte() 71 spin_lock(pvmw->ptl); in map_pte() 178 spinlock_t *ptl; in page_vma_mapped_walk() local 263 pvmw->ptl = NULL; in page_vma_mapped_walk() 276 spin_unlock(ptl); in page_vma_mapped_walk() 296 if (pvmw->ptl) { in page_vma_mapped_walk() 298 pvmw->ptl = NULL; in page_vma_mapped_walk() 307 if (!pvmw->ptl) { in page_vma_mapped_walk() 308 pvmw->ptl = ptl; in page_vma_mapped_walk() [all …]
|
| A D | mincore.c | 103 spinlock_t *ptl; in mincore_pte_range() local 109 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 110 if (ptl) { in mincore_pte_range() 112 spin_unlock(ptl); in mincore_pte_range() 116 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range() 151 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
| A D | huge_memory.c | 1343 spinlock_t *ptl; in insert_pfn_pmd() local 1444 spinlock_t *ptl; in insert_pfn_pud() local 1967 spinlock_t *ptl; in madvise_free_huge_pmd() local 1976 if (!ptl) in madvise_free_huge_pmd() 2047 spinlock_t *ptl; in zap_huge_pmd() local 2052 if (!ptl) in zap_huge_pmd() 2209 if (!ptl) in change_huge_pmd() 2352 if (!ptl) in change_huge_pud() 2521 return ptl; in __pmd_trans_huge_lock() 2538 return ptl; in __pud_trans_huge_lock() [all …]
|
| A D | pagewalk.c | 48 spinlock_t *ptl; in walk_pte_range() local 70 pte_unmap_unlock(pte, ptl); in walk_pte_range() 725 spinlock_t *ptl; in folio_walk_start() local 749 ptl = pud_lock(vma->vm_mm, pudp); in folio_walk_start() 762 spin_unlock(ptl); in folio_walk_start() 765 spin_unlock(ptl); in folio_walk_start() 793 spin_unlock(ptl); in folio_walk_start() 796 spin_unlock(ptl); in folio_walk_start() 816 spin_unlock(ptl); in folio_walk_start() 852 pte_unmap_unlock(ptep, ptl); in folio_walk_start() [all …]
|
| A D | khugepaged.c | 725 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 741 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 1002 spinlock_t *ptl; in __collapse_huge_page_swapin() local 1027 vmf.ptl = ptl; in __collapse_huge_page_swapin() 1267 spinlock_t *ptl; in hpage_collapse_scan_pmd() local 1608 spin_lock(ptl); in collapse_pte_mapped_thp() 1657 if (ptl != pml) in collapse_pte_mapped_thp() 1662 if (ptl != pml) in collapse_pte_mapped_thp() 1707 spinlock_t *ptl; in retract_page_tables() local 1746 if (ptl != pml) in retract_page_tables() [all …]
|
| A D | memory.c | 443 spin_unlock(ptl); in pmd_install() 1592 spinlock_t *ptl; in zap_pte_range() local 2070 spinlock_t *ptl; in insert_page() local 2321 spinlock_t *ptl; in insert_pfn() local 2550 spinlock_t *ptl; in remap_pte_range() local 2796 spinlock_t *ptl; in apply_to_pte_range() local 3908 &vmf->ptl); in remove_device_exclusive_entry() 4113 spinlock_t *ptl; in alloc_swap_folio() local 4407 &vmf->ptl); in do_swap_page() 6919 if (!ptl) in ptlock_alloc() [all …]
|
| A D | madvise.c | 177 spinlock_t *ptl; in swapin_walk_pmd_entry() local 198 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 208 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 346 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local 366 if (!ptl) in madvise_cold_or_pageout_pte_range() 392 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 424 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 447 pte_unmap_unlock(start_pte, ptl); in madvise_cold_or_pageout_pte_range() 547 pte_unmap_unlock(start_pte, ptl); in madvise_cold_or_pageout_pte_range() 650 spinlock_t *ptl; in madvise_free_pte_range() local [all …]
|
| A D | migrate_device.c | 66 spinlock_t *ptl; in migrate_vma_collect_pmd() local 76 ptl = pmd_lock(mm, pmdp); in migrate_vma_collect_pmd() 78 spin_unlock(ptl); in migrate_vma_collect_pmd() 84 spin_unlock(ptl); in migrate_vma_collect_pmd() 90 spin_unlock(ptl); in migrate_vma_collect_pmd() 103 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_collect_pmd() 277 pte_unmap_unlock(ptep - 1, ptl); in migrate_vma_collect_pmd() 576 spinlock_t *ptl; in migrate_vma_insert_page() local 636 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_insert_page() 673 pte_unmap_unlock(ptep, ptl); in migrate_vma_insert_page() [all …]
|
| A D | hmm.c | 418 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local 420 if (!ptl) in hmm_vma_walk_pud() 428 spin_unlock(ptl); in hmm_vma_walk_pud() 446 spin_unlock(ptl); in hmm_vma_walk_pud() 460 spin_unlock(ptl); in hmm_vma_walk_pud() 479 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local 482 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 494 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry() 513 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
|
| A D | pgtable-generic.c | 367 spinlock_t *ptl; in __pte_offset_map_lock() local 374 ptl = pte_lockptr(mm, &pmdval); in __pte_offset_map_lock() 375 spin_lock(ptl); in __pte_offset_map_lock() 377 *ptlp = ptl; in __pte_offset_map_lock() 380 pte_unmap_unlock(pte, ptl); in __pte_offset_map_lock()
|
| /linux/arch/arm/lib/ |
| A D | uaccess_with_memcpy.c | 31 spinlock_t *ptl; in pin_page_for_write() local 61 spin_lock(ptl); in pin_page_for_write() 64 spin_unlock(ptl); in pin_page_for_write() 69 *ptlp = ptl; in pin_page_for_write() 82 pte_unmap_unlock(pte, ptl); in pin_page_for_write() 87 *ptlp = ptl; in pin_page_for_write() 105 spinlock_t *ptl; in __copy_to_user_memcpy() local 129 pte_unmap_unlock(pte, ptl); in __copy_to_user_memcpy() 131 spin_unlock(ptl); in __copy_to_user_memcpy() 169 spinlock_t *ptl; in __clear_user_memset() local [all …]
|
| /linux/arch/arm/mm/ |
| A D | fault-armv.c | 70 static inline void do_pte_lock(spinlock_t *ptl) in do_pte_lock() argument 76 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in do_pte_lock() 79 static inline void do_pte_unlock(spinlock_t *ptl) in do_pte_unlock() argument 81 spin_unlock(ptl); in do_pte_unlock() 84 static inline void do_pte_lock(spinlock_t *ptl) {} in do_pte_lock() argument 85 static inline void do_pte_unlock(spinlock_t *ptl) {} in do_pte_unlock() argument 91 spinlock_t *ptl; in adjust_pte() local 120 pte = pte_offset_map_nolock(vma->vm_mm, pmd, address, &ptl); in adjust_pte() 124 do_pte_lock(ptl); in adjust_pte() 128 do_pte_unlock(ptl); in adjust_pte()
|
| /linux/mm/damon/ |
| A D | vaddr.c | 307 spinlock_t *ptl; in damon_mkold_pmd_entry() local 314 spin_unlock(ptl); in damon_mkold_pmd_entry() 320 spin_unlock(ptl); in damon_mkold_pmd_entry() 323 spin_unlock(ptl); in damon_mkold_pmd_entry() 374 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local 385 spin_unlock(ptl); in damon_mkold_hugetlb_entry() 444 spinlock_t *ptl; in damon_young_pmd_entry() local 456 spin_unlock(ptl); in damon_young_pmd_entry() 474 spin_unlock(ptl); in damon_young_pmd_entry() 510 spinlock_t *ptl; in damon_young_hugetlb_entry() local [all …]
|
| /linux/Documentation/translations/zh_CN/mm/ |
| A D | split_page_table_lock.rst | 63 面。这个区域与page->ptl共享存储。 79 page->ptl 82 page->ptl用于访问分割页表锁,其中'page'是包含该表的页面struct page。它 89 - 如果spinlock_t的大小大于long的大小,我们使用page->ptl作为spinlock_t 96 请不要直接访问page->ptl - -使用适当的辅助函数。
|
| /linux/arch/s390/mm/ |
| A D | pgtable.c | 806 spinlock_t *ptl; in set_guest_storage_key() local 826 spin_unlock(ptl); in set_guest_storage_key() 838 spin_unlock(ptl); in set_guest_storage_key() 841 spin_unlock(ptl); in set_guest_storage_key() 915 spinlock_t *ptl; in reset_guest_reference_bit() local 948 spin_unlock(ptl); in reset_guest_reference_bit() 979 spinlock_t *ptl; in get_guest_storage_key() local 1012 spin_unlock(ptl); in get_guest_storage_key() 1047 spinlock_t *ptl; in pgste_perform_essa() local 1152 spinlock_t *ptl; in set_pgste_bits() local [all …]
|
| A D | gmap.c | 556 spinlock_t *ptl; in __gmap_link() local 612 ptl = pmd_lock(mm, pmd); in __gmap_link() 634 spin_unlock(ptl); in __gmap_link() 690 spinlock_t *ptl; in __gmap_zap() local 869 spinlock_t **ptl) in gmap_pte_op_walk() argument 1028 spinlock_t *ptl; in gmap_protect_pte() local 1042 gmap_pte_op_end(ptep, ptl); in gmap_protect_pte() 1151 spinlock_t *ptl; in gmap_read_table() local 1237 spinlock_t *ptl; in gmap_protect_rmap() local 2142 spinlock_t *ptl; in gmap_shadow_page() local [all …]
|
| /linux/arch/powerpc/lib/ |
| A D | code-patching.c | 156 spinlock_t *ptl; in text_area_cpu_up_mm() local 177 pte = get_locked_pte(mm, addr, &ptl); in text_area_cpu_up_mm() 180 pte_unmap_unlock(pte, ptl); in text_area_cpu_up_mm() 290 spinlock_t *ptl; in __do_patch_mem_mm() local 296 pte = get_locked_pte(patching_mm, text_poke_addr, &ptl); in __do_patch_mem_mm() 322 pte_unmap_unlock(pte, ptl); in __do_patch_mem_mm() 473 spinlock_t *ptl; in __do_patch_instructions_mm() local 482 pte = get_locked_pte(patching_mm, text_poke_addr, &ptl); in __do_patch_instructions_mm() 508 pte_unmap_unlock(pte, ptl); in __do_patch_instructions_mm()
|
| /linux/fs/proc/ |
| A D | task_mmu.c | 886 spinlock_t *ptl; in smaps_pte_range() local 889 if (ptl) { in smaps_pte_range() 1454 spinlock_t *ptl; in clear_refs_pte_range() local 1458 if (ptl) { in clear_refs_pte_range() 1756 spinlock_t *ptl; in pagemap_pmd_range() local 1762 if (ptl) { in pagemap_pmd_range() 2404 spinlock_t *ptl; in pagemap_scan_thp_entry() local 2408 if (!ptl) in pagemap_scan_thp_entry() 2455 spinlock_t *ptl; in pagemap_scan_pmd_entry() local 2557 spinlock_t *ptl; in pagemap_scan_hugetlb_entry() local [all …]
|
| /linux/Documentation/mm/ |
| A D | split_page_table_lock.rst | 66 This field shares storage with page->ptl. 83 page->ptl 86 page->ptl is used to access split page table lock, where 'page' is struct 95 - if size of spinlock_t is bigger then size of long, we use page->ptl as 103 Please, never access page->ptl directly -- use appropriate helper.
|
| /linux/sound/soc/sof/intel/ |
| A D | Makefile | 37 snd-sof-pci-intel-ptl-y := pci-ptl.o 47 obj-$(CONFIG_SND_SOC_SOF_INTEL_PTL) += snd-sof-pci-intel-ptl.o
|