| /linux/drivers/platform/surface/aggregator/ |
| H A D | ssh_packet_layer.c | 335 ptl_info(packet->ptl, "packet error injection: dropping ACK packet %p\n", in __ssh_ptl_should_drop_ack_packet() 347 ptl_info(packet->ptl, "packet error injection: dropping NAK packet %p\n", in __ssh_ptl_should_drop_nak_packet() 359 ptl_info(packet->ptl, in __ssh_ptl_should_drop_dsq_packet() 387 static int ssh_ptl_write_buf(struct ssh_ptl *ptl, struct ssh_packet *packet, in ssh_ptl_write_buf() argument 395 ptl_info(packet->ptl, in ssh_ptl_write_buf() 402 return serdev_device_write_buf(ptl->serdev, buf, count); in ssh_ptl_write_buf() 419 ptl_info(packet->ptl, in ssh_ptl_tx_inject_invalid_data() 431 static void ssh_ptl_rx_inject_invalid_syn(struct ssh_ptl *ptl, in ssh_ptl_rx_inject_invalid_syn() argument 448 static void ssh_ptl_rx_inject_invalid_data(struct ssh_ptl *ptl, in ssh_ptl_rx_inject_invalid_data() argument 486 static inline int ssh_ptl_write_buf(struct ssh_ptl *ptl, in ssh_ptl_write_buf() argument [all …]
|
| H A D | ssh_packet_layer.h | 139 int ssh_ptl_init(struct ssh_ptl *ptl, struct serdev_device *serdev, 142 void ssh_ptl_destroy(struct ssh_ptl *ptl); 151 static inline struct device *ssh_ptl_get_device(struct ssh_ptl *ptl) in ssh_ptl_get_device() argument 153 return ptl->serdev ? &ptl->serdev->dev : NULL; in ssh_ptl_get_device() 156 int ssh_ptl_tx_start(struct ssh_ptl *ptl); 157 int ssh_ptl_tx_stop(struct ssh_ptl *ptl); 158 int ssh_ptl_rx_start(struct ssh_ptl *ptl); 159 int ssh_ptl_rx_stop(struct ssh_ptl *ptl); 160 void ssh_ptl_shutdown(struct ssh_ptl *ptl); 162 int ssh_ptl_submit(struct ssh_ptl *ptl, struct ssh_packet *p); [all …]
|
| H A D | ssh_request_layer.h | 66 struct ssh_ptl ptl; member 94 #define rtl_dbg(r, fmt, ...) ptl_dbg(&(r)->ptl, fmt, ##__VA_ARGS__) 95 #define rtl_info(p, fmt, ...) ptl_info(&(p)->ptl, fmt, ##__VA_ARGS__) 96 #define rtl_warn(r, fmt, ...) ptl_warn(&(r)->ptl, fmt, ##__VA_ARGS__) 97 #define rtl_err(r, fmt, ...) ptl_err(&(r)->ptl, fmt, ##__VA_ARGS__) 112 return ssh_ptl_get_device(&rtl->ptl); in ssh_rtl_get_device() 123 struct ssh_ptl *ptl; in ssh_request_rtl() local 125 ptl = READ_ONCE(rqst->packet.ptl); in ssh_request_rtl() 126 return likely(ptl) ? to_ssh_rtl(ptl, ptl) : NULL; in ssh_request_rtl()
|
| H A D | ssh_request_layer.c | 258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet); in ssh_rtl_tx_try_process_one() 391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) { in ssh_rtl_submit() 619 if (flags == fixed && !READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_nonpending() 674 if (!READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_pending() 911 struct ssh_rtl *rtl = to_ssh_rtl(p, ptl); in ssh_rtl_rx_command() 1033 status = ssh_ptl_init(&rtl->ptl, serdev, &ptl_ops); in ssh_rtl_init() 1067 ssh_ptl_destroy(&rtl->ptl); in ssh_rtl_destroy() 1080 status = ssh_ptl_tx_start(&rtl->ptl); in ssh_rtl_start() 1086 status = ssh_ptl_rx_start(&rtl->ptl); in ssh_rtl_start() 1089 ssh_ptl_tx_stop(&rtl->ptl); in ssh_rtl_start() [all …]
|
| /linux/drivers/media/platform/allegro-dvt/ |
| H A D | nal-hevc.c | 101 struct nal_hevc_profile_tier_level *ptl) in nal_hevc_rbsp_profile_tier_level() argument 106 rbsp_bits(rbsp, 2, &ptl->general_profile_space); in nal_hevc_rbsp_profile_tier_level() 107 rbsp_bit(rbsp, &ptl->general_tier_flag); in nal_hevc_rbsp_profile_tier_level() 108 rbsp_bits(rbsp, 5, &ptl->general_profile_idc); in nal_hevc_rbsp_profile_tier_level() 110 rbsp_bit(rbsp, &ptl->general_profile_compatibility_flag[i]); in nal_hevc_rbsp_profile_tier_level() 111 rbsp_bit(rbsp, &ptl->general_progressive_source_flag); in nal_hevc_rbsp_profile_tier_level() 112 rbsp_bit(rbsp, &ptl->general_interlaced_source_flag); in nal_hevc_rbsp_profile_tier_level() 113 rbsp_bit(rbsp, &ptl->general_non_packed_constraint_flag); in nal_hevc_rbsp_profile_tier_level() 114 rbsp_bit(rbsp, &ptl->general_frame_only_constraint_flag); in nal_hevc_rbsp_profile_tier_level() 115 if (ptl->general_profile_idc == 4 || in nal_hevc_rbsp_profile_tier_level() [all …]
|
| /linux/mm/ |
| H A D | page_vma_mapped.c | 25 pvmw->address, &pvmw->ptl); in map_pte() 26 *ptlp = pvmw->ptl; in map_pte() 80 pvmw->ptl = *ptlp; in map_pte() 185 spinlock_t *ptl; in page_vma_mapped_walk() local 210 pvmw->ptl = huge_pte_lock(hstate, mm, pvmw->pte); in page_vma_mapped_walk() 246 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 269 spin_unlock(pvmw->ptl); in page_vma_mapped_walk() 270 pvmw->ptl = NULL; in page_vma_mapped_walk() 280 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 288 spinlock_t *ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() local [all …]
|
| H A D | pt_reclaim.c | |
| H A D | mincore.c | 32 spinlock_t *ptl; in mincore_hugetlb() local 34 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); in mincore_hugetlb() 54 spin_unlock(ptl); in mincore_hugetlb() 166 spinlock_t *ptl; in mincore_pte_range() local 173 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 174 if (ptl) { in mincore_pte_range() 176 spin_unlock(ptl); in mincore_pte_range() 180 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range() 211 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
| H A D | migrate_device.c | 143 spinlock_t *ptl; in migrate_vma_collect_huge_pmd() local 147 ptl = pmd_lock(mm, pmdp); in migrate_vma_collect_huge_pmd() 149 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 155 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 161 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 174 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 179 migration_entry_wait_on_locked(entry, ptl); in migrate_vma_collect_huge_pmd() 180 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 187 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 193 spin_unlock(ptl); in migrate_vma_collect_huge_pmd() 256 spinlock_t *ptl; migrate_vma_collect_pmd() local 804 spinlock_t *ptl; migrate_vma_insert_huge_pmd_page() local 978 spinlock_t *ptl; migrate_vma_insert_page() local [all...] |
| H A D | huge_memory.c | 1338 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 1348 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 1358 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 1363 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 1376 spinlock_t *ptl; in do_huge_pmd_device_private() local 1386 ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_device_private() 1388 spin_unlock(ptl); in do_huge_pmd_device_private() 1399 spin_unlock(ptl); in do_huge_pmd_device_private() 1404 spin_unlock(ptl); in do_huge_pmd_device_private() 1487 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page() [all …]
|
| H A D | pagewalk.c | 67 spinlock_t *ptl; in walk_pte_range() local 86 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in walk_pte_range() 89 pte_unmap_unlock(pte, ptl); in walk_pte_range() 893 spinlock_t *ptl; in folio_walk_start() 917 ptl = pud_lock(vma->vm_mm, pudp); in folio_walk_start() 926 spin_unlock(ptl); in folio_walk_start() 929 spin_unlock(ptl); in folio_walk_start() 940 spin_unlock(ptl); in folio_walk_start() 952 ptl = pmd_lock(vma->vm_mm, pmdp); in folio_walk_start() 961 spin_unlock(ptl); in folio_walk_start() 892 spinlock_t *ptl; folio_walk_start() local [all...] |
| H A D | memory.c | 440 spinlock_t *ptl = pmd_lock(mm, pmd); in pmd_install() local 461 spin_unlock(ptl); in pmd_install() 1847 spinlock_t *ptl, pmd_t *pmdval) in zap_empty_pte_table() argument 1851 if (ptl != pml && !spin_trylock(pml)) in zap_empty_pte_table() 1856 if (ptl != pml) in zap_empty_pte_table() 1864 spinlock_t *pml, *ptl = NULL; in zap_pte_table_if_empty() local 1869 start_pte = pte_offset_map_rw_nolock(mm, pmd, addr, pmdval, &ptl); in zap_pte_table_if_empty() 1872 if (ptl != pml) in zap_pte_table_if_empty() 1873 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in zap_pte_table_if_empty() 1883 if (ptl != pml) in zap_pte_table_if_empty() [all …]
|
| H A D | khugepaged.c | 673 spinlock_t *ptl, in __collapse_huge_page_copy_succeeded() argument 693 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 695 spin_unlock(ptl); in __collapse_huge_page_copy_succeeded() 715 spin_lock(ptl); in __collapse_huge_page_copy_succeeded() 718 spin_unlock(ptl); in __collapse_huge_page_copy_succeeded() 775 unsigned long address, spinlock_t *ptl, in __collapse_huge_page_copy() argument 802 __collapse_huge_page_copy_succeeded(pte, vma, address, ptl, in __collapse_huge_page_copy() 983 spinlock_t *ptl; in __collapse_huge_page_swapin() local 999 pte = pte_offset_map_ro_nolock(mm, pmd, addr, &ptl); in __collapse_huge_page_swapin() 1013 vmf.ptl = ptl; in __collapse_huge_page_swapin() [all …]
|
| H A D | madvise.c | 191 spinlock_t *ptl; in swapin_walk_pmd_entry() local 200 ptep = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in swapin_walk_pmd_entry() 210 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 220 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 362 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local 381 ptl = pmd_trans_huge_lock(pmd, vma); in madvise_cold_or_pageout_pte_range() 382 if (!ptl) in madvise_cold_or_pageout_pte_range() 408 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 440 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 450 start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range() [all …]
|
| H A D | hmm.c | 491 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local 493 if (!ptl) in hmm_vma_walk_pud() 501 spin_unlock(ptl); in hmm_vma_walk_pud() 519 spin_unlock(ptl); in hmm_vma_walk_pud() 535 spin_unlock(ptl); in hmm_vma_walk_pud() 554 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local 557 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 569 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry() 590 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
|
| H A D | pgtable-generic.c | 346 * struct page). pte_unmap_unlock(pte, ptl) to unlock and unmap afterwards. 394 spinlock_t *ptl; in __pte_offset_map_lock() 401 ptl = pte_lockptr(mm, &pmdval); in __pte_offset_map_lock() 402 spin_lock(ptl); in __pte_offset_map_lock() 404 *ptlp = ptl; in __pte_offset_map_lock() 407 pte_unmap_unlock(pte, ptl); in __pte_offset_map_lock() 395 spinlock_t *ptl; __pte_offset_map_lock() local
|
| H A D | userfaultfd.c | 146 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */ 178 spinlock_t *ptl; in mfill_atomic_install_pte() local 193 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte() 236 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_install_pte() 354 spinlock_t *ptl; in mfill_atomic_pte_zeropage() local 363 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_zeropage() 378 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_pte_zeropage() 436 spinlock_t *ptl; in mfill_atomic_pte_poison() local 440 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_poison() 460 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_pte_poison() 1820 spinlock_t *ptl; move_pages() local [all...] |
| /linux/mm/damon/ |
| H A D | vaddr.c | 310 spinlock_t *ptl; in damon_mkold_pmd_entry() local 312 ptl = pmd_trans_huge_lock(pmd, walk->vma); in damon_mkold_pmd_entry() 313 if (ptl) { in damon_mkold_pmd_entry() 318 spin_unlock(ptl); in damon_mkold_pmd_entry() 322 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in damon_mkold_pmd_entry() 329 pte_unmap_unlock(pte, ptl); in damon_mkold_pmd_entry() 366 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local 369 ptl = huge_pte_lock(h, walk->mm, pte); in damon_mkold_hugetlb_entry() 377 spin_unlock(ptl); in damon_mkold_hugetlb_entry() 436 spinlock_t *ptl; in damon_young_pmd_entry() local [all …]
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | split_page_table_lock.rst | 63 面。这个区域与page->ptl共享存储。 79 page->ptl 82 page->ptl用于访问分割页表锁,其中'page'是包含该表的页面struct page。它 89 - 如果spinlock_t的大小大于long的大小,我们使用page->ptl作为spinlock_t 96 请不要直接访问page->ptl - -使用适当的辅助函数。
|
| /linux/arch/arm/mm/ |
| H A D | fault-armv.c | 67 spinlock_t *ptl; in adjust_pte() local 98 pte = pte_offset_map_rw_nolock(vma->vm_mm, pmd, address, &pmdval, &ptl); in adjust_pte() 107 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in adjust_pte() 109 pte_unmap_unlock(pte, ptl); in adjust_pte() 117 spin_unlock(ptl); in adjust_pte()
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 1113 spinlock_t *ptl; in smaps_pte_range() local 1115 ptl = pmd_trans_huge_lock(pmd, vma); in smaps_pte_range() 1116 if (ptl) { in smaps_pte_range() 1118 spin_unlock(ptl); in smaps_pte_range() 1122 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range() 1129 pte_unmap_unlock(pte - 1, ptl); in smaps_pte_range() 1240 spinlock_t *ptl; in smaps_hugetlb_range() local 1243 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in smaps_hugetlb_range() 1263 spin_unlock(ptl); in smaps_hugetlb_range() 1686 spinlock_t *ptl; in clear_refs_pte_range() local [all …]
|
| /linux/arch/powerpc/lib/ |
| H A D | code-patching.c | 156 spinlock_t *ptl; in text_area_cpu_up_mm() local 177 pte = get_locked_pte(mm, addr, &ptl); in text_area_cpu_up_mm() 180 pte_unmap_unlock(pte, ptl); in text_area_cpu_up_mm() 290 spinlock_t *ptl; in __do_patch_mem_mm() local 296 pte = get_locked_pte(patching_mm, text_poke_addr, &ptl); in __do_patch_mem_mm() 322 pte_unmap_unlock(pte, ptl); in __do_patch_mem_mm() 473 spinlock_t *ptl; in __do_patch_instructions_mm() local 482 pte = get_locked_pte(patching_mm, text_poke_addr, &ptl); in __do_patch_instructions_mm() 510 pte_unmap_unlock(pte, ptl); in __do_patch_instructions_mm()
|
| /linux/arch/x86/kernel/ |
| H A D | ldt.c | 292 spinlock_t *ptl; in map_ldt_struct() local 326 ptep = get_locked_pte(mm, va, &ptl); in map_ldt_struct() 339 pte_unmap_unlock(ptep, ptl); in map_ldt_struct() 365 spinlock_t *ptl; in unmap_ldt_struct() local 369 ptep = get_locked_pte(mm, va, &ptl); in unmap_ldt_struct() 372 pte_unmap_unlock(ptep, ptl); in unmap_ldt_struct()
|
| /linux/arch/m68k/kernel/ |
| H A D | sys_m68k.c | 474 spinlock_t *ptl; in sys_atomic_cmpxchg_32() local 490 pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl); in sys_atomic_cmpxchg_32() 495 pte_unmap_unlock(pte, ptl); in sys_atomic_cmpxchg_32() 507 pte_unmap_unlock(pte, ptl); in sys_atomic_cmpxchg_32()
|
| /linux/drivers/hid/intel-thc-hid/intel-quickspi/ |
| H A D | pci-quickspi.c | 32 struct quickspi_driver_data ptl = { variable 977 {PCI_DEVICE_DATA(INTEL, THC_PTL_H_DEVICE_ID_SPI_PORT1, &ptl), }, 978 {PCI_DEVICE_DATA(INTEL, THC_PTL_H_DEVICE_ID_SPI_PORT2, &ptl), }, 979 {PCI_DEVICE_DATA(INTEL, THC_PTL_U_DEVICE_ID_SPI_PORT1, &ptl), }, 980 {PCI_DEVICE_DATA(INTEL, THC_PTL_U_DEVICE_ID_SPI_PORT2, &ptl), }, 981 {PCI_DEVICE_DATA(INTEL, THC_WCL_DEVICE_ID_SPI_PORT1, &ptl), }, 982 {PCI_DEVICE_DATA(INTEL, THC_WCL_DEVICE_ID_SPI_PORT2, &ptl), },
|