/linux-6.3-rc2/fs/minix/ |
A D | namei.c | 205 struct page * new_page; in minix_rename() local 213 new_de = minix_find_entry(new_dentry, &new_page); in minix_rename() 216 err = minix_set_link(new_de, new_page, old_inode); in minix_rename() 217 kunmap(new_page); in minix_rename() 218 put_page(new_page); in minix_rename()
|
/linux-6.3-rc2/fs/sysv/ |
A D | namei.c | 218 struct page * new_page; in sysv_rename() local 226 new_de = sysv_find_entry(new_dentry, &new_page); in sysv_rename() 229 err = sysv_set_link(new_de, new_page, old_inode); in sysv_rename() 230 dir_put_page(new_page, new_de); in sysv_rename()
|
/linux-6.3-rc2/kernel/events/ |
A D | uprobes.c | 166 if (new_page) { in __replace_page() 167 new_folio = page_folio(new_page); in __replace_page() 182 if (new_page) { in __replace_page() 197 if (new_page) in __replace_page() 464 struct page *old_page, *new_page; in uprobe_write_opcode() local 511 if (!new_page) in uprobe_write_opcode() 514 __SetPageUptodate(new_page); in uprobe_write_opcode() 532 put_page(new_page); in uprobe_write_opcode() 533 new_page = NULL; in uprobe_write_opcode() 543 if (new_page) in uprobe_write_opcode() [all …]
|
/linux-6.3-rc2/arch/s390/mm/ |
A D | vmem.c | 174 void *new_page = vmemmap_alloc_block(PAGE_SIZE, NUMA_NO_NODE); in modify_pte_table() local 176 if (!new_page) in modify_pte_table() 178 set_pte(pte, __pte(__pa(new_page) | prot)); in modify_pte_table() 250 void *new_page; in modify_pmd_table() local 259 new_page = vmemmap_alloc_block(PMD_SIZE, NUMA_NO_NODE); in modify_pmd_table() 260 if (new_page) { in modify_pmd_table() 261 set_pmd(pmd, __pmd(__pa(new_page) | prot)); in modify_pmd_table()
|
/linux-6.3-rc2/fs/f2fs/ |
A D | namei.c | 948 struct page *old_page, *new_page = NULL; in f2fs_rename() local 1023 &new_page); in f2fs_rename() 1025 if (IS_ERR(new_page)) in f2fs_rename() 1026 err = PTR_ERR(new_page); in f2fs_rename() 1039 new_page = NULL; in f2fs_rename() 1119 f2fs_put_page(new_page, 0); in f2fs_rename() 1137 struct page *old_page, *new_page; in f2fs_cross_rename() local 1174 if (IS_ERR(new_page)) in f2fs_cross_rename() 1175 err = PTR_ERR(new_page); in f2fs_cross_rename() 1249 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_cross_rename() [all …]
|
/linux-6.3-rc2/fs/ext2/ |
A D | namei.c | 360 struct page *new_page; in ext2_rename() local 368 &new_page, &page_addr); in ext2_rename() 373 err = ext2_set_link(new_dir, new_de, new_page, page_addr, in ext2_rename() 375 ext2_put_page(new_page, page_addr); in ext2_rename()
|
/linux-6.3-rc2/fs/ufs/ |
A D | namei.c | 273 struct page *new_page; in ufs_rename() local 281 new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_page); in ufs_rename() 284 ufs_set_link(new_dir, new_de, new_page, old_inode, 1); in ufs_rename()
|
/linux-6.3-rc2/fs/ubifs/ |
A D | budget.c | 363 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 380 if (req->new_page) in calc_data_growth() 426 ubifs_assert(c, req->new_page <= 1); in ubifs_budget_space() 513 ubifs_assert(c, req->new_page <= 1); in ubifs_release_budget()
|
A D | file.c | 198 struct ubifs_budget_req req = { .recalculate = 1, .new_page = 1 }; in release_new_page_budget() 223 struct ubifs_budget_req req = { .new_page = 1 }; in write_begin_slow() 364 req.new_page = 1; in allocate_budget() 1507 struct ubifs_budget_req req = { .new_page = 1 }; in ubifs_vm_page_mkwrite()
|
A D | ubifs.h | 897 unsigned int new_page:1; member 907 unsigned int new_page; member
|
A D | debug.c | 591 req->new_page, req->dirtied_page); in ubifs_dump_budget_req()
|
/linux-6.3-rc2/mm/ |
A D | ksm.c | 2609 struct page *new_page; in ksm_might_need_to_copy() local 2625 if (new_page && in ksm_might_need_to_copy() 2627 put_page(new_page); in ksm_might_need_to_copy() 2628 new_page = NULL; in ksm_might_need_to_copy() 2630 if (new_page) { in ksm_might_need_to_copy() 2631 if (copy_mc_user_highpage(new_page, page, address, vma)) { in ksm_might_need_to_copy() 2632 put_page(new_page); in ksm_might_need_to_copy() 2636 SetPageDirty(new_page); in ksm_might_need_to_copy() 2637 __SetPageUptodate(new_page); in ksm_might_need_to_copy() 2638 __SetPageLocked(new_page); in ksm_might_need_to_copy() [all …]
|
A D | mempolicy.c | 1207 static struct page *new_page(struct page *page, unsigned long start) in new_page() function 1251 static struct page *new_page(struct page *page, unsigned long start) in new_page() function 1338 nr_failed = migrate_pages(&pagelist, new_page, NULL, in do_mbind()
|
/linux-6.3-rc2/drivers/net/ethernet/ti/ |
A D | cpsw.c | 346 struct page *new_page, *page = token; in cpsw_rx_handler() local 378 new_page = page; in cpsw_rx_handler() 387 new_page = page_pool_dev_alloc_pages(pool); in cpsw_rx_handler() 388 if (unlikely(!new_page)) { in cpsw_rx_handler() 389 new_page = page; in cpsw_rx_handler() 441 xmeta = page_address(new_page) + CPSW_XMETA_OFFSET; in cpsw_rx_handler() 445 dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM_NA; in cpsw_rx_handler() 446 ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, in cpsw_rx_handler() 450 page_pool_recycle_direct(pool, new_page); in cpsw_rx_handler()
|
A D | cpsw_new.c | 284 struct page *new_page, *page = token; in cpsw_rx_handler() local 322 new_page = page; in cpsw_rx_handler() 331 new_page = page_pool_dev_alloc_pages(pool); in cpsw_rx_handler() 332 if (unlikely(!new_page)) { in cpsw_rx_handler() 333 new_page = page; in cpsw_rx_handler() 385 xmeta = page_address(new_page) + CPSW_XMETA_OFFSET; in cpsw_rx_handler() 389 dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM_NA; in cpsw_rx_handler() 390 ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, in cpsw_rx_handler() 394 page_pool_recycle_direct(pool, new_page); in cpsw_rx_handler()
|
/linux-6.3-rc2/fs/nilfs2/ |
A D | namei.c | 377 struct page *new_page; in nilfs_rename() local 385 new_de = nilfs_find_entry(new_dir, &new_dentry->d_name, &new_page); in nilfs_rename() 388 nilfs_set_link(new_dir, new_de, new_page, old_inode); in nilfs_rename()
|
/linux-6.3-rc2/drivers/net/ethernet/microsoft/mana/ |
A D | mana_en.c | 1257 struct page *new_page; in mana_process_rx_cqe() local 1300 new_page = rxq->xdp_save_page; in mana_process_rx_cqe() 1303 new_page = alloc_page(GFP_ATOMIC); in mana_process_rx_cqe() 1306 if (new_page) { in mana_process_rx_cqe() 1307 da = dma_map_page(dev, new_page, XDP_PACKET_HEADROOM, rxq->datasize, in mana_process_rx_cqe() 1311 __free_page(new_page); in mana_process_rx_cqe() 1312 new_page = NULL; in mana_process_rx_cqe() 1316 new_buf = new_page ? page_to_virt(new_page) : NULL; in mana_process_rx_cqe()
|
/linux-6.3-rc2/drivers/tty/serial/ |
A D | icom.c | 607 unsigned char *new_page = NULL; in load_code() local 680 new_page = dma_alloc_coherent(&dev->dev, 4096, &temp_pci, GFP_KERNEL); in load_code() 682 if (!new_page) { in load_code() 702 new_page[index] = fw->data[index]; in load_code() 759 if (new_page != NULL) in load_code() 760 dma_free_coherent(&dev->dev, 4096, new_page, temp_pci); in load_code()
|
/linux-6.3-rc2/fs/jbd2/ |
A D | journal.c | 344 struct page *new_page; in jbd2_journal_write_metadata_buffer() local 373 new_page = virt_to_page(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 376 new_page = jh2bh(jh_in)->b_page; in jbd2_journal_write_metadata_buffer() 380 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 420 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 424 new_page = virt_to_page(tmp); in jbd2_journal_write_metadata_buffer() 441 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 446 set_bh_page(new_bh, new_page, new_offset); in jbd2_journal_write_metadata_buffer()
|
/linux-6.3-rc2/drivers/net/wireless/intel/iwlwifi/fw/ |
A D | dbg.c | 569 struct page *new_page; in alloc_sgtable() local 580 new_page = alloc_page(GFP_KERNEL); in alloc_sgtable() 581 if (!new_page) { in alloc_sgtable() 585 new_page = sg_page(iter); in alloc_sgtable() 586 if (new_page) in alloc_sgtable() 587 __free_page(new_page); in alloc_sgtable() 594 sg_set_page(iter, new_page, alloc_size, 0); in alloc_sgtable()
|
/linux-6.3-rc2/drivers/staging/rts5208/ |
A D | xd.c | 1102 u32 old_page, new_page; in xd_copy_page() local 1116 new_page = (new_blk << xd_card->block_shift) + start_page; in xd_copy_page() 1185 xd_assign_phy_addr(chip, new_page, XD_RW_ADDR); in xd_copy_page() 1208 new_page++; in xd_copy_page()
|
/linux-6.3-rc2/Documentation/networking/ |
A D | page_pool.rst | 197 new_page = page_pool_dev_alloc_pages(page_pool);
|
/linux-6.3-rc2/drivers/net/vmxnet3/ |
A D | vmxnet3_drv.c | 1433 struct page *new_page = NULL; in vmxnet3_rq_rx_complete() local 1584 new_page = alloc_page(GFP_ATOMIC); in vmxnet3_rq_rx_complete() 1590 if (unlikely(!new_page)) { in vmxnet3_rq_rx_complete() 1598 new_page, in vmxnet3_rq_rx_complete() 1603 put_page(new_page); in vmxnet3_rq_rx_complete() 1618 rbi->page = new_page; in vmxnet3_rq_rx_complete()
|
/linux-6.3-rc2/drivers/net/ethernet/freescale/ |
A D | fec_main.c | 1496 struct page *new_page; in fec_enet_update_cbd() local 1499 new_page = page_pool_dev_alloc_pages(rxq->page_pool); in fec_enet_update_cbd() 1500 WARN_ON(!new_page); in fec_enet_update_cbd() 1501 rxq->rx_skb_info[index].page = new_page; in fec_enet_update_cbd() 1504 phys_addr = page_pool_get_dma_addr(new_page) + FEC_ENET_XDP_HEADROOM; in fec_enet_update_cbd()
|