Lines Matching refs:spool

100 static inline bool subpool_is_free(struct hugepage_subpool *spool)  in subpool_is_free()  argument
102 if (spool->count) in subpool_is_free()
104 if (spool->max_hpages != -1) in subpool_is_free()
105 return spool->used_hpages == 0; in subpool_is_free()
106 if (spool->min_hpages != -1) in subpool_is_free()
107 return spool->rsv_hpages == spool->min_hpages; in subpool_is_free()
112 static inline void unlock_or_release_subpool(struct hugepage_subpool *spool, in unlock_or_release_subpool() argument
115 spin_unlock_irqrestore(&spool->lock, irq_flags); in unlock_or_release_subpool()
120 if (subpool_is_free(spool)) { in unlock_or_release_subpool()
121 if (spool->min_hpages != -1) in unlock_or_release_subpool()
122 hugetlb_acct_memory(spool->hstate, in unlock_or_release_subpool()
123 -spool->min_hpages); in unlock_or_release_subpool()
124 kfree(spool); in unlock_or_release_subpool()
131 struct hugepage_subpool *spool; in hugepage_new_subpool() local
133 spool = kzalloc(sizeof(*spool), GFP_KERNEL); in hugepage_new_subpool()
134 if (!spool) in hugepage_new_subpool()
137 spin_lock_init(&spool->lock); in hugepage_new_subpool()
138 spool->count = 1; in hugepage_new_subpool()
139 spool->max_hpages = max_hpages; in hugepage_new_subpool()
140 spool->hstate = h; in hugepage_new_subpool()
141 spool->min_hpages = min_hpages; in hugepage_new_subpool()
144 kfree(spool); in hugepage_new_subpool()
147 spool->rsv_hpages = min_hpages; in hugepage_new_subpool()
149 return spool; in hugepage_new_subpool()
152 void hugepage_put_subpool(struct hugepage_subpool *spool) in hugepage_put_subpool() argument
156 spin_lock_irqsave(&spool->lock, flags); in hugepage_put_subpool()
157 BUG_ON(!spool->count); in hugepage_put_subpool()
158 spool->count--; in hugepage_put_subpool()
159 unlock_or_release_subpool(spool, flags); in hugepage_put_subpool()
170 static long hugepage_subpool_get_pages(struct hugepage_subpool *spool, in hugepage_subpool_get_pages() argument
175 if (!spool) in hugepage_subpool_get_pages()
178 spin_lock_irq(&spool->lock); in hugepage_subpool_get_pages()
180 if (spool->max_hpages != -1) { /* maximum size accounting */ in hugepage_subpool_get_pages()
181 if ((spool->used_hpages + delta) <= spool->max_hpages) in hugepage_subpool_get_pages()
182 spool->used_hpages += delta; in hugepage_subpool_get_pages()
190 if (spool->min_hpages != -1 && spool->rsv_hpages) { in hugepage_subpool_get_pages()
191 if (delta > spool->rsv_hpages) { in hugepage_subpool_get_pages()
196 ret = delta - spool->rsv_hpages; in hugepage_subpool_get_pages()
197 spool->rsv_hpages = 0; in hugepage_subpool_get_pages()
200 spool->rsv_hpages -= delta; in hugepage_subpool_get_pages()
205 spin_unlock_irq(&spool->lock); in hugepage_subpool_get_pages()
215 static long hugepage_subpool_put_pages(struct hugepage_subpool *spool, in hugepage_subpool_put_pages() argument
221 if (!spool) in hugepage_subpool_put_pages()
224 spin_lock_irqsave(&spool->lock, flags); in hugepage_subpool_put_pages()
226 if (spool->max_hpages != -1) /* maximum size accounting */ in hugepage_subpool_put_pages()
227 spool->used_hpages -= delta; in hugepage_subpool_put_pages()
230 if (spool->min_hpages != -1 && spool->used_hpages < spool->min_hpages) { in hugepage_subpool_put_pages()
231 if (spool->rsv_hpages + delta <= spool->min_hpages) in hugepage_subpool_put_pages()
234 ret = spool->rsv_hpages + delta - spool->min_hpages; in hugepage_subpool_put_pages()
236 spool->rsv_hpages += delta; in hugepage_subpool_put_pages()
237 if (spool->rsv_hpages > spool->min_hpages) in hugepage_subpool_put_pages()
238 spool->rsv_hpages = spool->min_hpages; in hugepage_subpool_put_pages()
245 unlock_or_release_subpool(spool, flags); in hugepage_subpool_put_pages()
252 return HUGETLBFS_SB(inode->i_sb)->spool; in subpool_inode()
903 struct hugepage_subpool *spool = subpool_inode(inode); in hugetlb_fix_reserve_counts() local
907 rsv_adjust = hugepage_subpool_get_pages(spool, 1); in hugetlb_fix_reserve_counts()
1860 struct hugepage_subpool *spool = hugetlb_folio_subpool(folio); in free_huge_page() local
1889 if (hugepage_subpool_put_pages(spool, 1) == 0) in free_huge_page()
3022 struct hugepage_subpool *spool = subpool_vma(vma); in alloc_hugetlb_folio() local
3049 gbl_chg = hugepage_subpool_get_pages(spool, 1); in alloc_hugetlb_folio()
3114 hugetlb_set_folio_subpool(folio, spool); in alloc_hugetlb_folio()
3129 rsv_adjust = hugepage_subpool_put_pages(spool, 1); in alloc_hugetlb_folio()
3145 hugepage_subpool_put_pages(spool, 1); in alloc_hugetlb_folio()
4810 struct hugepage_subpool *spool = subpool_vma(vma); in hugetlb_vm_op_close() local
4830 gbl_reserve = hugepage_subpool_put_pages(spool, reserve); in hugetlb_vm_op_close()
6773 struct hugepage_subpool *spool = subpool_inode(inode); in hugetlb_reserve_pages() local
6844 gbl_reserve = hugepage_subpool_get_pages(spool, chg); in hugetlb_reserve_pages()
6890 rsv_adjust = hugepage_subpool_put_pages(spool, in hugetlb_reserve_pages()
6907 (void)hugepage_subpool_put_pages(spool, chg); in hugetlb_reserve_pages()
6930 struct hugepage_subpool *spool = subpool_inode(inode); in hugetlb_unreserve_pages() local
6959 gbl_reserve = hugepage_subpool_put_pages(spool, (chg - freed)); in hugetlb_unreserve_pages()