1 /* SPDX-License-Identifier: BSD-2-Clause */
2 /*
3 * Copyright (c) 2016-2021, Linaro Limited
4 * Copyright (c) 2014, STMicroelectronics International N.V.
5 */
6
7 #ifndef MM_TEE_PAGER_H
8 #define MM_TEE_PAGER_H
9
10 #include <kernel/abort.h>
11 #include <kernel/panic.h>
12 #include <kernel/user_ta.h>
13 #include <mm/core_mmu.h>
14 #include <mm/tee_mm.h>
15 #include <string.h>
16 #include <trace.h>
17
18 /*
19 * tee_pager_early_init() - Perform early initialization of pager
20 *
21 * Panics if some error occurs
22 */
23 void tee_pager_early_init(void);
24
25 /*
26 * tee_pager_get_table_info() - Fills in table info for address mapped in
27 * translation table managed by the pager.
28 * @va: address to look up
29 * @ti: filled in table info
30 *
31 * Returns true if address is in the pager translation tables else false
32 */
33 bool tee_pager_get_table_info(vaddr_t va, struct core_mmu_table_info *ti);
34
35 /*
36 * tee_pager_phys_to_virt() - Translate physical address to virtual address
37 * looking in the pager page tables
38 * @pa: address to translate
39 * @len: check for length is mapped linearly in CORE_MMU_PGDIR_SIZE range
40 *
41 * Returns found virtual address or NULL on error
42 */
43 void *tee_pager_phys_to_virt(paddr_t pa, size_t len);
44
45 /*
46 * tee_pager_set_alias_area() - Initialize pager alias area
47 * @mm_alias: The alias area where all physical pages managed by the
48 * pager are aliased
49 *
50 * Panics if called twice or some other error occurs.
51 */
52 void tee_pager_set_alias_area(tee_mm_entry_t *mm_alias);
53
54 /*
55 * tee_pager_init_iv_region() - Initialized pager region for tags IVs used by RW
56 * paged fobjs
57 * @fobj: fobj backing the region
58 *
59 * Panics if called twice or some other error occurs.
60 *
61 * Returns virtual address of start of IV region.
62 */
63 vaddr_t tee_pager_init_iv_region(struct fobj *fobj);
64
65 /*
66 * tee_pager_generate_authenc_key() - Generates authenc key for r/w paging
67 *
68 * Needs to draw random from RNG, panics if some error occurs.
69 */
70 #ifdef CFG_WITH_PAGER
71 void tee_pager_generate_authenc_key(void);
72 #else
tee_pager_generate_authenc_key(void)73 static inline void tee_pager_generate_authenc_key(void)
74 {
75 }
76 #endif
77
78 /*
79 * tee_pager_add_core_region() - Adds a pageable core region
80 * @base: base of covered memory region
81 * @type: type of memory region
82 * @fobj: fobj backing the region
83 *
84 * Non-page aligned base or size will cause a panic.
85 */
86 void tee_pager_add_core_region(vaddr_t base, enum vm_paged_region_type type,
87 struct fobj *fobj);
88
89 /*
90 * tee_pager_add_um_region() - Adds a pageable user TA region
91 * @uctx: user mode context of the region
92 * @base: base of covered memory region
93 * @fobj: fobj of the store backing the memory region
94 *
95 * The mapping is created suitable to initialize the memory content while
96 * loading the TA. Once the TA is properly loaded the regions should be
97 * finalized with tee_pager_set_um_region_attr() to get more strict settings.
98 *
99 * Return TEE_SUCCESS on success, anything else if the region can't be added
100 */
101 #ifdef CFG_PAGED_USER_TA
102 TEE_Result tee_pager_add_um_region(struct user_mode_ctx *uctx, vaddr_t base,
103 struct fobj *fobj, uint32_t prot);
104 #else
105 static inline TEE_Result
tee_pager_add_um_region(struct user_mode_ctx * uctx __unused,vaddr_t base __unused,struct fobj * fobj __unused,uint32_t prot __unused)106 tee_pager_add_um_region(struct user_mode_ctx *uctx __unused,
107 vaddr_t base __unused, struct fobj *fobj __unused,
108 uint32_t prot __unused)
109 {
110 return TEE_ERROR_NOT_SUPPORTED;
111 }
112 #endif
113
114 /*
115 * tee_pager_set_um_region_attr() - Set attributes of a initialized memory
116 * region
117 * @uctx: user mode context of the region
118 * @base: base of covered memory region
119 * @size: size of covered memory region
120 * @flags: TEE_MATTR_U* flags describing permissions of the region
121 *
122 * Return true on success of false if the region can't be updated
123 */
124 #ifdef CFG_PAGED_USER_TA
125 bool tee_pager_set_um_region_attr(struct user_mode_ctx *uctx, vaddr_t base,
126 size_t size, uint32_t flags);
127 #else
128 static inline bool
tee_pager_set_um_region_attr(struct user_mode_ctx * uctx __unused,vaddr_t base __unused,size_t size __unused,uint32_t flags __unused)129 tee_pager_set_um_region_attr(struct user_mode_ctx *uctx __unused,
130 vaddr_t base __unused, size_t size __unused,
131 uint32_t flags __unused)
132 {
133 return false;
134 }
135 #endif
136
137 #ifdef CFG_PAGED_USER_TA
138 void tee_pager_rem_um_region(struct user_mode_ctx *uctx, vaddr_t base,
139 size_t size);
140 #else
tee_pager_rem_um_region(struct user_mode_ctx * uctx __unused,vaddr_t base __unused,size_t size __unused)141 static inline void tee_pager_rem_um_region(struct user_mode_ctx *uctx __unused,
142 vaddr_t base __unused,
143 size_t size __unused)
144 {
145 }
146 #endif
147
148 #ifdef CFG_PAGED_USER_TA
149 TEE_Result tee_pager_split_um_region(struct user_mode_ctx *uctx, vaddr_t va);
150 #else
151 static inline TEE_Result
tee_pager_split_um_region(struct user_mode_ctx * uctx __unused,vaddr_t va __unused)152 tee_pager_split_um_region(struct user_mode_ctx *uctx __unused,
153 vaddr_t va __unused)
154 {
155 return TEE_ERROR_NOT_SUPPORTED;
156 }
157
158 #endif
159 #ifdef CFG_PAGED_USER_TA
160 void tee_pager_merge_um_region(struct user_mode_ctx *uctx, vaddr_t va,
161 size_t len);
162 #else
163 static inline void
tee_pager_merge_um_region(struct user_mode_ctx * uctx __unused,vaddr_t va __unused,size_t len __unused)164 tee_pager_merge_um_region(struct user_mode_ctx *uctx __unused,
165 vaddr_t va __unused, size_t len __unused)
166 {
167 }
168 #endif
169
170 /*
171 * tee_pager_rem_uma_regions() - Remove all user TA regions
172 * @uctx: user mode context
173 *
174 * This function is called when a user mode context is teared down.
175 */
176 #ifdef CFG_PAGED_USER_TA
177 void tee_pager_rem_um_regions(struct user_mode_ctx *uctx);
178 #else
tee_pager_rem_um_regions(struct user_mode_ctx * uctx __unused)179 static inline void tee_pager_rem_um_regions(struct user_mode_ctx *uctx __unused)
180 {
181 }
182 #endif
183
184 /*
185 * tee_pager_assign_um_tables() - Assigns translation table to a user ta
186 * @uctx: user mode context
187 *
188 * This function is called to assign translation tables for the pageable
189 * regions of a user TA.
190 */
191 #ifdef CFG_PAGED_USER_TA
192 void tee_pager_assign_um_tables(struct user_mode_ctx *uctx);
193 #else
194 static inline void
tee_pager_assign_um_tables(struct user_mode_ctx * uctx __unused)195 tee_pager_assign_um_tables(struct user_mode_ctx *uctx __unused)
196 {
197 }
198 #endif
199
200 /*
201 * Adds physical pages to the pager to use. The supplied virtual address range
202 * is searched for mapped physical pages and unmapped pages are ignored.
203 *
204 * vaddr is the first virtual address
205 * npages is the number of pages to add
206 */
207 void tee_pager_add_pages(vaddr_t vaddr, size_t npages, bool unmap);
208
209 /*
210 * tee_pager_alloc() - Allocate read-write virtual memory from pager.
211 * @size: size of memory in bytes
212 *
213 * @return NULL on failure or a pointer to the virtual memory on success.
214 */
215 void *tee_pager_alloc(size_t size);
216
217 #ifdef CFG_PAGED_USER_TA
218 /*
219 * tee_pager_pgt_save_and_release_entries() - Save dirty pages to backing store
220 * and remove physical page from translation table
221 * @pgt: page table descriptor
222 *
223 * This function is called when a translation table needs to be recycled
224 */
225 void tee_pager_pgt_save_and_release_entries(struct pgt *pgt);
226 #else
227 static inline void
tee_pager_pgt_save_and_release_entries(struct pgt * pgt __unused)228 tee_pager_pgt_save_and_release_entries(struct pgt *pgt __unused)
229 {
230 }
231 #endif
232
233 /*
234 * tee_pager_release_phys() - Release physical pages used for mapping
235 * @addr: virtual address of first page to release
236 * @size: number of bytes to release
237 *
238 * Only pages completely covered by the supplied range are affected. This
239 * function only supplies a hint to the pager that the physical page can be
240 * reused. The caller can't expect a released memory range to hold a
241 * specific bit pattern when used next time.
242 *
243 * Note that the virtual memory allocation is still valid after this
244 * function has returned, it's just the content that may or may not have
245 * changed.
246 */
247 #ifdef CFG_WITH_PAGER
248 void tee_pager_release_phys(void *addr, size_t size);
249 #else
tee_pager_release_phys(void * addr __unused,size_t size __unused)250 static inline void tee_pager_release_phys(void *addr __unused,
251 size_t size __unused)
252 {
253 }
254 #endif
255
256 /*
257 * Statistics on the pager
258 */
259 struct tee_pager_stats {
260 size_t hidden_hits;
261 size_t ro_hits;
262 size_t rw_hits;
263 size_t zi_released;
264 size_t npages; /* number of load pages */
265 size_t npages_all; /* number of pages */
266 };
267
268 #ifdef CFG_WITH_PAGER
269 void tee_pager_get_stats(struct tee_pager_stats *stats);
270 bool tee_pager_handle_fault(struct abort_info *ai);
271 #else /*CFG_WITH_PAGER*/
tee_pager_handle_fault(struct abort_info * ai __unused)272 static inline bool tee_pager_handle_fault(struct abort_info *ai __unused)
273 {
274 return false;
275 }
276
tee_pager_get_stats(struct tee_pager_stats * stats)277 static inline void tee_pager_get_stats(struct tee_pager_stats *stats)
278 {
279 memset(stats, 0, sizeof(struct tee_pager_stats));
280 }
281 #endif /*CFG_WITH_PAGER*/
282
283 void tee_pager_invalidate_fobj(struct fobj *fobj);
284
285 #endif /*MM_TEE_PAGER_H*/
286