1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2016 Facebook
3 */
4 #include <linux/cpumask.h>
5 #include <linux/spinlock.h>
6 #include <linux/percpu.h>
7
8 #include "bpf_lru_list.h"
9
10 #define LOCAL_FREE_TARGET (128)
11 #define LOCAL_NR_SCANS LOCAL_FREE_TARGET
12
13 #define PERCPU_FREE_TARGET (4)
14 #define PERCPU_NR_SCANS PERCPU_FREE_TARGET
15
16 /* Helpers to get the local list index */
17 #define LOCAL_LIST_IDX(t) ((t) - BPF_LOCAL_LIST_T_OFFSET)
18 #define LOCAL_FREE_LIST_IDX LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_FREE)
19 #define LOCAL_PENDING_LIST_IDX LOCAL_LIST_IDX(BPF_LRU_LOCAL_LIST_T_PENDING)
20 #define IS_LOCAL_LIST_TYPE(t) ((t) >= BPF_LOCAL_LIST_T_OFFSET)
21
get_next_cpu(int cpu)22 static int get_next_cpu(int cpu)
23 {
24 cpu = cpumask_next(cpu, cpu_possible_mask);
25 if (cpu >= nr_cpu_ids)
26 cpu = cpumask_first(cpu_possible_mask);
27 return cpu;
28 }
29
30 /* Local list helpers */
local_free_list(struct bpf_lru_locallist * loc_l)31 static struct list_head *local_free_list(struct bpf_lru_locallist *loc_l)
32 {
33 return &loc_l->lists[LOCAL_FREE_LIST_IDX];
34 }
35
local_pending_list(struct bpf_lru_locallist * loc_l)36 static struct list_head *local_pending_list(struct bpf_lru_locallist *loc_l)
37 {
38 return &loc_l->lists[LOCAL_PENDING_LIST_IDX];
39 }
40
41 /* bpf_lru_node helpers */
bpf_lru_node_is_ref(const struct bpf_lru_node * node)42 static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node)
43 {
44 return node->ref;
45 }
46
bpf_lru_list_count_inc(struct bpf_lru_list * l,enum bpf_lru_list_type type)47 static void bpf_lru_list_count_inc(struct bpf_lru_list *l,
48 enum bpf_lru_list_type type)
49 {
50 if (type < NR_BPF_LRU_LIST_COUNT)
51 l->counts[type]++;
52 }
53
bpf_lru_list_count_dec(struct bpf_lru_list * l,enum bpf_lru_list_type type)54 static void bpf_lru_list_count_dec(struct bpf_lru_list *l,
55 enum bpf_lru_list_type type)
56 {
57 if (type < NR_BPF_LRU_LIST_COUNT)
58 l->counts[type]--;
59 }
60
__bpf_lru_node_move_to_free(struct bpf_lru_list * l,struct bpf_lru_node * node,struct list_head * free_list,enum bpf_lru_list_type tgt_free_type)61 static void __bpf_lru_node_move_to_free(struct bpf_lru_list *l,
62 struct bpf_lru_node *node,
63 struct list_head *free_list,
64 enum bpf_lru_list_type tgt_free_type)
65 {
66 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
67 return;
68
69 /* If the removing node is the next_inactive_rotation candidate,
70 * move the next_inactive_rotation pointer also.
71 */
72 if (&node->list == l->next_inactive_rotation)
73 l->next_inactive_rotation = l->next_inactive_rotation->prev;
74
75 bpf_lru_list_count_dec(l, node->type);
76
77 node->type = tgt_free_type;
78 list_move(&node->list, free_list);
79 }
80
81 /* Move nodes from local list to the LRU list */
__bpf_lru_node_move_in(struct bpf_lru_list * l,struct bpf_lru_node * node,enum bpf_lru_list_type tgt_type)82 static void __bpf_lru_node_move_in(struct bpf_lru_list *l,
83 struct bpf_lru_node *node,
84 enum bpf_lru_list_type tgt_type)
85 {
86 if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) ||
87 WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
88 return;
89
90 bpf_lru_list_count_inc(l, tgt_type);
91 node->type = tgt_type;
92 node->ref = 0;
93 list_move(&node->list, &l->lists[tgt_type]);
94 }
95
96 /* Move nodes between or within active and inactive list (like
97 * active to inactive, inactive to active or tail of active back to
98 * the head of active).
99 */
__bpf_lru_node_move(struct bpf_lru_list * l,struct bpf_lru_node * node,enum bpf_lru_list_type tgt_type)100 static void __bpf_lru_node_move(struct bpf_lru_list *l,
101 struct bpf_lru_node *node,
102 enum bpf_lru_list_type tgt_type)
103 {
104 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) ||
105 WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(tgt_type)))
106 return;
107
108 if (node->type != tgt_type) {
109 bpf_lru_list_count_dec(l, node->type);
110 bpf_lru_list_count_inc(l, tgt_type);
111 node->type = tgt_type;
112 }
113 node->ref = 0;
114
115 /* If the moving node is the next_inactive_rotation candidate,
116 * move the next_inactive_rotation pointer also.
117 */
118 if (&node->list == l->next_inactive_rotation)
119 l->next_inactive_rotation = l->next_inactive_rotation->prev;
120
121 list_move(&node->list, &l->lists[tgt_type]);
122 }
123
bpf_lru_list_inactive_low(const struct bpf_lru_list * l)124 static bool bpf_lru_list_inactive_low(const struct bpf_lru_list *l)
125 {
126 return l->counts[BPF_LRU_LIST_T_INACTIVE] <
127 l->counts[BPF_LRU_LIST_T_ACTIVE];
128 }
129
130 /* Rotate the active list:
131 * 1. Start from tail
132 * 2. If the node has the ref bit set, it will be rotated
133 * back to the head of active list with the ref bit cleared.
134 * Give this node one more chance to survive in the active list.
135 * 3. If the ref bit is not set, move it to the head of the
136 * inactive list.
137 * 4. It will at most scan nr_scans nodes
138 */
__bpf_lru_list_rotate_active(struct bpf_lru * lru,struct bpf_lru_list * l)139 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru,
140 struct bpf_lru_list *l)
141 {
142 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE];
143 struct bpf_lru_node *node, *tmp_node, *first_node;
144 unsigned int i = 0;
145
146 first_node = list_first_entry(active, struct bpf_lru_node, list);
147 list_for_each_entry_safe_reverse(node, tmp_node, active, list) {
148 if (bpf_lru_node_is_ref(node))
149 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
150 else
151 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
152
153 if (++i == lru->nr_scans || node == first_node)
154 break;
155 }
156 }
157
158 /* Rotate the inactive list. It starts from the next_inactive_rotation
159 * 1. If the node has ref bit set, it will be moved to the head
160 * of active list with the ref bit cleared.
161 * 2. If the node does not have ref bit set, it will leave it
162 * at its current location (i.e. do nothing) so that it can
163 * be considered during the next inactive_shrink.
164 * 3. It will at most scan nr_scans nodes
165 */
__bpf_lru_list_rotate_inactive(struct bpf_lru * lru,struct bpf_lru_list * l)166 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru,
167 struct bpf_lru_list *l)
168 {
169 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
170 struct list_head *cur, *last, *next = inactive;
171 struct bpf_lru_node *node;
172 unsigned int i = 0;
173
174 if (list_empty(inactive))
175 return;
176
177 last = l->next_inactive_rotation->next;
178 if (last == inactive)
179 last = last->next;
180
181 cur = l->next_inactive_rotation;
182 while (i < lru->nr_scans) {
183 if (cur == inactive) {
184 cur = cur->prev;
185 continue;
186 }
187
188 node = list_entry(cur, struct bpf_lru_node, list);
189 next = cur->prev;
190 if (bpf_lru_node_is_ref(node))
191 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
192 if (cur == last)
193 break;
194 cur = next;
195 i++;
196 }
197
198 l->next_inactive_rotation = next;
199 }
200
201 /* Shrink the inactive list. It starts from the tail of the
202 * inactive list and only move the nodes without the ref bit
203 * set to the designated free list.
204 */
205 static unsigned int
__bpf_lru_list_shrink_inactive(struct bpf_lru * lru,struct bpf_lru_list * l,unsigned int tgt_nshrink,struct list_head * free_list,enum bpf_lru_list_type tgt_free_type)206 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru,
207 struct bpf_lru_list *l,
208 unsigned int tgt_nshrink,
209 struct list_head *free_list,
210 enum bpf_lru_list_type tgt_free_type)
211 {
212 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE];
213 struct bpf_lru_node *node, *tmp_node;
214 unsigned int nshrinked = 0;
215 unsigned int i = 0;
216
217 list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) {
218 if (bpf_lru_node_is_ref(node)) {
219 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
220 } else if (lru->del_from_htab(lru->del_arg, node)) {
221 __bpf_lru_node_move_to_free(l, node, free_list,
222 tgt_free_type);
223 if (++nshrinked == tgt_nshrink)
224 break;
225 }
226
227 if (++i == lru->nr_scans)
228 break;
229 }
230
231 return nshrinked;
232 }
233
234 /* 1. Rotate the active list (if needed)
235 * 2. Always rotate the inactive list
236 */
__bpf_lru_list_rotate(struct bpf_lru * lru,struct bpf_lru_list * l)237 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l)
238 {
239 if (bpf_lru_list_inactive_low(l))
240 __bpf_lru_list_rotate_active(lru, l);
241
242 __bpf_lru_list_rotate_inactive(lru, l);
243 }
244
245 /* Calls __bpf_lru_list_shrink_inactive() to shrink some
246 * ref-bit-cleared nodes and move them to the designated
247 * free list.
248 *
249 * If it cannot get a free node after calling
250 * __bpf_lru_list_shrink_inactive(). It will just remove
251 * one node from either inactive or active list without
252 * honoring the ref-bit. It prefers inactive list to active
253 * list in this situation.
254 */
__bpf_lru_list_shrink(struct bpf_lru * lru,struct bpf_lru_list * l,unsigned int tgt_nshrink,struct list_head * free_list,enum bpf_lru_list_type tgt_free_type)255 static unsigned int __bpf_lru_list_shrink(struct bpf_lru *lru,
256 struct bpf_lru_list *l,
257 unsigned int tgt_nshrink,
258 struct list_head *free_list,
259 enum bpf_lru_list_type tgt_free_type)
260
261 {
262 struct bpf_lru_node *node, *tmp_node;
263 struct list_head *force_shrink_list;
264 unsigned int nshrinked;
265
266 nshrinked = __bpf_lru_list_shrink_inactive(lru, l, tgt_nshrink,
267 free_list, tgt_free_type);
268 if (nshrinked)
269 return nshrinked;
270
271 /* Do a force shrink by ignoring the reference bit */
272 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE]))
273 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE];
274 else
275 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIVE];
276
277 list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list,
278 list) {
279 if (lru->del_from_htab(lru->del_arg, node)) {
280 __bpf_lru_node_move_to_free(l, node, free_list,
281 tgt_free_type);
282 return 1;
283 }
284 }
285
286 return 0;
287 }
288
289 /* Flush the nodes from the local pending list to the LRU list */
__local_list_flush(struct bpf_lru_list * l,struct bpf_lru_locallist * loc_l)290 static void __local_list_flush(struct bpf_lru_list *l,
291 struct bpf_lru_locallist *loc_l)
292 {
293 struct bpf_lru_node *node, *tmp_node;
294
295 list_for_each_entry_safe_reverse(node, tmp_node,
296 local_pending_list(loc_l), list) {
297 if (bpf_lru_node_is_ref(node))
298 __bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE);
299 else
300 __bpf_lru_node_move_in(l, node,
301 BPF_LRU_LIST_T_INACTIVE);
302 }
303 }
304
bpf_lru_list_push_free(struct bpf_lru_list * l,struct bpf_lru_node * node)305 static void bpf_lru_list_push_free(struct bpf_lru_list *l,
306 struct bpf_lru_node *node)
307 {
308 unsigned long flags;
309
310 if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
311 return;
312
313 raw_spin_lock_irqsave(&l->lock, flags);
314 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
315 raw_spin_unlock_irqrestore(&l->lock, flags);
316 }
317
bpf_lru_list_pop_free_to_local(struct bpf_lru * lru,struct bpf_lru_locallist * loc_l)318 static void bpf_lru_list_pop_free_to_local(struct bpf_lru *lru,
319 struct bpf_lru_locallist *loc_l)
320 {
321 struct bpf_lru_list *l = &lru->common_lru.lru_list;
322 struct bpf_lru_node *node, *tmp_node;
323 unsigned int nfree = 0;
324
325 raw_spin_lock(&l->lock);
326
327 __local_list_flush(l, loc_l);
328
329 __bpf_lru_list_rotate(lru, l);
330
331 list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE],
332 list) {
333 __bpf_lru_node_move_to_free(l, node, local_free_list(loc_l),
334 BPF_LRU_LOCAL_LIST_T_FREE);
335 if (++nfree == LOCAL_FREE_TARGET)
336 break;
337 }
338
339 if (nfree < LOCAL_FREE_TARGET)
340 __bpf_lru_list_shrink(lru, l, LOCAL_FREE_TARGET - nfree,
341 local_free_list(loc_l),
342 BPF_LRU_LOCAL_LIST_T_FREE);
343
344 raw_spin_unlock(&l->lock);
345 }
346
__local_list_add_pending(struct bpf_lru * lru,struct bpf_lru_locallist * loc_l,int cpu,struct bpf_lru_node * node,u32 hash)347 static void __local_list_add_pending(struct bpf_lru *lru,
348 struct bpf_lru_locallist *loc_l,
349 int cpu,
350 struct bpf_lru_node *node,
351 u32 hash)
352 {
353 *(u32 *)((void *)node + lru->hash_offset) = hash;
354 node->cpu = cpu;
355 node->type = BPF_LRU_LOCAL_LIST_T_PENDING;
356 node->ref = 0;
357 list_add(&node->list, local_pending_list(loc_l));
358 }
359
360 static struct bpf_lru_node *
__local_list_pop_free(struct bpf_lru_locallist * loc_l)361 __local_list_pop_free(struct bpf_lru_locallist *loc_l)
362 {
363 struct bpf_lru_node *node;
364
365 node = list_first_entry_or_null(local_free_list(loc_l),
366 struct bpf_lru_node,
367 list);
368 if (node)
369 list_del(&node->list);
370
371 return node;
372 }
373
374 static struct bpf_lru_node *
__local_list_pop_pending(struct bpf_lru * lru,struct bpf_lru_locallist * loc_l)375 __local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l)
376 {
377 struct bpf_lru_node *node;
378 bool force = false;
379
380 ignore_ref:
381 /* Get from the tail (i.e. older element) of the pending list. */
382 list_for_each_entry_reverse(node, local_pending_list(loc_l),
383 list) {
384 if ((!bpf_lru_node_is_ref(node) || force) &&
385 lru->del_from_htab(lru->del_arg, node)) {
386 list_del(&node->list);
387 return node;
388 }
389 }
390
391 if (!force) {
392 force = true;
393 goto ignore_ref;
394 }
395
396 return NULL;
397 }
398
bpf_percpu_lru_pop_free(struct bpf_lru * lru,u32 hash)399 static struct bpf_lru_node *bpf_percpu_lru_pop_free(struct bpf_lru *lru,
400 u32 hash)
401 {
402 struct list_head *free_list;
403 struct bpf_lru_node *node = NULL;
404 struct bpf_lru_list *l;
405 unsigned long flags;
406 int cpu = raw_smp_processor_id();
407
408 l = per_cpu_ptr(lru->percpu_lru, cpu);
409
410 raw_spin_lock_irqsave(&l->lock, flags);
411
412 __bpf_lru_list_rotate(lru, l);
413
414 free_list = &l->lists[BPF_LRU_LIST_T_FREE];
415 if (list_empty(free_list))
416 __bpf_lru_list_shrink(lru, l, PERCPU_FREE_TARGET, free_list,
417 BPF_LRU_LIST_T_FREE);
418
419 if (!list_empty(free_list)) {
420 node = list_first_entry(free_list, struct bpf_lru_node, list);
421 *(u32 *)((void *)node + lru->hash_offset) = hash;
422 node->ref = 0;
423 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
424 }
425
426 raw_spin_unlock_irqrestore(&l->lock, flags);
427
428 return node;
429 }
430
bpf_common_lru_pop_free(struct bpf_lru * lru,u32 hash)431 static struct bpf_lru_node *bpf_common_lru_pop_free(struct bpf_lru *lru,
432 u32 hash)
433 {
434 struct bpf_lru_locallist *loc_l, *steal_loc_l;
435 struct bpf_common_lru *clru = &lru->common_lru;
436 struct bpf_lru_node *node;
437 int steal, first_steal;
438 unsigned long flags;
439 int cpu = raw_smp_processor_id();
440
441 loc_l = per_cpu_ptr(clru->local_list, cpu);
442
443 raw_spin_lock_irqsave(&loc_l->lock, flags);
444
445 node = __local_list_pop_free(loc_l);
446 if (!node) {
447 bpf_lru_list_pop_free_to_local(lru, loc_l);
448 node = __local_list_pop_free(loc_l);
449 }
450
451 if (node)
452 __local_list_add_pending(lru, loc_l, cpu, node, hash);
453
454 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
455
456 if (node)
457 return node;
458
459 /* No free nodes found from the local free list and
460 * the global LRU list.
461 *
462 * Steal from the local free/pending list of the
463 * current CPU and remote CPU in RR. It starts
464 * with the loc_l->next_steal CPU.
465 */
466
467 first_steal = loc_l->next_steal;
468 steal = first_steal;
469 do {
470 steal_loc_l = per_cpu_ptr(clru->local_list, steal);
471
472 raw_spin_lock_irqsave(&steal_loc_l->lock, flags);
473
474 node = __local_list_pop_free(steal_loc_l);
475 if (!node)
476 node = __local_list_pop_pending(lru, steal_loc_l);
477
478 raw_spin_unlock_irqrestore(&steal_loc_l->lock, flags);
479
480 steal = get_next_cpu(steal);
481 } while (!node && steal != first_steal);
482
483 loc_l->next_steal = steal;
484
485 if (node) {
486 raw_spin_lock_irqsave(&loc_l->lock, flags);
487 __local_list_add_pending(lru, loc_l, cpu, node, hash);
488 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
489 }
490
491 return node;
492 }
493
bpf_lru_pop_free(struct bpf_lru * lru,u32 hash)494 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash)
495 {
496 if (lru->percpu)
497 return bpf_percpu_lru_pop_free(lru, hash);
498 else
499 return bpf_common_lru_pop_free(lru, hash);
500 }
501
bpf_common_lru_push_free(struct bpf_lru * lru,struct bpf_lru_node * node)502 static void bpf_common_lru_push_free(struct bpf_lru *lru,
503 struct bpf_lru_node *node)
504 {
505 u8 node_type = READ_ONCE(node->type);
506 unsigned long flags;
507
508 if (WARN_ON_ONCE(node_type == BPF_LRU_LIST_T_FREE) ||
509 WARN_ON_ONCE(node_type == BPF_LRU_LOCAL_LIST_T_FREE))
510 return;
511
512 if (node_type == BPF_LRU_LOCAL_LIST_T_PENDING) {
513 struct bpf_lru_locallist *loc_l;
514
515 loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
516
517 raw_spin_lock_irqsave(&loc_l->lock, flags);
518
519 if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) {
520 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
521 goto check_lru_list;
522 }
523
524 node->type = BPF_LRU_LOCAL_LIST_T_FREE;
525 node->ref = 0;
526 list_move(&node->list, local_free_list(loc_l));
527
528 raw_spin_unlock_irqrestore(&loc_l->lock, flags);
529 return;
530 }
531
532 check_lru_list:
533 bpf_lru_list_push_free(&lru->common_lru.lru_list, node);
534 }
535
bpf_percpu_lru_push_free(struct bpf_lru * lru,struct bpf_lru_node * node)536 static void bpf_percpu_lru_push_free(struct bpf_lru *lru,
537 struct bpf_lru_node *node)
538 {
539 struct bpf_lru_list *l;
540 unsigned long flags;
541
542 l = per_cpu_ptr(lru->percpu_lru, node->cpu);
543
544 raw_spin_lock_irqsave(&l->lock, flags);
545
546 __bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
547
548 raw_spin_unlock_irqrestore(&l->lock, flags);
549 }
550
bpf_lru_push_free(struct bpf_lru * lru,struct bpf_lru_node * node)551 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node)
552 {
553 if (lru->percpu)
554 bpf_percpu_lru_push_free(lru, node);
555 else
556 bpf_common_lru_push_free(lru, node);
557 }
558
bpf_common_lru_populate(struct bpf_lru * lru,void * buf,u32 node_offset,u32 elem_size,u32 nr_elems)559 static void bpf_common_lru_populate(struct bpf_lru *lru, void *buf,
560 u32 node_offset, u32 elem_size,
561 u32 nr_elems)
562 {
563 struct bpf_lru_list *l = &lru->common_lru.lru_list;
564 u32 i;
565
566 for (i = 0; i < nr_elems; i++) {
567 struct bpf_lru_node *node;
568
569 node = (struct bpf_lru_node *)(buf + node_offset);
570 node->type = BPF_LRU_LIST_T_FREE;
571 node->ref = 0;
572 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
573 buf += elem_size;
574 }
575 }
576
bpf_percpu_lru_populate(struct bpf_lru * lru,void * buf,u32 node_offset,u32 elem_size,u32 nr_elems)577 static void bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf,
578 u32 node_offset, u32 elem_size,
579 u32 nr_elems)
580 {
581 u32 i, pcpu_entries;
582 int cpu;
583 struct bpf_lru_list *l;
584
585 pcpu_entries = nr_elems / num_possible_cpus();
586
587 i = 0;
588
589 for_each_possible_cpu(cpu) {
590 struct bpf_lru_node *node;
591
592 l = per_cpu_ptr(lru->percpu_lru, cpu);
593 again:
594 node = (struct bpf_lru_node *)(buf + node_offset);
595 node->cpu = cpu;
596 node->type = BPF_LRU_LIST_T_FREE;
597 node->ref = 0;
598 list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
599 i++;
600 buf += elem_size;
601 if (i == nr_elems)
602 break;
603 if (i % pcpu_entries)
604 goto again;
605 }
606 }
607
bpf_lru_populate(struct bpf_lru * lru,void * buf,u32 node_offset,u32 elem_size,u32 nr_elems)608 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset,
609 u32 elem_size, u32 nr_elems)
610 {
611 if (lru->percpu)
612 bpf_percpu_lru_populate(lru, buf, node_offset, elem_size,
613 nr_elems);
614 else
615 bpf_common_lru_populate(lru, buf, node_offset, elem_size,
616 nr_elems);
617 }
618
bpf_lru_locallist_init(struct bpf_lru_locallist * loc_l,int cpu)619 static void bpf_lru_locallist_init(struct bpf_lru_locallist *loc_l, int cpu)
620 {
621 int i;
622
623 for (i = 0; i < NR_BPF_LRU_LOCAL_LIST_T; i++)
624 INIT_LIST_HEAD(&loc_l->lists[i]);
625
626 loc_l->next_steal = cpu;
627
628 raw_spin_lock_init(&loc_l->lock);
629 }
630
bpf_lru_list_init(struct bpf_lru_list * l)631 static void bpf_lru_list_init(struct bpf_lru_list *l)
632 {
633 int i;
634
635 for (i = 0; i < NR_BPF_LRU_LIST_T; i++)
636 INIT_LIST_HEAD(&l->lists[i]);
637
638 for (i = 0; i < NR_BPF_LRU_LIST_COUNT; i++)
639 l->counts[i] = 0;
640
641 l->next_inactive_rotation = &l->lists[BPF_LRU_LIST_T_INACTIVE];
642
643 raw_spin_lock_init(&l->lock);
644 }
645
bpf_lru_init(struct bpf_lru * lru,bool percpu,u32 hash_offset,del_from_htab_func del_from_htab,void * del_arg)646 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
647 del_from_htab_func del_from_htab, void *del_arg)
648 {
649 int cpu;
650
651 if (percpu) {
652 lru->percpu_lru = alloc_percpu(struct bpf_lru_list);
653 if (!lru->percpu_lru)
654 return -ENOMEM;
655
656 for_each_possible_cpu(cpu) {
657 struct bpf_lru_list *l;
658
659 l = per_cpu_ptr(lru->percpu_lru, cpu);
660 bpf_lru_list_init(l);
661 }
662 lru->nr_scans = PERCPU_NR_SCANS;
663 } else {
664 struct bpf_common_lru *clru = &lru->common_lru;
665
666 clru->local_list = alloc_percpu(struct bpf_lru_locallist);
667 if (!clru->local_list)
668 return -ENOMEM;
669
670 for_each_possible_cpu(cpu) {
671 struct bpf_lru_locallist *loc_l;
672
673 loc_l = per_cpu_ptr(clru->local_list, cpu);
674 bpf_lru_locallist_init(loc_l, cpu);
675 }
676
677 bpf_lru_list_init(&clru->lru_list);
678 lru->nr_scans = LOCAL_NR_SCANS;
679 }
680
681 lru->percpu = percpu;
682 lru->del_from_htab = del_from_htab;
683 lru->del_arg = del_arg;
684 lru->hash_offset = hash_offset;
685
686 return 0;
687 }
688
bpf_lru_destroy(struct bpf_lru * lru)689 void bpf_lru_destroy(struct bpf_lru *lru)
690 {
691 if (lru->percpu)
692 free_percpu(lru->percpu_lru);
693 else
694 free_percpu(lru->common_lru.local_list);
695 }
696