Lines Matching refs:nc

242 	struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache);  in __napi_alloc_frag_align()  local
248 data = __page_frag_alloc_align(&nc->page, fragsz, in __napi_alloc_frag_align()
261 struct page_frag_cache *nc = this_cpu_ptr(&netdev_alloc_cache); in __netdev_alloc_frag_align() local
264 data = __page_frag_alloc_align(nc, fragsz, in __netdev_alloc_frag_align()
278 struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache); in napi_skb_cache_get() local
282 if (unlikely(!nc->skb_count)) { in napi_skb_cache_get()
283 nc->skb_count = kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, in napi_skb_cache_get()
286 nc->skb_cache); in napi_skb_cache_get()
287 if (unlikely(!nc->skb_count)) { in napi_skb_cache_get()
293 skb = nc->skb_cache[--nc->skb_count]; in napi_skb_cache_get()
318 struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache); in napi_skb_cache_get_bulk() local
323 if (nc->skb_count >= n) in napi_skb_cache_get_bulk()
327 bulk = min(NAPI_SKB_CACHE_SIZE - nc->skb_count, NAPI_SKB_CACHE_BULK); in napi_skb_cache_get_bulk()
328 nc->skb_count += kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, in napi_skb_cache_get_bulk()
330 &nc->skb_cache[nc->skb_count]); in napi_skb_cache_get_bulk()
331 if (likely(nc->skb_count >= n)) in napi_skb_cache_get_bulk()
337 n - nc->skb_count, &skbs[nc->skb_count]); in napi_skb_cache_get_bulk()
338 if (likely(nc->skb_count >= n)) in napi_skb_cache_get_bulk()
342 total -= n - nc->skb_count; in napi_skb_cache_get_bulk()
343 n = nc->skb_count; in napi_skb_cache_get_bulk()
346 for (u32 base = nc->skb_count - n, i = 0; i < n; i++) { in napi_skb_cache_get_bulk()
349 skbs[i] = nc->skb_cache[base + i]; in napi_skb_cache_get_bulk()
355 nc->skb_count -= n; in napi_skb_cache_get_bulk()
720 struct page_frag_cache *nc; in __netdev_alloc_skb() local
745 nc = this_cpu_ptr(&netdev_alloc_cache); in __netdev_alloc_skb()
746 data = page_frag_alloc(nc, len, gfp_mask); in __netdev_alloc_skb()
747 pfmemalloc = page_frag_cache_is_pfmemalloc(nc); in __netdev_alloc_skb()
752 nc = this_cpu_ptr(&napi_alloc_cache.page); in __netdev_alloc_skb()
753 data = page_frag_alloc(nc, len, gfp_mask); in __netdev_alloc_skb()
754 pfmemalloc = page_frag_cache_is_pfmemalloc(nc); in __netdev_alloc_skb()
797 struct napi_alloc_cache *nc; in napi_alloc_skb() local
824 nc = this_cpu_ptr(&napi_alloc_cache); in napi_alloc_skb()
826 data = page_frag_alloc(&nc->page, len, gfp_mask); in napi_alloc_skb()
827 pfmemalloc = page_frag_cache_is_pfmemalloc(&nc->page); in napi_alloc_skb()
1418 struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache); in napi_skb_cache_put() local
1425 nc->skb_cache[nc->skb_count++] = skb; in napi_skb_cache_put()
1427 if (unlikely(nc->skb_count == NAPI_SKB_CACHE_SIZE)) { in napi_skb_cache_put()
1429 kasan_mempool_unpoison_object(nc->skb_cache[i], in napi_skb_cache_put()
1433 nc->skb_cache + NAPI_SKB_CACHE_HALF); in napi_skb_cache_put()
1434 nc->skb_count = NAPI_SKB_CACHE_HALF; in napi_skb_cache_put()