Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
  1// SPDX-License-Identifier: GPL-2.0-only
  2/* Page fragment allocator
  3 *
  4 * Page Fragment:
  5 *  An arbitrary-length arbitrary-offset area of memory which resides within a
  6 *  0 or higher order page.  Multiple fragments within that page are
  7 *  individually refcounted, in the page's reference counter.
  8 *
  9 * The page_frag functions provide a simple allocation framework for page
 10 * fragments.  This is used by the network stack and network device drivers to
 11 * provide a backing region of memory for use as either an sk_buff->head, or to
 12 * be used in the "frags" portion of skb_shared_info.
 13 */
 14
 15#include <linux/build_bug.h>
 16#include <linux/export.h>
 17#include <linux/gfp_types.h>
 18#include <linux/init.h>
 19#include <linux/mm.h>
 20#include <linux/page_frag_cache.h>
 21#include "internal.h"
 22
 23static unsigned long encoded_page_create(struct page *page, unsigned int order,
 24					 bool pfmemalloc)
 25{
 26	BUILD_BUG_ON(PAGE_FRAG_CACHE_MAX_ORDER > PAGE_FRAG_CACHE_ORDER_MASK);
 27	BUILD_BUG_ON(PAGE_FRAG_CACHE_PFMEMALLOC_BIT >= PAGE_SIZE);
 28
 29	return (unsigned long)page_address(page) |
 30		(order & PAGE_FRAG_CACHE_ORDER_MASK) |
 31		((unsigned long)pfmemalloc * PAGE_FRAG_CACHE_PFMEMALLOC_BIT);
 32}
 33
 34static unsigned long encoded_page_decode_order(unsigned long encoded_page)
 35{
 36	return encoded_page & PAGE_FRAG_CACHE_ORDER_MASK;
 37}
 38
 39static void *encoded_page_decode_virt(unsigned long encoded_page)
 40{
 41	return (void *)(encoded_page & PAGE_MASK);
 42}
 43
 44static struct page *encoded_page_decode_page(unsigned long encoded_page)
 45{
 46	return virt_to_page((void *)encoded_page);
 47}
 48
 49static struct page *__page_frag_cache_refill(struct page_frag_cache *nc,
 50					     gfp_t gfp_mask)
 51{
 52	unsigned long order = PAGE_FRAG_CACHE_MAX_ORDER;
 53	struct page *page = NULL;
 54	gfp_t gfp = gfp_mask;
 55
 56#if (PAGE_SIZE < PAGE_FRAG_CACHE_MAX_SIZE)
 57	gfp_mask = (gfp_mask & ~__GFP_DIRECT_RECLAIM) |  __GFP_COMP |
 58		   __GFP_NOWARN | __GFP_NORETRY | __GFP_NOMEMALLOC;
 59	page = __alloc_pages(gfp_mask, PAGE_FRAG_CACHE_MAX_ORDER,
 60			     numa_mem_id(), NULL);
 61#endif
 62	if (unlikely(!page)) {
 63		page = __alloc_pages(gfp, 0, numa_mem_id(), NULL);
 64		order = 0;
 65	}
 66
 67	nc->encoded_page = page ?
 68		encoded_page_create(page, order, page_is_pfmemalloc(page)) : 0;
 69
 70	return page;
 71}
 72
 73void page_frag_cache_drain(struct page_frag_cache *nc)
 74{
 75	if (!nc->encoded_page)
 76		return;
 77
 78	__page_frag_cache_drain(encoded_page_decode_page(nc->encoded_page),
 79				nc->pagecnt_bias);
 80	nc->encoded_page = 0;
 81}
 82EXPORT_SYMBOL(page_frag_cache_drain);
 83
 84void __page_frag_cache_drain(struct page *page, unsigned int count)
 85{
 86	VM_BUG_ON_PAGE(page_ref_count(page) == 0, page);
 87
 88	if (page_ref_sub_and_test(page, count))
 89		free_unref_page(page, compound_order(page));
 90}
 91EXPORT_SYMBOL(__page_frag_cache_drain);
 92
 93void *__page_frag_alloc_align(struct page_frag_cache *nc,
 94			      unsigned int fragsz, gfp_t gfp_mask,
 95			      unsigned int align_mask)
 96{
 97	unsigned long encoded_page = nc->encoded_page;
 98	unsigned int size, offset;
 99	struct page *page;
100
101	if (unlikely(!encoded_page)) {
102refill:
103		page = __page_frag_cache_refill(nc, gfp_mask);
104		if (!page)
105			return NULL;
106
107		encoded_page = nc->encoded_page;
108
109		/* Even if we own the page, we do not use atomic_set().
110		 * This would break get_page_unless_zero() users.
111		 */
112		page_ref_add(page, PAGE_FRAG_CACHE_MAX_SIZE);
113
114		/* reset page count bias and offset to start of new frag */
115		nc->pagecnt_bias = PAGE_FRAG_CACHE_MAX_SIZE + 1;
116		nc->offset = 0;
117	}
118
119	size = PAGE_SIZE << encoded_page_decode_order(encoded_page);
120	offset = __ALIGN_KERNEL_MASK(nc->offset, ~align_mask);
121	if (unlikely(offset + fragsz > size)) {
122		if (unlikely(fragsz > PAGE_SIZE)) {
123			/*
124			 * The caller is trying to allocate a fragment
125			 * with fragsz > PAGE_SIZE but the cache isn't big
126			 * enough to satisfy the request, this may
127			 * happen in low memory conditions.
128			 * We don't release the cache page because
129			 * it could make memory pressure worse
130			 * so we simply return NULL here.
131			 */
132			return NULL;
133		}
134
135		page = encoded_page_decode_page(encoded_page);
136
137		if (!page_ref_sub_and_test(page, nc->pagecnt_bias))
138			goto refill;
139
140		if (unlikely(encoded_page_decode_pfmemalloc(encoded_page))) {
141			free_unref_page(page,
142					encoded_page_decode_order(encoded_page));
143			goto refill;
144		}
145
146		/* OK, page count is 0, we can safely set it */
147		set_page_count(page, PAGE_FRAG_CACHE_MAX_SIZE + 1);
148
149		/* reset page count bias and offset to start of new frag */
150		nc->pagecnt_bias = PAGE_FRAG_CACHE_MAX_SIZE + 1;
151		offset = 0;
152	}
153
154	nc->pagecnt_bias--;
155	nc->offset = offset + fragsz;
156
157	return encoded_page_decode_virt(encoded_page) + offset;
158}
159EXPORT_SYMBOL(__page_frag_alloc_align);
160
161/*
162 * Frees a page fragment allocated out of either a compound or order 0 page.
163 */
164void page_frag_free(void *addr)
165{
166	struct page *page = virt_to_head_page(addr);
167
168	if (unlikely(put_page_testzero(page)))
169		free_unref_page(page, compound_order(page));
170}
171EXPORT_SYMBOL(page_frag_free);