Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
4 * All Rights Reserved.
5 */
6#include "xfs.h"
7#include <linux/backing-dev.h>
8#include "xfs_message.h"
9#include "xfs_trace.h"
10
11void *
12kmem_alloc(size_t size, xfs_km_flags_t flags)
13{
14 int retries = 0;
15 gfp_t lflags = kmem_flags_convert(flags);
16 void *ptr;
17
18 trace_kmem_alloc(size, flags, _RET_IP_);
19
20 do {
21 ptr = kmalloc(size, lflags);
22 if (ptr || (flags & KM_MAYFAIL))
23 return ptr;
24 if (!(++retries % 100))
25 xfs_err(NULL,
26 "%s(%u) possible memory allocation deadlock size %u in %s (mode:0x%x)",
27 current->comm, current->pid,
28 (unsigned int)size, __func__, lflags);
29 congestion_wait(BLK_RW_ASYNC, HZ/50);
30 } while (1);
31}
32
33
34/*
35 * __vmalloc() will allocate data pages and auxiliary structures (e.g.
36 * pagetables) with GFP_KERNEL, yet we may be under GFP_NOFS context here. Hence
37 * we need to tell memory reclaim that we are in such a context via
38 * PF_MEMALLOC_NOFS to prevent memory reclaim re-entering the filesystem here
39 * and potentially deadlocking.
40 */
41static void *
42__kmem_vmalloc(size_t size, xfs_km_flags_t flags)
43{
44 unsigned nofs_flag = 0;
45 void *ptr;
46 gfp_t lflags = kmem_flags_convert(flags);
47
48 if (flags & KM_NOFS)
49 nofs_flag = memalloc_nofs_save();
50
51 ptr = __vmalloc(size, lflags);
52
53 if (flags & KM_NOFS)
54 memalloc_nofs_restore(nofs_flag);
55
56 return ptr;
57}
58
59/*
60 * Same as kmem_alloc_large, except we guarantee the buffer returned is aligned
61 * to the @align_mask. We only guarantee alignment up to page size, we'll clamp
62 * alignment at page size if it is larger. vmalloc always returns a PAGE_SIZE
63 * aligned region.
64 */
65void *
66kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags)
67{
68 void *ptr;
69
70 trace_kmem_alloc_io(size, flags, _RET_IP_);
71
72 if (WARN_ON_ONCE(align_mask >= PAGE_SIZE))
73 align_mask = PAGE_SIZE - 1;
74
75 ptr = kmem_alloc(size, flags | KM_MAYFAIL);
76 if (ptr) {
77 if (!((uintptr_t)ptr & align_mask))
78 return ptr;
79 kfree(ptr);
80 }
81 return __kmem_vmalloc(size, flags);
82}
83
84void *
85kmem_alloc_large(size_t size, xfs_km_flags_t flags)
86{
87 void *ptr;
88
89 trace_kmem_alloc_large(size, flags, _RET_IP_);
90
91 ptr = kmem_alloc(size, flags | KM_MAYFAIL);
92 if (ptr)
93 return ptr;
94 return __kmem_vmalloc(size, flags);
95}
1/*
2 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
3 * All Rights Reserved.
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
17 */
18#include <linux/mm.h>
19#include <linux/highmem.h>
20#include <linux/slab.h>
21#include <linux/swap.h>
22#include <linux/blkdev.h>
23#include <linux/backing-dev.h>
24#include "time.h"
25#include "kmem.h"
26#include "xfs_message.h"
27
28/*
29 * Greedy allocation. May fail and may return vmalloced memory.
30 *
31 * Must be freed using kmem_free_large.
32 */
33void *
34kmem_zalloc_greedy(size_t *size, size_t minsize, size_t maxsize)
35{
36 void *ptr;
37 size_t kmsize = maxsize;
38
39 while (!(ptr = kmem_zalloc_large(kmsize))) {
40 if ((kmsize >>= 1) <= minsize)
41 kmsize = minsize;
42 }
43 if (ptr)
44 *size = kmsize;
45 return ptr;
46}
47
48void *
49kmem_alloc(size_t size, xfs_km_flags_t flags)
50{
51 int retries = 0;
52 gfp_t lflags = kmem_flags_convert(flags);
53 void *ptr;
54
55 do {
56 ptr = kmalloc(size, lflags);
57 if (ptr || (flags & (KM_MAYFAIL|KM_NOSLEEP)))
58 return ptr;
59 if (!(++retries % 100))
60 xfs_err(NULL,
61 "possible memory allocation deadlock in %s (mode:0x%x)",
62 __func__, lflags);
63 congestion_wait(BLK_RW_ASYNC, HZ/50);
64 } while (1);
65}
66
67void *
68kmem_zalloc(size_t size, xfs_km_flags_t flags)
69{
70 void *ptr;
71
72 ptr = kmem_alloc(size, flags);
73 if (ptr)
74 memset((char *)ptr, 0, (int)size);
75 return ptr;
76}
77
78void
79kmem_free(const void *ptr)
80{
81 if (!is_vmalloc_addr(ptr)) {
82 kfree(ptr);
83 } else {
84 vfree(ptr);
85 }
86}
87
88void *
89kmem_realloc(const void *ptr, size_t newsize, size_t oldsize,
90 xfs_km_flags_t flags)
91{
92 void *new;
93
94 new = kmem_alloc(newsize, flags);
95 if (ptr) {
96 if (new)
97 memcpy(new, ptr,
98 ((oldsize < newsize) ? oldsize : newsize));
99 kmem_free(ptr);
100 }
101 return new;
102}
103
104void *
105kmem_zone_alloc(kmem_zone_t *zone, xfs_km_flags_t flags)
106{
107 int retries = 0;
108 gfp_t lflags = kmem_flags_convert(flags);
109 void *ptr;
110
111 do {
112 ptr = kmem_cache_alloc(zone, lflags);
113 if (ptr || (flags & (KM_MAYFAIL|KM_NOSLEEP)))
114 return ptr;
115 if (!(++retries % 100))
116 xfs_err(NULL,
117 "possible memory allocation deadlock in %s (mode:0x%x)",
118 __func__, lflags);
119 congestion_wait(BLK_RW_ASYNC, HZ/50);
120 } while (1);
121}
122
123void *
124kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags)
125{
126 void *ptr;
127
128 ptr = kmem_zone_alloc(zone, flags);
129 if (ptr)
130 memset((char *)ptr, 0, kmem_cache_size(zone));
131 return ptr;
132}