Linux Audio

Check our new training course

Loading...
v3.1
  1/*
  2 * JFFS2 -- Journalling Flash File System, Version 2.
  3 *
  4 * Copyright © 2001-2007 Red Hat, Inc.
  5 *
  6 * Created by David Woodhouse <dwmw2@infradead.org>
  7 *
  8 * For licensing information, see the file 'LICENCE' in this directory.
  9 *
 10 */
 11
 
 
 12#include <linux/kernel.h>
 13#include <linux/slab.h>
 14#include <linux/init.h>
 15#include <linux/jffs2.h>
 16#include "nodelist.h"
 17
 18/* These are initialised to NULL in the kernel startup code.
 19   If you're porting to other operating systems, beware */
 20static struct kmem_cache *full_dnode_slab;
 21static struct kmem_cache *raw_dirent_slab;
 22static struct kmem_cache *raw_inode_slab;
 23static struct kmem_cache *tmp_dnode_info_slab;
 24static struct kmem_cache *raw_node_ref_slab;
 25static struct kmem_cache *node_frag_slab;
 26static struct kmem_cache *inode_cache_slab;
 27#ifdef CONFIG_JFFS2_FS_XATTR
 28static struct kmem_cache *xattr_datum_cache;
 29static struct kmem_cache *xattr_ref_cache;
 30#endif
 31
 32int __init jffs2_create_slab_caches(void)
 33{
 34	full_dnode_slab = kmem_cache_create("jffs2_full_dnode",
 35					    sizeof(struct jffs2_full_dnode),
 36					    0, 0, NULL);
 37	if (!full_dnode_slab)
 38		goto err;
 39
 40	raw_dirent_slab = kmem_cache_create("jffs2_raw_dirent",
 41					    sizeof(struct jffs2_raw_dirent),
 42					    0, SLAB_HWCACHE_ALIGN, NULL);
 43	if (!raw_dirent_slab)
 44		goto err;
 45
 46	raw_inode_slab = kmem_cache_create("jffs2_raw_inode",
 47					   sizeof(struct jffs2_raw_inode),
 48					   0, SLAB_HWCACHE_ALIGN, NULL);
 49	if (!raw_inode_slab)
 50		goto err;
 51
 52	tmp_dnode_info_slab = kmem_cache_create("jffs2_tmp_dnode",
 53						sizeof(struct jffs2_tmp_dnode_info),
 54						0, 0, NULL);
 55	if (!tmp_dnode_info_slab)
 56		goto err;
 57
 58	raw_node_ref_slab = kmem_cache_create("jffs2_refblock",
 59					      sizeof(struct jffs2_raw_node_ref) * (REFS_PER_BLOCK + 1),
 60					      0, 0, NULL);
 61	if (!raw_node_ref_slab)
 62		goto err;
 63
 64	node_frag_slab = kmem_cache_create("jffs2_node_frag",
 65					   sizeof(struct jffs2_node_frag),
 66					   0, 0, NULL);
 67	if (!node_frag_slab)
 68		goto err;
 69
 70	inode_cache_slab = kmem_cache_create("jffs2_inode_cache",
 71					     sizeof(struct jffs2_inode_cache),
 72					     0, 0, NULL);
 73	if (!inode_cache_slab)
 74		goto err;
 75
 76#ifdef CONFIG_JFFS2_FS_XATTR
 77	xattr_datum_cache = kmem_cache_create("jffs2_xattr_datum",
 78					     sizeof(struct jffs2_xattr_datum),
 79					     0, 0, NULL);
 80	if (!xattr_datum_cache)
 81		goto err;
 82
 83	xattr_ref_cache = kmem_cache_create("jffs2_xattr_ref",
 84					   sizeof(struct jffs2_xattr_ref),
 85					   0, 0, NULL);
 86	if (!xattr_ref_cache)
 87		goto err;
 88#endif
 89
 90	return 0;
 91 err:
 92	jffs2_destroy_slab_caches();
 93	return -ENOMEM;
 94}
 95
 96void jffs2_destroy_slab_caches(void)
 97{
 98	if(full_dnode_slab)
 99		kmem_cache_destroy(full_dnode_slab);
100	if(raw_dirent_slab)
101		kmem_cache_destroy(raw_dirent_slab);
102	if(raw_inode_slab)
103		kmem_cache_destroy(raw_inode_slab);
104	if(tmp_dnode_info_slab)
105		kmem_cache_destroy(tmp_dnode_info_slab);
106	if(raw_node_ref_slab)
107		kmem_cache_destroy(raw_node_ref_slab);
108	if(node_frag_slab)
109		kmem_cache_destroy(node_frag_slab);
110	if(inode_cache_slab)
111		kmem_cache_destroy(inode_cache_slab);
112#ifdef CONFIG_JFFS2_FS_XATTR
113	if (xattr_datum_cache)
114		kmem_cache_destroy(xattr_datum_cache);
115	if (xattr_ref_cache)
116		kmem_cache_destroy(xattr_ref_cache);
117#endif
118}
119
120struct jffs2_full_dirent *jffs2_alloc_full_dirent(int namesize)
121{
122	struct jffs2_full_dirent *ret;
123	ret = kmalloc(sizeof(struct jffs2_full_dirent) + namesize, GFP_KERNEL);
124	dbg_memalloc("%p\n", ret);
125	return ret;
126}
127
128void jffs2_free_full_dirent(struct jffs2_full_dirent *x)
129{
130	dbg_memalloc("%p\n", x);
131	kfree(x);
132}
133
134struct jffs2_full_dnode *jffs2_alloc_full_dnode(void)
135{
136	struct jffs2_full_dnode *ret;
137	ret = kmem_cache_alloc(full_dnode_slab, GFP_KERNEL);
138	dbg_memalloc("%p\n", ret);
139	return ret;
140}
141
142void jffs2_free_full_dnode(struct jffs2_full_dnode *x)
143{
144	dbg_memalloc("%p\n", x);
145	kmem_cache_free(full_dnode_slab, x);
146}
147
148struct jffs2_raw_dirent *jffs2_alloc_raw_dirent(void)
149{
150	struct jffs2_raw_dirent *ret;
151	ret = kmem_cache_alloc(raw_dirent_slab, GFP_KERNEL);
152	dbg_memalloc("%p\n", ret);
153	return ret;
154}
155
156void jffs2_free_raw_dirent(struct jffs2_raw_dirent *x)
157{
158	dbg_memalloc("%p\n", x);
159	kmem_cache_free(raw_dirent_slab, x);
160}
161
162struct jffs2_raw_inode *jffs2_alloc_raw_inode(void)
163{
164	struct jffs2_raw_inode *ret;
165	ret = kmem_cache_alloc(raw_inode_slab, GFP_KERNEL);
166	dbg_memalloc("%p\n", ret);
167	return ret;
168}
169
170void jffs2_free_raw_inode(struct jffs2_raw_inode *x)
171{
172	dbg_memalloc("%p\n", x);
173	kmem_cache_free(raw_inode_slab, x);
174}
175
176struct jffs2_tmp_dnode_info *jffs2_alloc_tmp_dnode_info(void)
177{
178	struct jffs2_tmp_dnode_info *ret;
179	ret = kmem_cache_alloc(tmp_dnode_info_slab, GFP_KERNEL);
180	dbg_memalloc("%p\n",
181		ret);
182	return ret;
183}
184
185void jffs2_free_tmp_dnode_info(struct jffs2_tmp_dnode_info *x)
186{
187	dbg_memalloc("%p\n", x);
188	kmem_cache_free(tmp_dnode_info_slab, x);
189}
190
191static struct jffs2_raw_node_ref *jffs2_alloc_refblock(void)
192{
193	struct jffs2_raw_node_ref *ret;
194
195	ret = kmem_cache_alloc(raw_node_ref_slab, GFP_KERNEL);
196	if (ret) {
197		int i = 0;
198		for (i=0; i < REFS_PER_BLOCK; i++) {
199			ret[i].flash_offset = REF_EMPTY_NODE;
200			ret[i].next_in_ino = NULL;
201		}
202		ret[i].flash_offset = REF_LINK_NODE;
203		ret[i].next_in_ino = NULL;
204	}
205	return ret;
206}
207
208int jffs2_prealloc_raw_node_refs(struct jffs2_sb_info *c,
209				 struct jffs2_eraseblock *jeb, int nr)
210{
211	struct jffs2_raw_node_ref **p, *ref;
212	int i = nr;
213
214	dbg_memalloc("%d\n", nr);
215
216	p = &jeb->last_node;
217	ref = *p;
218
219	dbg_memalloc("Reserving %d refs for block @0x%08x\n", nr, jeb->offset);
220
221	/* If jeb->last_node is really a valid node then skip over it */
222	if (ref && ref->flash_offset != REF_EMPTY_NODE)
223		ref++;
224
225	while (i) {
226		if (!ref) {
227			dbg_memalloc("Allocating new refblock linked from %p\n", p);
228			ref = *p = jffs2_alloc_refblock();
229			if (!ref)
230				return -ENOMEM;
231		}
232		if (ref->flash_offset == REF_LINK_NODE) {
233			p = &ref->next_in_ino;
234			ref = *p;
235			continue;
236		}
237		i--;
238		ref++;
239	}
240	jeb->allocated_refs = nr;
241
242	dbg_memalloc("Reserved %d refs for block @0x%08x, last_node is %p (%08x,%p)\n",
243		  nr, jeb->offset, jeb->last_node, jeb->last_node->flash_offset,
244		  jeb->last_node->next_in_ino);
245
246	return 0;
247}
248
249void jffs2_free_refblock(struct jffs2_raw_node_ref *x)
250{
251	dbg_memalloc("%p\n", x);
252	kmem_cache_free(raw_node_ref_slab, x);
253}
254
255struct jffs2_node_frag *jffs2_alloc_node_frag(void)
256{
257	struct jffs2_node_frag *ret;
258	ret = kmem_cache_alloc(node_frag_slab, GFP_KERNEL);
259	dbg_memalloc("%p\n", ret);
260	return ret;
261}
262
263void jffs2_free_node_frag(struct jffs2_node_frag *x)
264{
265	dbg_memalloc("%p\n", x);
266	kmem_cache_free(node_frag_slab, x);
267}
268
269struct jffs2_inode_cache *jffs2_alloc_inode_cache(void)
270{
271	struct jffs2_inode_cache *ret;
272	ret = kmem_cache_alloc(inode_cache_slab, GFP_KERNEL);
273	dbg_memalloc("%p\n", ret);
274	return ret;
275}
276
277void jffs2_free_inode_cache(struct jffs2_inode_cache *x)
278{
279	dbg_memalloc("%p\n", x);
280	kmem_cache_free(inode_cache_slab, x);
281}
282
283#ifdef CONFIG_JFFS2_FS_XATTR
284struct jffs2_xattr_datum *jffs2_alloc_xattr_datum(void)
285{
286	struct jffs2_xattr_datum *xd;
287	xd = kmem_cache_zalloc(xattr_datum_cache, GFP_KERNEL);
288	dbg_memalloc("%p\n", xd);
 
 
289
290	xd->class = RAWNODE_CLASS_XATTR_DATUM;
291	xd->node = (void *)xd;
292	INIT_LIST_HEAD(&xd->xindex);
293	return xd;
294}
295
296void jffs2_free_xattr_datum(struct jffs2_xattr_datum *xd)
297{
298	dbg_memalloc("%p\n", xd);
299	kmem_cache_free(xattr_datum_cache, xd);
300}
301
302struct jffs2_xattr_ref *jffs2_alloc_xattr_ref(void)
303{
304	struct jffs2_xattr_ref *ref;
305	ref = kmem_cache_zalloc(xattr_ref_cache, GFP_KERNEL);
306	dbg_memalloc("%p\n", ref);
 
 
307
308	ref->class = RAWNODE_CLASS_XATTR_REF;
309	ref->node = (void *)ref;
310	return ref;
311}
312
313void jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref)
314{
315	dbg_memalloc("%p\n", ref);
316	kmem_cache_free(xattr_ref_cache, ref);
317}
318#endif
v3.15
  1/*
  2 * JFFS2 -- Journalling Flash File System, Version 2.
  3 *
  4 * Copyright © 2001-2007 Red Hat, Inc.
  5 *
  6 * Created by David Woodhouse <dwmw2@infradead.org>
  7 *
  8 * For licensing information, see the file 'LICENCE' in this directory.
  9 *
 10 */
 11
 12#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
 13
 14#include <linux/kernel.h>
 15#include <linux/slab.h>
 16#include <linux/init.h>
 17#include <linux/jffs2.h>
 18#include "nodelist.h"
 19
 20/* These are initialised to NULL in the kernel startup code.
 21   If you're porting to other operating systems, beware */
 22static struct kmem_cache *full_dnode_slab;
 23static struct kmem_cache *raw_dirent_slab;
 24static struct kmem_cache *raw_inode_slab;
 25static struct kmem_cache *tmp_dnode_info_slab;
 26static struct kmem_cache *raw_node_ref_slab;
 27static struct kmem_cache *node_frag_slab;
 28static struct kmem_cache *inode_cache_slab;
 29#ifdef CONFIG_JFFS2_FS_XATTR
 30static struct kmem_cache *xattr_datum_cache;
 31static struct kmem_cache *xattr_ref_cache;
 32#endif
 33
 34int __init jffs2_create_slab_caches(void)
 35{
 36	full_dnode_slab = kmem_cache_create("jffs2_full_dnode",
 37					    sizeof(struct jffs2_full_dnode),
 38					    0, 0, NULL);
 39	if (!full_dnode_slab)
 40		goto err;
 41
 42	raw_dirent_slab = kmem_cache_create("jffs2_raw_dirent",
 43					    sizeof(struct jffs2_raw_dirent),
 44					    0, SLAB_HWCACHE_ALIGN, NULL);
 45	if (!raw_dirent_slab)
 46		goto err;
 47
 48	raw_inode_slab = kmem_cache_create("jffs2_raw_inode",
 49					   sizeof(struct jffs2_raw_inode),
 50					   0, SLAB_HWCACHE_ALIGN, NULL);
 51	if (!raw_inode_slab)
 52		goto err;
 53
 54	tmp_dnode_info_slab = kmem_cache_create("jffs2_tmp_dnode",
 55						sizeof(struct jffs2_tmp_dnode_info),
 56						0, 0, NULL);
 57	if (!tmp_dnode_info_slab)
 58		goto err;
 59
 60	raw_node_ref_slab = kmem_cache_create("jffs2_refblock",
 61					      sizeof(struct jffs2_raw_node_ref) * (REFS_PER_BLOCK + 1),
 62					      0, 0, NULL);
 63	if (!raw_node_ref_slab)
 64		goto err;
 65
 66	node_frag_slab = kmem_cache_create("jffs2_node_frag",
 67					   sizeof(struct jffs2_node_frag),
 68					   0, 0, NULL);
 69	if (!node_frag_slab)
 70		goto err;
 71
 72	inode_cache_slab = kmem_cache_create("jffs2_inode_cache",
 73					     sizeof(struct jffs2_inode_cache),
 74					     0, 0, NULL);
 75	if (!inode_cache_slab)
 76		goto err;
 77
 78#ifdef CONFIG_JFFS2_FS_XATTR
 79	xattr_datum_cache = kmem_cache_create("jffs2_xattr_datum",
 80					     sizeof(struct jffs2_xattr_datum),
 81					     0, 0, NULL);
 82	if (!xattr_datum_cache)
 83		goto err;
 84
 85	xattr_ref_cache = kmem_cache_create("jffs2_xattr_ref",
 86					   sizeof(struct jffs2_xattr_ref),
 87					   0, 0, NULL);
 88	if (!xattr_ref_cache)
 89		goto err;
 90#endif
 91
 92	return 0;
 93 err:
 94	jffs2_destroy_slab_caches();
 95	return -ENOMEM;
 96}
 97
 98void jffs2_destroy_slab_caches(void)
 99{
100	if(full_dnode_slab)
101		kmem_cache_destroy(full_dnode_slab);
102	if(raw_dirent_slab)
103		kmem_cache_destroy(raw_dirent_slab);
104	if(raw_inode_slab)
105		kmem_cache_destroy(raw_inode_slab);
106	if(tmp_dnode_info_slab)
107		kmem_cache_destroy(tmp_dnode_info_slab);
108	if(raw_node_ref_slab)
109		kmem_cache_destroy(raw_node_ref_slab);
110	if(node_frag_slab)
111		kmem_cache_destroy(node_frag_slab);
112	if(inode_cache_slab)
113		kmem_cache_destroy(inode_cache_slab);
114#ifdef CONFIG_JFFS2_FS_XATTR
115	if (xattr_datum_cache)
116		kmem_cache_destroy(xattr_datum_cache);
117	if (xattr_ref_cache)
118		kmem_cache_destroy(xattr_ref_cache);
119#endif
120}
121
122struct jffs2_full_dirent *jffs2_alloc_full_dirent(int namesize)
123{
124	struct jffs2_full_dirent *ret;
125	ret = kmalloc(sizeof(struct jffs2_full_dirent) + namesize, GFP_KERNEL);
126	dbg_memalloc("%p\n", ret);
127	return ret;
128}
129
130void jffs2_free_full_dirent(struct jffs2_full_dirent *x)
131{
132	dbg_memalloc("%p\n", x);
133	kfree(x);
134}
135
136struct jffs2_full_dnode *jffs2_alloc_full_dnode(void)
137{
138	struct jffs2_full_dnode *ret;
139	ret = kmem_cache_alloc(full_dnode_slab, GFP_KERNEL);
140	dbg_memalloc("%p\n", ret);
141	return ret;
142}
143
144void jffs2_free_full_dnode(struct jffs2_full_dnode *x)
145{
146	dbg_memalloc("%p\n", x);
147	kmem_cache_free(full_dnode_slab, x);
148}
149
150struct jffs2_raw_dirent *jffs2_alloc_raw_dirent(void)
151{
152	struct jffs2_raw_dirent *ret;
153	ret = kmem_cache_alloc(raw_dirent_slab, GFP_KERNEL);
154	dbg_memalloc("%p\n", ret);
155	return ret;
156}
157
158void jffs2_free_raw_dirent(struct jffs2_raw_dirent *x)
159{
160	dbg_memalloc("%p\n", x);
161	kmem_cache_free(raw_dirent_slab, x);
162}
163
164struct jffs2_raw_inode *jffs2_alloc_raw_inode(void)
165{
166	struct jffs2_raw_inode *ret;
167	ret = kmem_cache_alloc(raw_inode_slab, GFP_KERNEL);
168	dbg_memalloc("%p\n", ret);
169	return ret;
170}
171
172void jffs2_free_raw_inode(struct jffs2_raw_inode *x)
173{
174	dbg_memalloc("%p\n", x);
175	kmem_cache_free(raw_inode_slab, x);
176}
177
178struct jffs2_tmp_dnode_info *jffs2_alloc_tmp_dnode_info(void)
179{
180	struct jffs2_tmp_dnode_info *ret;
181	ret = kmem_cache_alloc(tmp_dnode_info_slab, GFP_KERNEL);
182	dbg_memalloc("%p\n",
183		ret);
184	return ret;
185}
186
187void jffs2_free_tmp_dnode_info(struct jffs2_tmp_dnode_info *x)
188{
189	dbg_memalloc("%p\n", x);
190	kmem_cache_free(tmp_dnode_info_slab, x);
191}
192
193static struct jffs2_raw_node_ref *jffs2_alloc_refblock(void)
194{
195	struct jffs2_raw_node_ref *ret;
196
197	ret = kmem_cache_alloc(raw_node_ref_slab, GFP_KERNEL);
198	if (ret) {
199		int i = 0;
200		for (i=0; i < REFS_PER_BLOCK; i++) {
201			ret[i].flash_offset = REF_EMPTY_NODE;
202			ret[i].next_in_ino = NULL;
203		}
204		ret[i].flash_offset = REF_LINK_NODE;
205		ret[i].next_in_ino = NULL;
206	}
207	return ret;
208}
209
210int jffs2_prealloc_raw_node_refs(struct jffs2_sb_info *c,
211				 struct jffs2_eraseblock *jeb, int nr)
212{
213	struct jffs2_raw_node_ref **p, *ref;
214	int i = nr;
215
216	dbg_memalloc("%d\n", nr);
217
218	p = &jeb->last_node;
219	ref = *p;
220
221	dbg_memalloc("Reserving %d refs for block @0x%08x\n", nr, jeb->offset);
222
223	/* If jeb->last_node is really a valid node then skip over it */
224	if (ref && ref->flash_offset != REF_EMPTY_NODE)
225		ref++;
226
227	while (i) {
228		if (!ref) {
229			dbg_memalloc("Allocating new refblock linked from %p\n", p);
230			ref = *p = jffs2_alloc_refblock();
231			if (!ref)
232				return -ENOMEM;
233		}
234		if (ref->flash_offset == REF_LINK_NODE) {
235			p = &ref->next_in_ino;
236			ref = *p;
237			continue;
238		}
239		i--;
240		ref++;
241	}
242	jeb->allocated_refs = nr;
243
244	dbg_memalloc("Reserved %d refs for block @0x%08x, last_node is %p (%08x,%p)\n",
245		  nr, jeb->offset, jeb->last_node, jeb->last_node->flash_offset,
246		  jeb->last_node->next_in_ino);
247
248	return 0;
249}
250
251void jffs2_free_refblock(struct jffs2_raw_node_ref *x)
252{
253	dbg_memalloc("%p\n", x);
254	kmem_cache_free(raw_node_ref_slab, x);
255}
256
257struct jffs2_node_frag *jffs2_alloc_node_frag(void)
258{
259	struct jffs2_node_frag *ret;
260	ret = kmem_cache_alloc(node_frag_slab, GFP_KERNEL);
261	dbg_memalloc("%p\n", ret);
262	return ret;
263}
264
265void jffs2_free_node_frag(struct jffs2_node_frag *x)
266{
267	dbg_memalloc("%p\n", x);
268	kmem_cache_free(node_frag_slab, x);
269}
270
271struct jffs2_inode_cache *jffs2_alloc_inode_cache(void)
272{
273	struct jffs2_inode_cache *ret;
274	ret = kmem_cache_alloc(inode_cache_slab, GFP_KERNEL);
275	dbg_memalloc("%p\n", ret);
276	return ret;
277}
278
279void jffs2_free_inode_cache(struct jffs2_inode_cache *x)
280{
281	dbg_memalloc("%p\n", x);
282	kmem_cache_free(inode_cache_slab, x);
283}
284
285#ifdef CONFIG_JFFS2_FS_XATTR
286struct jffs2_xattr_datum *jffs2_alloc_xattr_datum(void)
287{
288	struct jffs2_xattr_datum *xd;
289	xd = kmem_cache_zalloc(xattr_datum_cache, GFP_KERNEL);
290	dbg_memalloc("%p\n", xd);
291	if (!xd)
292		return NULL;
293
294	xd->class = RAWNODE_CLASS_XATTR_DATUM;
295	xd->node = (void *)xd;
296	INIT_LIST_HEAD(&xd->xindex);
297	return xd;
298}
299
300void jffs2_free_xattr_datum(struct jffs2_xattr_datum *xd)
301{
302	dbg_memalloc("%p\n", xd);
303	kmem_cache_free(xattr_datum_cache, xd);
304}
305
306struct jffs2_xattr_ref *jffs2_alloc_xattr_ref(void)
307{
308	struct jffs2_xattr_ref *ref;
309	ref = kmem_cache_zalloc(xattr_ref_cache, GFP_KERNEL);
310	dbg_memalloc("%p\n", ref);
311	if (!ref)
312		return NULL;
313
314	ref->class = RAWNODE_CLASS_XATTR_REF;
315	ref->node = (void *)ref;
316	return ref;
317}
318
319void jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref)
320{
321	dbg_memalloc("%p\n", ref);
322	kmem_cache_free(xattr_ref_cache, ref);
323}
324#endif