 d65f6f7009
			
		
	
	
		d65f6f7009
		
	
	
	
	
		
			
			. common/include/arch/i386 is not actually an imported sys/arch/i386/include but leftover Minix files; remove and move to include/ . move include/ufs to sys/ufs, where it came from, now that we have a sys/ hierarchy . move mdocml/ to external/bsd/, now we have that . single sys/arch/i386/stand/ import for boot stuff
		
			
				
	
	
		
			397 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			397 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*	$NetBSD: chfs_malloc.c,v 1.1 2011/11/24 15:51:31 ahoka Exp $	*/
 | |
| 
 | |
| /*-
 | |
|  * Copyright (c) 2010 Department of Software Engineering,
 | |
|  *		      University of Szeged, Hungary
 | |
|  * Copyright (C) 2010 Tamas Toth <ttoth@inf.u-szeged.hu>
 | |
|  * Copyright (C) 2010 Adam Hoka <ahoka@NetBSD.org>
 | |
|  * All rights reserved.
 | |
|  *
 | |
|  * This code is derived from software contributed to The NetBSD Foundation
 | |
|  * by the Department of Software Engineering, University of Szeged, Hungary
 | |
|  *
 | |
|  * Redistribution and use in source and binary forms, with or without
 | |
|  * modification, are permitted provided that the following conditions
 | |
|  * are met:
 | |
|  * 1. Redistributions of source code must retain the above copyright
 | |
|  *    notice, this list of conditions and the following disclaimer.
 | |
|  * 2. Redistributions in binary form must reproduce the above copyright
 | |
|  *    notice, this list of conditions and the following disclaimer in the
 | |
|  *    documentation and/or other materials provided with the distribution.
 | |
|  *
 | |
|  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
 | |
|  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 | |
|  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 | |
|  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
 | |
|  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
 | |
|  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
 | |
|  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
 | |
|  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
 | |
|  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 | |
|  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 | |
|  * SUCH DAMAGE.
 | |
|  */
 | |
| 
 | |
| #include "chfs.h"
 | |
| #include <sys/pool.h>
 | |
| 
 | |
| pool_cache_t chfs_vnode_cache;
 | |
| pool_cache_t chfs_nrefs_cache;
 | |
| pool_cache_t chfs_flash_vnode_cache;
 | |
| pool_cache_t chfs_flash_dirent_cache;
 | |
| pool_cache_t chfs_flash_dnode_cache;
 | |
| pool_cache_t chfs_node_frag_cache;
 | |
| pool_cache_t chfs_tmp_dnode_cache;
 | |
| pool_cache_t chfs_tmp_dnode_info_cache;
 | |
| 
 | |
| int
 | |
| chfs_alloc_pool_caches()
 | |
| {
 | |
| 	chfs_vnode_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_vnode_cache),
 | |
| 		0, 0, 0, "chfs_vnode_cache", NULL, IPL_NONE, NULL, NULL,
 | |
| 		NULL);
 | |
| 	if (!chfs_vnode_cache)
 | |
| 		goto err_vnode;
 | |
| 
 | |
| 	chfs_nrefs_cache = pool_cache_init(
 | |
| 		(REFS_BLOCK_LEN + 1) * sizeof(struct chfs_node_ref), 0, 0,
 | |
| 		0, "chfs_nrefs_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_nrefs_cache)
 | |
| 		goto err_nrefs;
 | |
| 
 | |
| 	chfs_flash_vnode_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_flash_vnode), 0, 0, 0,
 | |
| 		"chfs_flash_vnode_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_flash_vnode_cache)
 | |
| 		goto err_flash_vnode;
 | |
| 
 | |
| 	chfs_flash_dirent_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_flash_dirent_node), 0, 0, 0,
 | |
| 		"chfs_flash_dirent_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_flash_dirent_cache)
 | |
| 		goto err_flash_dirent;
 | |
| 
 | |
| 	chfs_flash_dnode_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_flash_data_node), 0, 0, 0,
 | |
| 		"chfs_flash_dnode_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_flash_dnode_cache)
 | |
| 		goto err_flash_dnode;
 | |
| 
 | |
| 	chfs_node_frag_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_node_frag), 0, 0, 0,
 | |
| 		"chfs_node_frag_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_node_frag_cache)
 | |
| 		goto err_node_frag;
 | |
| 
 | |
| 	chfs_tmp_dnode_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_tmp_dnode), 0, 0, 0,
 | |
| 		"chfs_tmp_dnode_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_tmp_dnode_cache)
 | |
| 		goto err_tmp_dnode;
 | |
| 
 | |
| 	chfs_tmp_dnode_info_cache = pool_cache_init(
 | |
| 		sizeof(struct chfs_tmp_dnode_info), 0, 0, 0,
 | |
| 		"chfs_tmp_dnode_info_pool", NULL, IPL_NONE, NULL, NULL, NULL);
 | |
| 	if (!chfs_tmp_dnode_info_cache)
 | |
| 		goto err_tmp_dnode_info;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err_tmp_dnode_info:
 | |
| 	pool_cache_destroy(chfs_tmp_dnode_cache);
 | |
| err_tmp_dnode:
 | |
| 	pool_cache_destroy(chfs_node_frag_cache);
 | |
| err_node_frag:
 | |
| 	pool_cache_destroy(chfs_flash_dnode_cache);
 | |
| err_flash_dnode:
 | |
| 	pool_cache_destroy(chfs_flash_dirent_cache);
 | |
| err_flash_dirent:
 | |
| 	pool_cache_destroy(chfs_flash_vnode_cache);
 | |
| err_flash_vnode:
 | |
| 	pool_cache_destroy(chfs_nrefs_cache);
 | |
| err_nrefs:
 | |
| 	pool_cache_destroy(chfs_vnode_cache);
 | |
| err_vnode:
 | |
| 
 | |
| 	return ENOMEM;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_destroy_pool_caches()
 | |
| {
 | |
| 	if (chfs_vnode_cache)
 | |
| 		pool_cache_destroy(chfs_vnode_cache);
 | |
| 
 | |
| 	if (chfs_nrefs_cache)
 | |
| 		pool_cache_destroy(chfs_nrefs_cache);
 | |
| 
 | |
| 	if (chfs_flash_vnode_cache)
 | |
| 		pool_cache_destroy(chfs_flash_vnode_cache);
 | |
| 
 | |
| 	if (chfs_flash_dirent_cache)
 | |
| 		pool_cache_destroy(chfs_flash_dirent_cache);
 | |
| 
 | |
| 	if (chfs_flash_dnode_cache)
 | |
| 		pool_cache_destroy(chfs_flash_dnode_cache);
 | |
| 
 | |
| 	if (chfs_node_frag_cache)
 | |
| 		pool_cache_destroy(chfs_node_frag_cache);
 | |
| 
 | |
| 	if (chfs_tmp_dnode_cache)
 | |
| 		pool_cache_destroy(chfs_tmp_dnode_cache);
 | |
| 
 | |
| 	if (chfs_tmp_dnode_info_cache)
 | |
| 		pool_cache_destroy(chfs_tmp_dnode_info_cache);
 | |
| }
 | |
| 
 | |
| struct chfs_vnode_cache *
 | |
| chfs_vnode_cache_alloc(ino_t vno)
 | |
| {
 | |
| 	struct chfs_vnode_cache* vc;
 | |
| 	vc = pool_cache_get(chfs_vnode_cache, PR_WAITOK);
 | |
| 
 | |
| 	memset(vc, 0, sizeof(*vc));
 | |
| 	vc->vno = vno;
 | |
| 	vc->v = (void *)vc;
 | |
| 	vc->dirents = (void *)vc;
 | |
| 	vc->dnode = (void *)vc;
 | |
| 	TAILQ_INIT(&vc->scan_dirents);
 | |
| 	vc->highest_version = 0;
 | |
| 
 | |
| 	return vc;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_vnode_cache_free(struct chfs_vnode_cache *vc)
 | |
| {
 | |
| 	//kmem_free(vc->vno_version, sizeof(uint64_t));
 | |
| 	pool_cache_put(chfs_vnode_cache, vc);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * chfs_alloc_refblock - allocating a refblock
 | |
|  *
 | |
|  * Returns a pointer of the first element in the block.
 | |
|  *
 | |
|  * We are not allocating just one node ref, instead we allocating REFS_BLOCK_LEN
 | |
|  * number of node refs, the last element will be a pointer to the next block.
 | |
|  * We do this, because we need a chain of nodes which have been ordered by the
 | |
|  * physical address of them.
 | |
|  *
 | |
|  */
 | |
| struct chfs_node_ref*
 | |
| chfs_alloc_refblock(void)
 | |
| {
 | |
| 	int i;
 | |
| 	struct chfs_node_ref *nref;
 | |
| 	nref = pool_cache_get(chfs_nrefs_cache, PR_WAITOK);
 | |
| 
 | |
| 	for (i = 0; i < REFS_BLOCK_LEN; i++) {
 | |
| 		nref[i].nref_lnr = REF_EMPTY_NODE;
 | |
| 		nref[i].nref_next = NULL;
 | |
| 	}
 | |
| 	i = REFS_BLOCK_LEN;
 | |
| 	nref[i].nref_lnr = REF_LINK_TO_NEXT;
 | |
| 	nref[i].nref_next = NULL;
 | |
| 
 | |
| 	return nref;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * chfs_free_refblock - freeing a refblock
 | |
|  */
 | |
| void
 | |
| chfs_free_refblock(struct chfs_node_ref *nref)
 | |
| {
 | |
| 	pool_cache_put(chfs_nrefs_cache, nref);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * chfs_alloc_node_ref - allocating a node ref from a refblock
 | |
|  * @cheb: eraseblock information structure
 | |
|  *
 | |
|  * Allocating a node ref from a refblock, it there isn't any free element in the
 | |
|  * block, a new block will be allocated and be linked to the current block.
 | |
|  */
 | |
| struct chfs_node_ref*
 | |
| chfs_alloc_node_ref(struct chfs_eraseblock *cheb)
 | |
| {
 | |
| 	struct chfs_node_ref *nref, *new, *old;
 | |
| 	old = cheb->last_node;
 | |
| 	nref = cheb->last_node;
 | |
| 
 | |
| 	if (!nref) {
 | |
| 		//There haven't been any nref allocated for this block yet
 | |
| 		nref = chfs_alloc_refblock();
 | |
| 
 | |
| 		cheb->first_node = nref;
 | |
| 		cheb->last_node = nref;
 | |
| 		nref->nref_lnr = cheb->lnr;
 | |
| 		KASSERT(cheb->lnr == nref->nref_lnr);
 | |
| 
 | |
| 		return nref;
 | |
| 	}
 | |
| 
 | |
| 	nref++;
 | |
| 	if (nref->nref_lnr == REF_LINK_TO_NEXT) {
 | |
| 		new = chfs_alloc_refblock();
 | |
| 		nref->nref_next = new;
 | |
| 		nref = new;
 | |
| 	}
 | |
| 
 | |
| 	cheb->last_node = nref;
 | |
| 	nref->nref_lnr = cheb->lnr;
 | |
| 
 | |
| 	KASSERT(old->nref_lnr == nref->nref_lnr &&
 | |
| 	    nref->nref_lnr == cheb->lnr);
 | |
| 
 | |
| 	return nref;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * chfs_free_node_refs - freeing an eraseblock's node refs
 | |
|  * @cheb: eraseblock information structure
 | |
|  */
 | |
| void
 | |
| chfs_free_node_refs(struct chfs_eraseblock *cheb)
 | |
| {
 | |
| 	struct chfs_node_ref *nref, *block;
 | |
| 
 | |
| 	block = nref = cheb->first_node;
 | |
| 
 | |
| 	while (nref) {
 | |
| 		if (nref->nref_lnr == REF_LINK_TO_NEXT) {
 | |
| 			nref = nref->nref_next;
 | |
| 			chfs_free_refblock(block);
 | |
| 			block = nref;
 | |
| 			continue;
 | |
| 		}
 | |
| 		nref++;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| struct chfs_dirent*
 | |
| chfs_alloc_dirent(int namesize)
 | |
| {
 | |
| 	struct chfs_dirent *ret;
 | |
| 	size_t size = sizeof(struct chfs_dirent) + namesize;
 | |
| 
 | |
| 	ret = kmem_alloc(size, KM_SLEEP);
 | |
| 	//ret->alloc_size = size;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_dirent(struct chfs_dirent *dirent)
 | |
| {
 | |
| 	//size_t size = dirent->alloc_size;
 | |
| 	size_t size = sizeof(struct chfs_dirent) + dirent->nsize + 1;
 | |
| 
 | |
| 	kmem_free(dirent, size);
 | |
| }
 | |
| 
 | |
| struct chfs_full_dnode*
 | |
| chfs_alloc_full_dnode()
 | |
| {
 | |
| 	struct chfs_full_dnode *ret;
 | |
| 	ret = kmem_alloc(sizeof(struct chfs_full_dnode), KM_SLEEP);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_full_dnode(struct chfs_full_dnode *fd)
 | |
| {
 | |
| 	kmem_free(fd,(sizeof(struct chfs_full_dnode)));
 | |
| }
 | |
| 
 | |
| struct chfs_flash_vnode*
 | |
| chfs_alloc_flash_vnode()
 | |
| {
 | |
| 	struct chfs_flash_vnode *ret;
 | |
| 	ret = pool_cache_get(chfs_flash_vnode_cache, 0);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_flash_vnode(struct chfs_flash_vnode *fvnode)
 | |
| {
 | |
| 	pool_cache_put(chfs_flash_vnode_cache, fvnode);
 | |
| }
 | |
| 
 | |
| struct chfs_flash_dirent_node*
 | |
| chfs_alloc_flash_dirent()
 | |
| {
 | |
| 	struct chfs_flash_dirent_node *ret;
 | |
| 	ret = pool_cache_get(chfs_flash_dirent_cache, 0);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_flash_dirent(struct chfs_flash_dirent_node *fdnode)
 | |
| {
 | |
| 	pool_cache_put(chfs_flash_dirent_cache, fdnode);
 | |
| }
 | |
| 
 | |
| struct chfs_flash_data_node*
 | |
| chfs_alloc_flash_dnode()
 | |
| {
 | |
| 	struct chfs_flash_data_node *ret;
 | |
| 	ret = pool_cache_get(chfs_flash_dnode_cache, 0);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_flash_dnode(struct chfs_flash_data_node *fdnode)
 | |
| {
 | |
| 	pool_cache_put(chfs_flash_dnode_cache, fdnode);
 | |
| }
 | |
| 
 | |
| 
 | |
| struct chfs_node_frag*
 | |
| chfs_alloc_node_frag()
 | |
| {
 | |
| 	struct chfs_node_frag *ret;
 | |
| 	ret = pool_cache_get(chfs_node_frag_cache, 0);
 | |
| 	return ret;
 | |
| 
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_node_frag(struct chfs_node_frag *frag)
 | |
| {
 | |
| 	pool_cache_put(chfs_node_frag_cache, frag);
 | |
| }
 | |
| 
 | |
| struct chfs_tmp_dnode *
 | |
| chfs_alloc_tmp_dnode()
 | |
| {
 | |
| 	struct chfs_tmp_dnode *ret;
 | |
| 	ret = pool_cache_get(chfs_tmp_dnode_cache, 0);
 | |
| 	ret->next = NULL;
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_tmp_dnode(struct chfs_tmp_dnode *td)
 | |
| {
 | |
| 	pool_cache_put(chfs_tmp_dnode_cache, td);
 | |
| }
 | |
| 
 | |
| struct chfs_tmp_dnode_info *
 | |
| chfs_alloc_tmp_dnode_info()
 | |
| {
 | |
| 	struct chfs_tmp_dnode_info *ret;
 | |
| 	ret = pool_cache_get(chfs_tmp_dnode_info_cache, 0);
 | |
| 	ret->tmpnode = NULL;
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void
 | |
| chfs_free_tmp_dnode_info(struct chfs_tmp_dnode_info *di)
 | |
| {
 | |
| 	pool_cache_put(chfs_tmp_dnode_info_cache, di);
 | |
| }
 | |
| 
 |