Things should be able to depend on object.h without pulling in all of cache.h. Move an enum to allow this. Note that a couple files previously depended on things brought in through cache.h indirectly (revision.h -> commit.h -> object.h -> cache.h). As such, this change requires making existing dependencies more explicit in half a dozen files. The inclusion of strbuf.h in some headers if of particular note: these headers directly embedded a strbuf in some new structs, meaning they should have been including strbuf.h all along but were indirectly getting the necessary definitions. Signed-off-by: Elijah Newren <newren@gmail.com> Signed-off-by: Junio C Hamano <gitster@pobox.com>
		
			
				
	
	
		
			123 lines
		
	
	
		
			2.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			123 lines
		
	
	
		
			2.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * alloc.c  - specialized allocator for internal objects
 | 
						|
 *
 | 
						|
 * Copyright (C) 2006 Linus Torvalds
 | 
						|
 *
 | 
						|
 * The standard malloc/free wastes too much space for objects, partly because
 | 
						|
 * it maintains all the allocation infrastructure, but even more because it ends
 | 
						|
 * up with maximal alignment because it doesn't know what the object alignment
 | 
						|
 * for the new allocation is.
 | 
						|
 */
 | 
						|
#include "git-compat-util.h"
 | 
						|
#include "object.h"
 | 
						|
#include "blob.h"
 | 
						|
#include "tree.h"
 | 
						|
#include "commit.h"
 | 
						|
#include "tag.h"
 | 
						|
#include "alloc.h"
 | 
						|
 | 
						|
#define BLOCKING 1024
 | 
						|
 | 
						|
union any_object {
 | 
						|
	struct object object;
 | 
						|
	struct blob blob;
 | 
						|
	struct tree tree;
 | 
						|
	struct commit commit;
 | 
						|
	struct tag tag;
 | 
						|
};
 | 
						|
 | 
						|
struct alloc_state {
 | 
						|
	int nr;    /* number of nodes left in current allocation */
 | 
						|
	void *p;   /* first free node in current allocation */
 | 
						|
 | 
						|
	/* bookkeeping of allocations */
 | 
						|
	void **slabs;
 | 
						|
	int slab_nr, slab_alloc;
 | 
						|
};
 | 
						|
 | 
						|
struct alloc_state *allocate_alloc_state(void)
 | 
						|
{
 | 
						|
	return xcalloc(1, sizeof(struct alloc_state));
 | 
						|
}
 | 
						|
 | 
						|
void clear_alloc_state(struct alloc_state *s)
 | 
						|
{
 | 
						|
	while (s->slab_nr > 0) {
 | 
						|
		s->slab_nr--;
 | 
						|
		free(s->slabs[s->slab_nr]);
 | 
						|
	}
 | 
						|
 | 
						|
	FREE_AND_NULL(s->slabs);
 | 
						|
}
 | 
						|
 | 
						|
static inline void *alloc_node(struct alloc_state *s, size_t node_size)
 | 
						|
{
 | 
						|
	void *ret;
 | 
						|
 | 
						|
	if (!s->nr) {
 | 
						|
		s->nr = BLOCKING;
 | 
						|
		s->p = xmalloc(BLOCKING * node_size);
 | 
						|
 | 
						|
		ALLOC_GROW(s->slabs, s->slab_nr + 1, s->slab_alloc);
 | 
						|
		s->slabs[s->slab_nr++] = s->p;
 | 
						|
	}
 | 
						|
	s->nr--;
 | 
						|
	ret = s->p;
 | 
						|
	s->p = (char *)s->p + node_size;
 | 
						|
	memset(ret, 0, node_size);
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
void *alloc_blob_node(struct repository *r)
 | 
						|
{
 | 
						|
	struct blob *b = alloc_node(r->parsed_objects->blob_state, sizeof(struct blob));
 | 
						|
	b->object.type = OBJ_BLOB;
 | 
						|
	return b;
 | 
						|
}
 | 
						|
 | 
						|
void *alloc_tree_node(struct repository *r)
 | 
						|
{
 | 
						|
	struct tree *t = alloc_node(r->parsed_objects->tree_state, sizeof(struct tree));
 | 
						|
	t->object.type = OBJ_TREE;
 | 
						|
	return t;
 | 
						|
}
 | 
						|
 | 
						|
void *alloc_tag_node(struct repository *r)
 | 
						|
{
 | 
						|
	struct tag *t = alloc_node(r->parsed_objects->tag_state, sizeof(struct tag));
 | 
						|
	t->object.type = OBJ_TAG;
 | 
						|
	return t;
 | 
						|
}
 | 
						|
 | 
						|
void *alloc_object_node(struct repository *r)
 | 
						|
{
 | 
						|
	struct object *obj = alloc_node(r->parsed_objects->object_state, sizeof(union any_object));
 | 
						|
	obj->type = OBJ_NONE;
 | 
						|
	return obj;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * The returned count is to be used as an index into commit slabs,
 | 
						|
 * that are *NOT* maintained per repository, and that is why a single
 | 
						|
 * global counter is used.
 | 
						|
 */
 | 
						|
static unsigned int alloc_commit_index(void)
 | 
						|
{
 | 
						|
	static unsigned int parsed_commits_count;
 | 
						|
	return parsed_commits_count++;
 | 
						|
}
 | 
						|
 | 
						|
void init_commit_node(struct commit *c)
 | 
						|
{
 | 
						|
	c->object.type = OBJ_COMMIT;
 | 
						|
	c->index = alloc_commit_index();
 | 
						|
}
 | 
						|
 | 
						|
void *alloc_commit_node(struct repository *r)
 | 
						|
{
 | 
						|
	struct commit *c = alloc_node(r->parsed_objects->commit_state, sizeof(struct commit));
 | 
						|
	init_commit_node(c);
 | 
						|
	return c;
 | 
						|
}
 |