Update hashwrite() and friends to use the unsafe_-variants of hashing
functions, calling for e.g., "the_hash_algo->unsafe_update_fn()" instead
of "the_hash_algo->update_fn()".
These callers only use the_hash_algo to produce a checksum, which we
depend on for data integrity, but not for cryptographic purposes, so
these callers are safe to use the unsafe (non-collision detecting) SHA-1
implementation.
To time this, I took a freshly packed copy of linux.git, and ran the
following with and without the OPENSSL_SHA1_UNSAFE=1 build-knob. Both
versions were compiled with -O3:
    $ git for-each-ref --format='%(objectname)' refs/heads refs/tags >in
    $ valgrind --tool=callgrind ~/src/git/git-pack-objects \
        --revs --stdout --all-progress --use-bitmap-index <in >/dev/null
Without OPENSSL_SHA1_UNSAFE=1 (that is, using the collision-detecting
SHA-1 implementation for both cryptographic and non-cryptographic
purposes), we spend a significant amount of our instruction count in
hashwrite():
    $ callgrind_annotate --inclusive=yes | grep hashwrite | head -n1
    159,998,868,413 (79.42%)  /home/ttaylorr/src/git/csum-file.c:hashwrite [/home/ttaylorr/src/git/git-pack-objects]
, and the resulting "clone" takes 19.219 seconds of wall clock time,
18.94 seconds of user time and 0.28 seconds of system time.
Compiling with OPENSSL_SHA1_UNSAFE=1, we spend ~60% fewer instructions
in hashwrite():
    $ callgrind_annotate --inclusive=yes | grep hashwrite | head -n1
     59,164,001,176 (58.79%)  /home/ttaylorr/src/git/csum-file.c:hashwrite [/home/ttaylorr/src/git/git-pack-objects]
, and generate the resulting "clone" much faster, in only 11.597 seconds
of wall time, 11.37 seconds of user time, and 0.23 seconds of system
time, for a ~40% speed-up.
Signed-off-by: Taylor Blau <me@ttaylorr.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
		
	
		
			
				
	
	
		
			254 lines
		
	
	
		
			5.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			254 lines
		
	
	
		
			5.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * csum-file.c
 | 
						|
 *
 | 
						|
 * Copyright (C) 2005 Linus Torvalds
 | 
						|
 *
 | 
						|
 * Simple file write infrastructure for writing SHA1-summed
 | 
						|
 * files. Useful when you write a file that you want to be
 | 
						|
 * able to verify hasn't been messed with afterwards.
 | 
						|
 */
 | 
						|
 | 
						|
#define USE_THE_REPOSITORY_VARIABLE
 | 
						|
 | 
						|
#include "git-compat-util.h"
 | 
						|
#include "progress.h"
 | 
						|
#include "csum-file.h"
 | 
						|
#include "hash.h"
 | 
						|
 | 
						|
static void verify_buffer_or_die(struct hashfile *f,
 | 
						|
				 const void *buf,
 | 
						|
				 unsigned int count)
 | 
						|
{
 | 
						|
	ssize_t ret = read_in_full(f->check_fd, f->check_buffer, count);
 | 
						|
 | 
						|
	if (ret < 0)
 | 
						|
		die_errno("%s: sha1 file read error", f->name);
 | 
						|
	if (ret != count)
 | 
						|
		die("%s: sha1 file truncated", f->name);
 | 
						|
	if (memcmp(buf, f->check_buffer, count))
 | 
						|
		die("sha1 file '%s' validation error", f->name);
 | 
						|
}
 | 
						|
 | 
						|
static void flush(struct hashfile *f, const void *buf, unsigned int count)
 | 
						|
{
 | 
						|
	if (0 <= f->check_fd && count)
 | 
						|
		verify_buffer_or_die(f, buf, count);
 | 
						|
 | 
						|
	if (write_in_full(f->fd, buf, count) < 0) {
 | 
						|
		if (errno == ENOSPC)
 | 
						|
			die("sha1 file '%s' write error. Out of diskspace", f->name);
 | 
						|
		die_errno("sha1 file '%s' write error", f->name);
 | 
						|
	}
 | 
						|
 | 
						|
	f->total += count;
 | 
						|
	display_throughput(f->tp, f->total);
 | 
						|
}
 | 
						|
 | 
						|
void hashflush(struct hashfile *f)
 | 
						|
{
 | 
						|
	unsigned offset = f->offset;
 | 
						|
 | 
						|
	if (offset) {
 | 
						|
		if (!f->skip_hash)
 | 
						|
			the_hash_algo->unsafe_update_fn(&f->ctx, f->buffer, offset);
 | 
						|
		flush(f, f->buffer, offset);
 | 
						|
		f->offset = 0;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
void free_hashfile(struct hashfile *f)
 | 
						|
{
 | 
						|
	free(f->buffer);
 | 
						|
	free(f->check_buffer);
 | 
						|
	free(f);
 | 
						|
}
 | 
						|
 | 
						|
int finalize_hashfile(struct hashfile *f, unsigned char *result,
 | 
						|
		      enum fsync_component component, unsigned int flags)
 | 
						|
{
 | 
						|
	int fd;
 | 
						|
 | 
						|
	hashflush(f);
 | 
						|
 | 
						|
	if (f->skip_hash)
 | 
						|
		hashclr(f->buffer, the_repository->hash_algo);
 | 
						|
	else
 | 
						|
		the_hash_algo->unsafe_final_fn(f->buffer, &f->ctx);
 | 
						|
 | 
						|
	if (result)
 | 
						|
		hashcpy(result, f->buffer, the_repository->hash_algo);
 | 
						|
	if (flags & CSUM_HASH_IN_STREAM)
 | 
						|
		flush(f, f->buffer, the_hash_algo->rawsz);
 | 
						|
	if (flags & CSUM_FSYNC)
 | 
						|
		fsync_component_or_die(component, f->fd, f->name);
 | 
						|
	if (flags & CSUM_CLOSE) {
 | 
						|
		if (close(f->fd))
 | 
						|
			die_errno("%s: sha1 file error on close", f->name);
 | 
						|
		fd = 0;
 | 
						|
	} else
 | 
						|
		fd = f->fd;
 | 
						|
	if (0 <= f->check_fd) {
 | 
						|
		char discard;
 | 
						|
		int cnt = read_in_full(f->check_fd, &discard, 1);
 | 
						|
		if (cnt < 0)
 | 
						|
			die_errno("%s: error when reading the tail of sha1 file",
 | 
						|
				  f->name);
 | 
						|
		if (cnt)
 | 
						|
			die("%s: sha1 file has trailing garbage", f->name);
 | 
						|
		if (close(f->check_fd))
 | 
						|
			die_errno("%s: sha1 file error on close", f->name);
 | 
						|
	}
 | 
						|
	free_hashfile(f);
 | 
						|
	return fd;
 | 
						|
}
 | 
						|
 | 
						|
void discard_hashfile(struct hashfile *f)
 | 
						|
{
 | 
						|
	if (0 <= f->check_fd)
 | 
						|
		close(f->check_fd);
 | 
						|
	if (0 <= f->fd)
 | 
						|
		close(f->fd);
 | 
						|
	free_hashfile(f);
 | 
						|
}
 | 
						|
 | 
						|
void hashwrite(struct hashfile *f, const void *buf, unsigned int count)
 | 
						|
{
 | 
						|
	while (count) {
 | 
						|
		unsigned left = f->buffer_len - f->offset;
 | 
						|
		unsigned nr = count > left ? left : count;
 | 
						|
 | 
						|
		if (f->do_crc)
 | 
						|
			f->crc32 = crc32(f->crc32, buf, nr);
 | 
						|
 | 
						|
		if (nr == f->buffer_len) {
 | 
						|
			/*
 | 
						|
			 * Flush a full batch worth of data directly
 | 
						|
			 * from the input, skipping the memcpy() to
 | 
						|
			 * the hashfile's buffer. In this block,
 | 
						|
			 * f->offset is necessarily zero.
 | 
						|
			 */
 | 
						|
			if (!f->skip_hash)
 | 
						|
				the_hash_algo->unsafe_update_fn(&f->ctx, buf, nr);
 | 
						|
			flush(f, buf, nr);
 | 
						|
		} else {
 | 
						|
			/*
 | 
						|
			 * Copy to the hashfile's buffer, flushing only
 | 
						|
			 * if it became full.
 | 
						|
			 */
 | 
						|
			memcpy(f->buffer + f->offset, buf, nr);
 | 
						|
			f->offset += nr;
 | 
						|
			left -= nr;
 | 
						|
			if (!left)
 | 
						|
				hashflush(f);
 | 
						|
		}
 | 
						|
 | 
						|
		count -= nr;
 | 
						|
		buf = (char *) buf + nr;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
struct hashfile *hashfd_check(const char *name)
 | 
						|
{
 | 
						|
	int sink, check;
 | 
						|
	struct hashfile *f;
 | 
						|
 | 
						|
	sink = xopen("/dev/null", O_WRONLY);
 | 
						|
	check = xopen(name, O_RDONLY);
 | 
						|
	f = hashfd(sink, name);
 | 
						|
	f->check_fd = check;
 | 
						|
	f->check_buffer = xmalloc(f->buffer_len);
 | 
						|
 | 
						|
	return f;
 | 
						|
}
 | 
						|
 | 
						|
static struct hashfile *hashfd_internal(int fd, const char *name,
 | 
						|
					struct progress *tp,
 | 
						|
					size_t buffer_len)
 | 
						|
{
 | 
						|
	struct hashfile *f = xmalloc(sizeof(*f));
 | 
						|
	f->fd = fd;
 | 
						|
	f->check_fd = -1;
 | 
						|
	f->offset = 0;
 | 
						|
	f->total = 0;
 | 
						|
	f->tp = tp;
 | 
						|
	f->name = name;
 | 
						|
	f->do_crc = 0;
 | 
						|
	f->skip_hash = 0;
 | 
						|
	the_hash_algo->unsafe_init_fn(&f->ctx);
 | 
						|
 | 
						|
	f->buffer_len = buffer_len;
 | 
						|
	f->buffer = xmalloc(buffer_len);
 | 
						|
	f->check_buffer = NULL;
 | 
						|
 | 
						|
	return f;
 | 
						|
}
 | 
						|
 | 
						|
struct hashfile *hashfd(int fd, const char *name)
 | 
						|
{
 | 
						|
	/*
 | 
						|
	 * Since we are not going to use a progress meter to
 | 
						|
	 * measure the rate of data passing through this hashfile,
 | 
						|
	 * use a larger buffer size to reduce fsync() calls.
 | 
						|
	 */
 | 
						|
	return hashfd_internal(fd, name, NULL, 128 * 1024);
 | 
						|
}
 | 
						|
 | 
						|
struct hashfile *hashfd_throughput(int fd, const char *name, struct progress *tp)
 | 
						|
{
 | 
						|
	/*
 | 
						|
	 * Since we are expecting to report progress of the
 | 
						|
	 * write into this hashfile, use a smaller buffer
 | 
						|
	 * size so the progress indicators arrive at a more
 | 
						|
	 * frequent rate.
 | 
						|
	 */
 | 
						|
	return hashfd_internal(fd, name, tp, 8 * 1024);
 | 
						|
}
 | 
						|
 | 
						|
void hashfile_checkpoint(struct hashfile *f, struct hashfile_checkpoint *checkpoint)
 | 
						|
{
 | 
						|
	hashflush(f);
 | 
						|
	checkpoint->offset = f->total;
 | 
						|
	the_hash_algo->unsafe_clone_fn(&checkpoint->ctx, &f->ctx);
 | 
						|
}
 | 
						|
 | 
						|
int hashfile_truncate(struct hashfile *f, struct hashfile_checkpoint *checkpoint)
 | 
						|
{
 | 
						|
	off_t offset = checkpoint->offset;
 | 
						|
 | 
						|
	if (ftruncate(f->fd, offset) ||
 | 
						|
	    lseek(f->fd, offset, SEEK_SET) != offset)
 | 
						|
		return -1;
 | 
						|
	f->total = offset;
 | 
						|
	the_hash_algo->unsafe_clone_fn(&f->ctx, &checkpoint->ctx);
 | 
						|
	f->offset = 0; /* hashflush() was called in checkpoint */
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
void crc32_begin(struct hashfile *f)
 | 
						|
{
 | 
						|
	f->crc32 = crc32(0, NULL, 0);
 | 
						|
	f->do_crc = 1;
 | 
						|
}
 | 
						|
 | 
						|
uint32_t crc32_end(struct hashfile *f)
 | 
						|
{
 | 
						|
	f->do_crc = 0;
 | 
						|
	return f->crc32;
 | 
						|
}
 | 
						|
 | 
						|
int hashfile_checksum_valid(const unsigned char *data, size_t total_len)
 | 
						|
{
 | 
						|
	unsigned char got[GIT_MAX_RAWSZ];
 | 
						|
	git_hash_ctx ctx;
 | 
						|
	size_t data_len = total_len - the_hash_algo->rawsz;
 | 
						|
 | 
						|
	if (total_len < the_hash_algo->rawsz)
 | 
						|
		return 0; /* say "too short"? */
 | 
						|
 | 
						|
	the_hash_algo->unsafe_init_fn(&ctx);
 | 
						|
	the_hash_algo->unsafe_update_fn(&ctx, data, data_len);
 | 
						|
	the_hash_algo->unsafe_final_fn(got, &ctx);
 | 
						|
 | 
						|
	return hasheq(got, data + data_len, the_repository->hash_algo);
 | 
						|
}
 |