[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: duplicate entries on ext3 when using readdir/readdir64

On Tue, Aug 05, 2008 at 12:53:51PM +0200, Thomas Trauner wrote:
> On Fri, 2008-08-01 at 08:16 -0400, Theodore Tso wrote:
> > On Fri, Aug 01, 2008 at 11:43:40AM +0200, Thomas Trauner wrote:
> > > 
> > > I have a problem with directories that contain more than 10000 entries
> > > (Ubuntu 8.04.1) or with more than 70000 entries (RHEL 5.2). If you use
> > > readdir(3) or readdir64(3) you get one entry twice, with same name and
> > > inode.
> > > 
> I made new tests with the code under
> <http://www.unet.univie.ac.at/~a9100884/readdir.c> on a bunch of freshly
> generated and empty filesystems, every about 38GB large, of type fat
> (aborted after about 22000 entries because it took to long), ext2, xfs,
> jfs and again ext3....

OK, I have a workaroud for you.  It appears there's a kernel bug
hiding here, since there shouldn't be duplicates returned by readdir()
even if we have hash collisions.  

It turns out though that the TEA hash we are currently using as the
default is a really sucky hash.  I can't remember who suggested it; I
may go looking in the archives just out of curiosity.  My fault,
though, I should have tested it much more thoroughly, although it
*looked* good, and it was take from the core of an encryption
algorithm, so I thought it would be OK.

The claim was that it was just as good for our purposes as the
cut-down md4 hash we were using, but it was faster (so it would burn
less cpu cycles).  Unfortunately, (a) at least on modern hardware (I
tested on an X61s laptop) the TEA hash is in fact a little slower, and
(b) for small filenames with small hamming distances between them,,
such as what you are using in your test, it's generating lots of

Anyway, the workaround is as follows:

debugfs -w /dev/sdXXX
debugfs: set_super_value def_hash_version half_md4
debugfs: quit

Then completely delete any directories where you were having problems,
and recreate them.  (You can do the "mkdir foo.new; mv foo/* foo.new;
rmdir foo; mv foo.new foo" trick if you want to preserve the files in
that directory.)

In any case, here's the test case which shows the hash collision
problem much more quickly.  You can also use it for benchmarks, like

time tst_hash -q -a tea -n 3000000
time tst_hash -q -a half_md4 -n 3000000

With the following options, we can also see with the right filename
lengths, the tea algorithm doesn't create any hash collisions, so
maybe whoever tested the algorithm before they suggested it just got
unlucky with the set of filenames that he/she chose:

   tst_hash -p 0000 -a tea -n 3000000

In any case, unless someone comes up with a really good reason, I
probably will change the default hash algorithm for mke2fs to
half_md4, since it is both faster and a better hash function.

This doesn't change the fact that the kernel should do the right thing
with hash collisions, at least in the simple case without
telldir/seekdir.  When I merged the htree code I had tested it with
the Douglas Adams hash (always returns a hash value of
0x00000042:0000000 no matter what its inputs), and it did the right
thing, so we must have regressed somewhere along the line...

    	  	       	    	 	   - Ted

 * tst_htree.c
 * Copyright (C) 2008 by Theodore Ts'o.
 * This file may be redistributed under the terms of the GNU Public
 * License, Version 2
 * Compile command:
 *	cc -g -O2 -o tst_hash tst_hash.c -lext2fs -lcom_err -luuid -le2p

#include <stdio.h>
#include <unistd.h>
#include <stdlib.h>
#include <ctype.h>
#include <string.h>
#include <time.h>
#include <errno.h>
#include <sys/types.h>
#include <getopt.h>

#include "ext2fs/ext2fs.h"
#include "uuid/uuid.h"
#include "et/com_err.h"

#define SEED "87fd5d61-4612-4147-8bf5-a21948e7e909"

struct hash {
	int num;
	ext2_dirhash_t hash, minor_hash;

static EXT2_QSORT_TYPE hash_cmp(const void *a, const void *b)
	const struct hash *db_a =
		(const struct hash *) a;
	const struct hash *db_b =
		(const struct hash *) b;

	if (db_a->hash != db_b->hash)
		return (int) (db_a->hash - db_b->hash);
	return (int) (db_a->minor_hash - db_b->minor_hash);

main(int argc, char **argv)
	errcode_t errcode;
	ext2_dirhash_t	hash, minor_hash;
	int hash_alg = EXT2_HASH_TEA;
	char name[200], *tmp, prefix[100];
	unsigned char uuid[16];
	int thislen, i, c, quiet = 0, num_hashes = 300000;
	struct hash *hash_array;

	uuid_parse(SEED, uuid);
	prefix[0] = 0;

	while ((c = getopt(argc, argv, "s:a:n:qp:")) != EOF)
		switch (c) {
		case 's':
			uuid_parse(optarg, uuid);
		case 'a':
			hash_alg = e2p_string2hash(optarg);
			if (hash_alg < 0) {
				fprintf(stderr, "Invalid hash algorithm: %s\n",
		case 'n':
			num_hashes = strtoul (optarg, &tmp, 0);
			if (*tmp) {
				com_err (argv[0], 0, "count - %s", optarg);
		case 'p':
			if (strlen(optarg)+1 > sizeof(prefix)) {
				fprintf(stderr, "%s: prefix too large!\n",
			strcpy(prefix, optarg);
		case 'q':
			quiet = 1;
			fprintf(stderr, "Usage: %s [-q] [-s hash_seed] "
				"[-a hash_alg] [-n num_hashes]\n", argv[0]);

	hash_array = malloc(num_hashes * sizeof(struct hash));
	if (hash_array == NULL) {
		fprintf(stderr, "Couldn't allocate hash_array\n");
	for (i=0; i < num_hashes; i++) {
		sprintf(name, "%s%d", prefix, i);
		errcode = ext2fs_dirhash(hash_alg, name, strlen(name),
					 (__u32 *) uuid, 
		if (errcode) {
			com_err("ext2fs_dirhash", errcode, 
				"while trying to hash '%s'", name);
		hash_array[i].num = i;

	qsort(hash_array, (size_t) num_hashes, sizeof(struct hash), hash_cmp);

	for (c=0,i=0; i < num_hashes-1; i++) {
		if ((hash_array[i].hash == hash_array[i+1].hash) &&
		    (hash_array[i].minor_hash == hash_array[i+1].minor_hash)) {
			if (quiet)
			printf("hash collision: %d, %d: %08x:%08x\n", 
			       hash_array[i].num, hash_array[i+1].num,
			       hash_array[i].hash, hash_array[i].minor_hash);
	printf("%d collisions\n", c);


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]