OK, I guess I should contribute the solution I finally used.  The
problem at hand was a 34 MB mailing list archive containing lots of
news excerpts (about 10000 messages).  For an article I'm going to
write, I wanted to scan that archive for contributions concerning
certain topics.

My solution was this: First, using mutt, I stored that archive to a
maildir folder.  I then created a sub-directory hashed/, and used a
small shell script (horribly inefficient) to move the message files
into a certain number of subdirectories under hashed/.  That way,
file system overhead was under control again.  I then indexed the
hashed/ subdirectory using glimpseindex, and finally used the
attached script to create hard links to the cur sub-directory, which
could then be read using mutt.

Note that this script has a kludge to emulate glimpse -l - that's
because glimpse -l dumps core on my system.

-- 
http://www.guug.de/~roessler/
#!/bin/sh --

glimpse -H . -1 -i "$@" | cut -f1 -d: | while read filename ; do
        ln $filename cur/
done

Reply via email to