How badly compressed are they? Problems in the codec, or in the data that
comes out of the codec?

We've had some lzo corruption problems, and so far have simply been dealing
with that by doing correctness tests in our log mover pipeline before moving
into the "data warehouse" area.

Skipping bad files silently seems like asking for trouble (at some point the
problem quietly grows and you wind up skipping most of your data), so I've
been avoiding putting something like that in so that when things are badly
broken, we get some early pain rather than lots of late pain.

D

On Tue, Jan 25, 2011 at 2:54 PM, Kim Vogt <k...@simplegeo.com> wrote:

> Hi,
>
> I'm processing gzipped compressed files in a directory, but some files are
> corrupted and can't be decompressed.  Is there a way to skip the bad files
> with a custom load func?
>
> -Kim
>

Reply via email to