If you don't care to store those characters then maybe you should change
the database character set to use SQL_ASCII instead of UTF-8.  I believe
ASCII will quietly discard those characters when converting from UTF-8.

Mike

 On Wed, 2006-03-15 at 23:20 -0600, Wes wrote:
> Based on a couple of other posts, I think I'm out of luck, but I'm hoping
> something might have changed recently.
> 
> I'm loading a very high volume of data with COPY using libpq - about 100+
> million rows per day. The problem is that the COPY sometimes aborts with
> 
>   invalid UTF-8 byte sequence detected
> 
> I'd prefer not to add to the overhead by pre-validating every string, since
> PostgreSQL validates it already.  Is there a way to get the server to
> replace invalid characters with something like blank instead of generating a
> fatal error?
> 
> Wes
> 
> 
> 
> ---------------------------(end of broadcast)---------------------------
> TIP 1: if posting/reading through Usenet, please send an appropriate
>        subscribe-nomail command to [EMAIL PROTECTED] so that your
>        message can get through to the mailing list cleanly

---------------------------(end of broadcast)---------------------------
TIP 9: In versions below 8.0, the planner will ignore your desire to
       choose an index scan if your joining column's datatypes do not
       match

Reply via email to