> And again: the same row is correct if I export a small set of data, and
>incorrect if I export a large set - so I think that file/data size has
>something to do with this.
My Phoenix vs LLAP benchmark hit size related issues in ETL.
In my case, the tipping point was >1 hdfs block per CSV file.
around for this?
-Original Message-
From: Nicholas Hakobian [mailto:nicholas.hakob...@rallyhealth.com]
Sent: Thursday, January 28, 2016 3:15 PM
To: user@hive.apache.org
Subject: Re: "Create external table" nulling data from source table
Do you have any fields with embedded newline charac
Do you have any fields with embedded newline characters? If so,
certain hive output formats will parse the newline character as the
end of row, and when importing, chances are the missing fields (now
part of the next row) will be padded with nulls. This happens in Hive
as well if you are using a Te