Hi Prash,
Try this:
create external table mslog
(
time_stamp string,
seq string
) row format delimited fields terminated by '\t' stored as textfile location
's3://your/bucket/path/'
;
Important: your s3 bucket can only contain files that have the same schema
format. Hive doesn't like
Can you try using using streaming from Hive?
http://wiki.apache.org/hadoop/Hive/LanguageManual/Transform
Cheers,
~Avram
On Apr 7, 2011, at 10:28 AM, Henry Liu wrote:
>
> I'd like to report some counters in my customized map and reduce ruby scripts
> for hive, just like hadoop strea
You can probably do this in Bash now, try something like this:
zz=`hive -S -e "select count(1) from t"`
~Avram
~ Avram Aelony|Sr. Analyst|eHarmony.com|(424) 258-1199|x1098|skype:
avram.aelony ~
On Mar 21, 2011, at 3:55 PM, Sameer Kalburgi wrote:
> Any plans, if possible, to
mentation or an example so I can give it
a try?
Many thanks,
Avram
On Jan 24, 2011, at 3:03 PM, Edward Capriolo wrote:
> On Mon, Jan 24, 2011 at 5:58 PM, Avram Aelony
> wrote:
>> Hi,
>>
>> I really like the virtual column feature in 0.7 that allows me to request
&g
Hi,
I really like the virtual column feature in 0.7 that allows me to request
INPUT__FILE__NAME and see the names of files that are being acted on.
Because I can see the files that are being read, I see that I am spending time
querying many, many very large files, most of which I do not need