Hive query result in sequence file

2012-01-24 Thread jingjung Ng
Hi, I have following hive query (pseudo hive query code) select name, address, phone from t1 join t2 Executing above query will end up file stored in the format of name, address, phone format on the fie system (hdfs or local). However I'd like to write to either to a sequence file (key: name,

Re: Wiki Write Access

2012-01-24 Thread Aniket Mokashi
Hi Carl, It would be helpful for me too. My wiki username: aniket486. Thanks, Aniket On Tue, Jan 24, 2012 at 11:57 AM, Carl Steinbach wrote: > Hi Matt, > > Great! > > Please sign up for a wiki account here: > https://cwiki.apache.org/confluence/signup.action > > Then email me your wiki usernam

Re: dropping an "external" table without deleting the data

2012-01-24 Thread Jerome Boulon
… And you may want to double check your table properties using: describe formatted ; /Jerome From: Steven Wong mailto:sw...@netflix.com>> Reply-To: mailto:user@hive.apache.org>> Date: Wed, 25 Jan 2012 00:27:26 + To: "user@hive.apache.org" mailto:user@hive.apache.

RE: dropping an "external" table without deleting the data

2012-01-24 Thread Steven Wong
You can change your table to external first and then drop it: ALTER TABLE my_table SET TBLPROPERTIES ('EXTERNAL'='FALSE'); Please test it on an unimportant table first. From: Igor Tatarinov [mailto:i...@decide.com] Sent: Tuesday, January 24, 2012 3:55 PM To: user@hive.apache.org Subject: droppi

dropping an "external" table without deleting the data

2012-01-24 Thread Igor Tatarinov
I keep making the same mistake of creating an external table without saying "EXTERNAL". As a result, as you can imagine, I am stuck with a table that I cannot drop or I'll lose the external (S3) data. I end up terminating my EMR instance and creating a new one to avoid dropping the table accidental

RE: derby related files show up when using MySQL

2012-01-24 Thread Mark Schramm (tetrascend)
Edward - Mystery solved, thanks! - Mark From: Edward Capriolo [mailto:edlinuxg...@gmail.com] Sent: Tuesday, January 24, 2012 11:17 AM To: user@hive.apache.org Subject: Re: derby related files show up when using MySQL This took me a while to figure out. Hive has this feature where it will a

Re: Hive FAQ

2012-01-24 Thread Carl Steinbach
Hi Nicolas, Thanks for pointing this out. If you would like to provide answers I can give you edit access to the wiki. Please create an account and send me your username. Thanks. Carl 2012/1/24 Nicolas Lalevée > It seems to be there: > https://cwiki.apache.org/confluence/display/Hive/User+FAQ

Re: Wiki Write Access

2012-01-24 Thread Carl Steinbach
Hi Matt, Great! Please sign up for a wiki account here: https://cwiki.apache.org/confluence/signup.action Then email me your wiki username and I will add you to the Hive wiki ACL. Thanks. Carl On Tue, Jan 24, 2012 at 7:10 AM, Tucker, Matt wrote: > Hi, > > ** ** > > I would like to get wr

Re: derby related files show up when using MySQL

2012-01-24 Thread Edward Capriolo
This took me a while to figure out. Hive has this feature where it will automatically collect table stats ONLY during 'insert overwrite' queries. This feature is still being worked on and is only useful for indexing. You should set this in your hive-site.xml hive.stats.autogather false

derby related files show up when using MySQL

2012-01-24 Thread Mark Schramm (tetrascend)
I have configured hive to use MySQL. This all seems to work fine, but occasionally derby.log files and TempStatsStore directories get created as if I was using the default derby database. Is this an indication that I have something misconfigured? Can anyone explain why this happens? Thanks,

RE: Performance problems with Hive script

2012-01-24 Thread Butani, Harish
Hi Benjamin, In case you are interested SQL Windowing(https://github.com/hbutani/SQLWindowing) is designed for these kinds of use cases. Your query would be expressed as: from < select symbol, dt, cast(close AS FLOAT) as close from raw_symbols > partition by symbol order by dt wit

Re: Question on bucketed map join

2012-01-24 Thread Avrilia Floratou
Hi Bejoy, Thanks a lot for your help:) I'm still a little confused. In my script I always have the set.auto.convert.join = true. Then I did what you suggested: 1st -> set hive.optimize.bucketmapjoin=true 2nd -> set hive.optimize.bucketmapjoin=false I ran explain with these 2 options and the

Hive FAQ

2012-01-24 Thread Nicolas Lalevée
It seems to be there: https://cwiki.apache.org/confluence/display/Hive/User+FAQ But humm..how should I phrase it. shouldn't questions be followed by answers ? :D Nicolas

Wiki Write Access

2012-01-24 Thread Tucker, Matt
Hi, I would like to get write access to the Hive wiki, so that I can add documentation on existing UDFs. Thanks Matt Tucker Associate eBusiness Analyst Walt Disney Parks and Resorts Online Ph: 407-566-2545 Tie: 8-296-2545

Re: JobTracker url shwoing less no of nodes available

2012-01-24 Thread hadoop hive
*now On Tue, Jan 24, 2012 at 3:37 PM, hadoop hive wrote: > hey thanks Alex, > > not its working fine . > > Regards > Vikas Srivastava > > > On Tue, Jan 24, 2012 at 1:58 PM, alo alt wrote: > >> +common user BCC >> >> please post to the correct mailing lists. Added common users. >> >> that mean t

Re: JobTracker url shwoing less no of nodes available

2012-01-24 Thread hadoop hive
hey thanks Alex, not its working fine . Regards Vikas Srivastava On Tue, Jan 24, 2012 at 1:58 PM, alo alt wrote: > +common user BCC > > please post to the correct mailing lists. Added common users. > > that mean that some DN daemons not running. FIrst place for that are the > logs of the DNs.

Embedding Hive

2012-01-24 Thread Tharindu Mathew
Hi, I was wondering whether it's possible to embed a Hive runtime inside another Java program. Basically, if the hadoop*xml are present in the classpath I expect it to submit the job to the Hadoop cluster, but if not it will default to local mode and execute the job. Thanks in advance. -- Rega

Re: JobTracker url shwoing less no of nodes available

2012-01-24 Thread alo alt
+common user BCC please post to the correct mailing lists. Added common users. that mean that some DN daemons not running. FIrst place for that are the logs of the DNs. What says that? - Alex -- Alexander Lorenz http://mapredit.blogspot.com On Jan 24, 2012, at 7:55 AM, hadoop hive wrote: >