query parameters in hive

2012-02-13 Thread Wojciech Langiewicz
x27;t find any clues about such feature. If you know how to pass parameters to hive queries, please let me know. -- Wojciech Langiewicz

Re: why 1 reducer on simple join?

2012-01-12 Thread Wojciech Langiewicz
ect count(*) from table2 group by column1 limit 10; query runs with 6 mappers and 3 reducers On Thu, Jan 12, 2012 at 6:09 PM, Wojciech Langiewicz wrote: What do you mean by "Select runs fine" - is it using number of reducers that you set? It might help if you could show actual query.

Re: why 1 reducer on simple join?

2012-01-12 Thread Wojciech Langiewicz
hu, Jan 12, 2012 at 6:00 PM, Wojciech Langiewicz wrote: Hello, Have you tried running only select, without creating table? What are results? How did you tried to set number of reducers? Have you used this: set mapred.reduce.tasks = xyz; How many mappers does this query use? On 12.01.2012 23:53, Koe

Re: why 1 reducer on simple join?

2012-01-12 Thread Wojciech Langiewicz
.col5 ); both tables are backed by multiple files / blocks / chunks -- Wojciech Langiewicz

Re: Hive query taking too much time

2011-12-08 Thread Wojciech Langiewicz
ake a look at-- https://issues.apache.org/jira/browse/HIVE-74 On Wed, Dec 7, 2011 at 9:05 PM, Savant, Keshav< keshav.c.sav...@fisglobal.com> wrote: You are right Wojciech Langiewicz, we did the same thing and posted my result yesterday. Now we are planning to do this using a shell script

Re: Hive query taking too much time

2011-12-07 Thread Wojciech Langiewicz
Hi, In this case it's much easier and faster to merge all files using this command: cat *.csv > output.csv hive -e "load data local inpath 'output.csv' into table $table" On 07.12.2011 07:00, Vikas Srivastava wrote: hey if u having the same col of all the files then you can easily merge by s

Re: Hive query taking too much time

2011-12-06 Thread Wojciech Langiewicz
ut it was not significant in my case. set mapred.reduce.tasks=xyz doesn't change that behavior, but for example using GROUP with COUNT works much faster. I hope this helps. -- Wojciech Langiewicz On 06.12.2011 12:00, Savant, Keshav wrote: Hi All, My setup is hadoop-0.20.203.0 hive-0.7

Re: Hive for large statistics tables?

2011-09-27 Thread Wojciech Langiewicz
Hello, I'm using Hive to query data like yours. In my case I have about 300 - 500GB data per day, so it is much larger. We use Flume to load data into Hive - data is rolled every day (this can be changed). Hive queries - ad-hoc or scheduled usually take at least 10-20s or more (possibly hours

Re: how to store EOL characters in String field?

2011-09-07 Thread Wojciech Langiewicz
tpow...@tealeaf.com Tealeaf Technology / http://www.tealeaf.com -Original Message- From: Wojciech Langiewicz [mailto:wlangiew...@gmail.com] Sent: Wednesday, September 07, 2011 4:31 AM To: user@hive.apache.org Subject: how to store EOL characters in String field? Hello, How can I store EOL characters in

how to store EOL characters in String field?

2011-09-07 Thread Wojciech Langiewicz
x27;\n' is loaded as next line, which is unreadable (all NULLs) Thanks for help -- Wojciech Langiewicz

Re: Hive 0.7 using only one mapper

2011-07-29 Thread Wojciech Langiewicz
HiveInputFormat in my case. Thanks again. -- Wojciech Langiewicz On 29.07.2011 05:43, Carl Steinbach wrote: Hi Wojciech, Vaibhav is correct. There's a configuration problem in the copy of hive-default.xml that ships with CDH3u1 which sets hive.input.format=CombineHiveInputFormat, but l

Hive 0.7 using only one mapper

2011-07-28 Thread Wojciech Langiewicz
ersion 0.5 everything worked correctly. I'm attaching my hive-site.xml: https://gist.github.com/531 I have tested also jobs with Pig, and those jobs use multiple mappers - so I guess this is a Hive issue. Thank you for all your help. -- Wojciech Langiewicz