Re: pig,hive install over hadoop

2014-04-01 Thread Shengjun Xin
Install hive and pig completely and configure hive and pig to use the existing hadoop cluster On Tue, Apr 1, 2014 at 6:17 PM, Chinna Rao Lalam wrote: > Hi, > > You need to install hive completely because both are different > processes(Installing hive is very simple and it require few > configur

Deserializing into multiple records

2014-04-01 Thread David Quigley
We are currently streaming complex documents to hdfs with the hope of being able to query. Each single document logically breaks down into a set of individual records. In order to use Hive, we preprocess each input document into a set of discreet records, which we save on HDFS and create an externa

Re: regarding HIVE-6469

2014-04-01 Thread Jacky
Yes, I agree, hive by design provides a feature to put stuff in trash while deleting it. but I found some people wanting this feature to skipTrash when necessary (by providing extra token on cmd line). and as you pointed out, I feel the same, there is no reason not to have this feature. Thanks Jay

Re: regarding HIVE-6469

2014-04-01 Thread Edward Capriolo
Due to the nature of hadoop most people usually have substantial overhead, and there trash is set accordingly to prevent accidental data loss. That being said there is no reason not to have this feature. On Tue, Apr 1, 2014 at 3:20 PM, Jacky wrote: > Hi, > > what do you guys think about this s

regarding HIVE-6469

2014-04-01 Thread Jacky
Hi, what do you guys think about this small feature : https://issues.apache.org/jira/browse/HIVE-6469 Please comment. Thanks Jay

RE: FAILED: SemanticException [Error 10001]: Table not found answer

2014-04-01 Thread Anthony Hsu
Hi Rishabh, You can get more logging info, including the full stack trace, by changing the logging level when starting Hive: hive -hiveconf hive.root.logger=INFO,console Cheers, Anthony From: Rishabh Bhardwaj [rbnex...@yahoo.com] Sent: Monday, March 31, 2014 11:5

bugs in 0.12 version

2014-04-01 Thread Lior Schachter
Hi all, We are randomly getting 2 types of exceptions while inserting data to hive. Seems like we encountered : HIVE-6114 , HIVE-6113 issues. Both issues are critical but there is no patch nor worka

Re: pig,hive install over hadoop

2014-04-01 Thread Chinna Rao Lalam
Hi, You need to install hive completely because both are different processes(Installing hive is very simple and it require few configurations). Hive process to work, it required Hadoop processes. I think same with the pig also(I did't try)... Hope It Helps, Chinna.. On Tue, Apr 1, 2014 at 3:3

Re: pig,hive install over hadoop

2014-04-01 Thread Nitin Pawar
Pig and hive do not come in bare minimum version. Its complete pig or hive package You can use existing hadoop cluster with pig and hive. If you do not need persistent storage for hive tables, then you dont need to configure much . Search for hive with derby and that should get you started. On pi

pig,hive install over hadoop

2014-04-01 Thread Rahul Singh
Hi, I have installed and configured hadoop. Now, I want to install hive and pig, As per my understanding pig and hive internally uses hadoop. So is there a way i can just install bare minimum hive or pig and take advantage of already installed hadoop or i need to separately install and configure

Re: READING FILE FROM MONGO DB

2014-04-01 Thread Nitin Pawar
you can always write customUDF for your needs On Tue, Apr 1, 2014 at 1:35 PM, Swagatika Tripathy wrote: > Do we hv a for loop concept in hive to iterate through the array elements > n display them. We need an alternative for explode method > Well you cN use Json serde for this > > Sent from my i

Re: READING FILE FROM MONGO DB

2014-04-01 Thread Swagatika Tripathy
Do we hv a for loop concept in hive to iterate through the array elements n display them. We need an alternative for explode method Well you cN use Json serde for this Sent from my iPhone On Mar 26, 2014, at 8:40 PM, "Swagatika Tripathy" wrote: Hi , The use case is we have some unstructured dat

Re: Hive vs Pig against number of files spawned

2014-04-01 Thread Navis류승우
try hive.hadoop.supports.splittable.combineinputformat=true; Thanks, Navis 2014-04-01 15:55 GMT+09:00 Sreenath : > Hi all, > I have a partitioned table in hive where each partition will have 630 gzip > compressed files each of average size 100kb. If I query over these files > using hive it will g