A couple of days back, Erik Sammer at the Hadoop Hands On Lab at the Cloudera 
Sessions demonstrated how to achieve dynamic partitioning using Flume and 
created those partitioned directories on HDFS which are then readily usable by 
Hive

Understanding what I can from the two lines of your mail below, I would 
configure Flume to do dynamic partitioning (YEAR, MONTH, DAY, HOUR) and create 
those directories in HDFS and then create Hive tables with  those partitions 
and run the queries

As Stephen said earlier , experiment like crazy - and share please - it will 
make all of us better as well !


Thanks

sanjay

From: ch huang <justlo...@gmail.com<mailto:justlo...@gmail.com>>
Reply-To: "user@hive.apache.org<mailto:user@hive.apache.org>" 
<user@hive.apache.org<mailto:user@hive.apache.org>>
Date: Thursday, September 12, 2013 6:55 PM
To: "user@hive.apache.org<mailto:user@hive.apache.org>" 
<user@hive.apache.org<mailto:user@hive.apache.org>>
Subject: question about partition table in hive

hi,all:
        i use flume collect log data and put it in hdfs ,i want to use hive to 
do some caculate, query based on timerange,i want to use parttion table ,
but the data file in hdfs is a big file ,how can i put it into pratition table 
in hive?

CONFIDENTIALITY NOTICE
======================
This email message and any attachments are for the exclusive use of the 
intended recipient(s) and may contain confidential and privileged information. 
Any unauthorized review, use, disclosure or distribution is prohibited. If you 
are not the intended recipient, please contact the sender by reply email and 
destroy all copies of the original message along with any attachments, from 
your computer system. If you are the intended recipient, please be advised that 
the content of this message is subject to access, review and disclosure by the 
sender's Email System Administrator.

Reply via email to