A couple of days back, Erik Sammer at the Hadoop Hands On Lab at the Cloudera Sessions demonstrated how to achieve dynamic partitioning using Flume and created those partitioned directories on HDFS which are then readily usable by Hive
Understanding what I can from the two lines of your mail below, I would configure Flume to do dynamic partitioning (YEAR, MONTH, DAY, HOUR) and create those directories in HDFS and then create Hive tables with those partitions and run the queries As Stephen said earlier , experiment like crazy - and share please - it will make all of us better as well ! Thanks sanjay From: ch huang <justlo...@gmail.com<mailto:justlo...@gmail.com>> Reply-To: "user@hive.apache.org<mailto:user@hive.apache.org>" <user@hive.apache.org<mailto:user@hive.apache.org>> Date: Thursday, September 12, 2013 6:55 PM To: "user@hive.apache.org<mailto:user@hive.apache.org>" <user@hive.apache.org<mailto:user@hive.apache.org>> Subject: question about partition table in hive hi,all: i use flume collect log data and put it in hdfs ,i want to use hive to do some caculate, query based on timerange,i want to use parttion table , but the data file in hdfs is a big file ,how can i put it into pratition table in hive? CONFIDENTIALITY NOTICE ====================== This email message and any attachments are for the exclusive use of the intended recipient(s) and may contain confidential and privileged information. Any unauthorized review, use, disclosure or distribution is prohibited. If you are not the intended recipient, please contact the sender by reply email and destroy all copies of the original message along with any attachments, from your computer system. If you are the intended recipient, please be advised that the content of this message is subject to access, review and disclosure by the sender's Email System Administrator.