The HWI will create a cli session for each query through hive libs, so
several queries can run concurrently.


2013/1/11 Tom Brown <tombrow...@gmail.com>

> How is concurrency achieved with this solution?
>
>
> On Thursday, January 10, 2013, Qiang Wang wrote:
>
>> I believe the HWI (Hive Web Interface) can give you a hand.
>>
>> https://github.com/anjuke/hwi
>>
>> You can use the HWI to submit and run queries concurrently.
>> Partition management can be achieved by creating crontabs using the HWI.
>>
>> It's simple and easy to use. Hope it helps.
>>
>> Regards,
>> Qiang
>>
>>
>> 2013/1/11 Tom Brown <tombrow...@gmail.com>
>>
>>> All,
>>>
>>> I want to automate jobs against Hive (using an external table with
>>> ever growing partitions), and I'm running into a few challenges:
>>>
>>> Concurrency - If I run Hive as a thrift server, I can only safely run
>>> one job at a time. As such, it seems like my best bet will be to run
>>> it from the command line and setup a brand new instance for each job.
>>> That quite a bit of a hassle to solves a seemingly common problem, so
>>> I want to know if there are any accepted patterns or best practices
>>> for this?
>>>
>>> Partition management - New partitions will be added regularly. If I
>>> have to setup multiple instances of Hive for each (potentially)
>>> overlapping job, it will be difficult to keep track of the partitions
>>> that have been added. In the context of the preceding question, what
>>> is the best way to add metadata about new partitions?
>>>
>>> Thanks in advance!
>>>
>>> --Tom
>>>
>>
>>

Reply via email to