How is concurrency achieved with this solution?

On Thursday, January 10, 2013, Qiang Wang wrote:

> I believe the HWI (Hive Web Interface) can give you a hand.
>
> https://github.com/anjuke/hwi
>
> You can use the HWI to submit and run queries concurrently.
> Partition management can be achieved by creating crontabs using the HWI.
>
> It's simple and easy to use. Hope it helps.
>
> Regards,
> Qiang
>
>
> 2013/1/11 Tom Brown <tombrow...@gmail.com <javascript:_e({}, 'cvml',
> 'tombrow...@gmail.com');>>
>
>> All,
>>
>> I want to automate jobs against Hive (using an external table with
>> ever growing partitions), and I'm running into a few challenges:
>>
>> Concurrency - If I run Hive as a thrift server, I can only safely run
>> one job at a time. As such, it seems like my best bet will be to run
>> it from the command line and setup a brand new instance for each job.
>> That quite a bit of a hassle to solves a seemingly common problem, so
>> I want to know if there are any accepted patterns or best practices
>> for this?
>>
>> Partition management - New partitions will be added regularly. If I
>> have to setup multiple instances of Hive for each (potentially)
>> overlapping job, it will be difficult to keep track of the partitions
>> that have been added. In the context of the preceding question, what
>> is the best way to add metadata about new partitions?
>>
>> Thanks in advance!
>>
>> --Tom
>>
>
>

Reply via email to