I think it may not work in scenario where Hadoop security is enabled and each 
HCFS setup is configured differently, unless if there is a way to isolate the 
Hadoop configurations used in this case?

Regards,
Vijay

Sent from my iPhone

> On Aug 24, 2017, at 2:51 AM, Stephan Ewen <se...@apache.org> wrote:
> 
> Hi!
> 
> I think it can work if you fully qualify the URIs.
> 
> For the checkpoint configuration, specify one namenode (in the
> flink-config.yml or in the constructor of the state backend).
> Example:   statebackend.fs.checkpoint.dir:
> hdfs://dfsOneNamenode:port/flink/checkpoints
> 
> For the result (for example rolling sink), configure it with
> hdfs://dfsTwoNamenode:otherport/flink/result
> 
> Is that what you are looking for?
> 
> Stephan
> 
> 
> On Thu, Aug 24, 2017 at 11:47 AM, Stefan Richter <
> s.rich...@data-artisans.com> wrote:
> 
>> Hi,
>> 
>> I don’t think that this is currently supported. If you see a use case for
>> this (over creating different root directories for checkpoint data and
>> result data) then I suggest that you open a JIRA issue with a new feature
>> request.
>> 
>> Best,
>> Stefan
>> 
>>> Am 23.08.2017 um 20:17 schrieb Vijay Srinivasaraghavan <
>> vijikar...@yahoo.com>:
>>> 
>>> Hi Ted,
>>> 
>>> I believe HDFS-6584 is more of an HDFS feature supporting archive use
>> case through some policy configurations.
>>> 
>>> My ask is that I have two distinct HCFS File systems which are
>> independent but the Flink job will decide which one to use for sink while
>> the Flink infrastructure is by default configured with one of these HCFS as
>> state backend store.
>>> 
>>> Hope this helps.
>>> 
>>> Regards
>>> Vijay
>>> 
>>> 
>>> On Wednesday, August 23, 2017 11:06 AM, Ted Yu <yuzhih...@gmail.com>
>> wrote:
>>> 
>>> 
>>> Would HDFS-6584 help with your use case ?
>>> 
>>> On Wed, Aug 23, 2017 at 11:00 AM, Vijay Srinivasaraghavan <
>>> vijikar...@yahoo.com.invalid <mailto:vijikar...@yahoo.com.invalid>>
>> wrote:
>>> 
>>>> Hello,
>>>> Is it possible for a Flink cluster to use multiple HDFS repository
>> (HDFS-1
>>>> for managing Flink state backend, HDFS-2 for syncing results from user
>>>> job)?
>>>> The scenario can be viewed in the context of running some jobs that are
>>>> meant to push the results to an archive repository (cold storage).
>>>> Since the hadoop configuration is static, I am thinking it is hard to
>>>> achieve this but I could be wrong.
>>>> Please share any thoughts.
>>>> RegardsVijay
>>> 
>>> 
>> 
>> 

Reply via email to