You too!

On Fri, Aug 26, 2016 at 4:15 PM, Niels Basjes <ni...@basjes.nl> wrote:
> Thanks!
> I'm going to work with this next week.
>
> Have a nice weekend.
>
> Niels
>
> On Fri, Aug 26, 2016 at 2:49 PM, Maximilian Michels <m...@apache.org> wrote:
>>
>> It is a bit more involved as I thought. We could simply the API further:
>>
>> import org.apache.flink.client.program.PackagedProgram;
>> import org.apache.flink.configuration.Configuration;
>> import org.apache.flink.configuration.GlobalConfiguration;
>> import org.apache.hadoop.fs.Path;
>>
>> import java.io.File;
>> import java.net.URL;
>> import java.util.Collections;
>>
>> public class Niels {
>>
>>    public static void main(String[] args) throws Exception {
>>
>>       final String basePath = "/Users/max/Dev/flink/build-target/";
>>       final String confDir = basePath + "/conf";
>>       final String flinkJarPath = basePath +
>> "/lib/flink-dist_2.10-1.2-SNAPSHOT.jar";
>>
>>       final PackagedProgram packagedProgram =
>>          new PackagedProgram(
>>             new File(basePath + "/examples/streaming/WordCount.jar"),
>>             // We need the Flink jar here because we want to lookup the
>> main method
>>             // which might contains dependencies to Flink which are not in
>> the user jar
>>             Collections.singletonList(new URL("file://" + flinkJarPath)));
>>
>>       final YarnClusterDescriptor descriptor = new
>> YarnClusterDescriptor();
>>
>>       Configuration configuration =
>> GlobalConfiguration.loadConfiguration(confDir);
>>       descriptor.setFlinkConfiguration(configuration);
>>
>>       descriptor.setConfigurationDirectory(confDir);
>>       descriptor.setConfigurationFilePath(new Path(confDir +
>> "/flink-conf.yaml"));
>>
>>       descriptor.setLocalJarPath(new Path(flinkJarPath));
>>       descriptor.setTaskManagerCount(2);
>>       descriptor.setName("Testing the YarnClusterClient");
>>
>>       final YarnClusterClient client = descriptor.deploy();
>>       client.run(packagedProgram, 2);
>>       client.shutdown();
>>    }
>> }
>>
>>
>> On Thu, Aug 25, 2016 at 5:06 PM, Niels Basjes <ni...@basjes.nl> wrote:
>>>
>>> Sounds good.
>>> Is there a basic example somewhere I can have a look at?
>>>
>>> Niels
>>>
>>> On Thu, Aug 25, 2016 at 2:55 PM, Maximilian Michels <m...@apache.org>
>>> wrote:
>>>>
>>>> Hi Niels,
>>>>
>>>> If you're using 1.1.1, then you can instantiate the
>>>> YarnClusterDescriptor and supply it with the Flink jar and
>>>> configuration and subsequently call `deploy()` on it to receive a
>>>> ClusterClient for Yarn which you can submit programs using the
>>>> `run(PackagedProgram program, String args)` method. You can also
>>>> cancel jobs or shutdown the cluster from the ClusterClient.
>>>>
>>>> Cheers,
>>>> Max
>>>>
>>>> On Thu, Aug 25, 2016 at 10:24 AM, Niels Basjes <ni...@basjes.nl> wrote:
>>>> > Hi,
>>>> >
>>>> > We have a situation where we need to start a flink batch job on a yarn
>>>> > cluster the moment an event arrives over a queue.
>>>> > These events occur at a very low rate (like once or twice a week).
>>>> >
>>>> > The idea we have is to run an application that listens to the queue
>>>> > and
>>>> > executes the batch when it receives a message.
>>>> >
>>>> > We found that if we start this using 'flink run -m yarn-cluster ..."
>>>> > the
>>>> > moment we run this the jobmanager in yarn is started and the resources
>>>> > for
>>>> > these batches is claimed immediately.
>>>> >
>>>> > What is the recommended way to only claim these resources when we
>>>> > actually
>>>> > have a job to run?
>>>> > Can we 'manually' start and stop the jobmanager in yarn in some way
>>>> > from our
>>>> > java code?
>>>> >
>>>> > --
>>>> > Best regards / Met vriendelijke groeten,
>>>> >
>>>> > Niels Basjes
>>>
>>>
>>>
>>>
>>> --
>>> Best regards / Met vriendelijke groeten,
>>>
>>> Niels Basjes
>>
>>
>
>
>
> --
> Best regards / Met vriendelijke groeten,
>
> Niels Basjes

Reply via email to