Thanks for you attention.  The input of sink is 500, and there is no order
by and limit.

Jingsong Li <jingsongl...@gmail.com> 于 2020年3月6日周五 下午6:15写道:

> Hi faaron,
>
> For sink parallelism.
> - What is parallelism of the input of sink? The sink parallelism should be
> same.
> - Does you sql have order by or limit ?
> Flink batch sql not support range partition now, so it will use single
> parallelism to run order by.
>
> For the memory of taskmanager.
> There is manage memory option to configure.
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-master/ops/memory/mem_setup.html#managed-memory
>
> Best,
> Jingsong Lee
>
> On Fri, Mar 6, 2020 at 5:38 PM faaron zheng <faaronzh...@gmail.com> wrote:
>
>> Hi all,
>>
>> I am trying to use flink sql to run hive task. I use tEnv.sqlUpdate to
>> execute my sql which looks like "insert overtwrite ... select ...". But I
>> find the parallelism of sink is always 1, it's intolerable for large data.
>> Why it happens? Otherwise, Is there any guide to decide the memory of
>> taskmanager when I have two huge table to hashjoin, for example, each table
>> has several TB data?
>>
>> Thanks,
>> Faaron
>>
>
>
> --
> Best, Jingsong Lee
>

Reply via email to