Hi Fabian,

After a bit more documentation reading I have a better understanding of how
InputFormat interface works.
Indeed I've better to wrap a custom InputFormat implementation in my source.
This article helps a lot
https://brewing.codes/2017/02/06/implementing-flink-batch-data-connector/

connect() will be for a next sprint

All the best

François

Le ven. 15 févr. 2019 à 09:37, Fabian Hueske <fhue...@gmail.com> a écrit :

> H François,
>
> The TableEnvironment.connect() method can only be used if you provide
> (quite a bit) more code.
> It requires a TableSourceFactory and handling of all the properties that
> are defined in the other builder methods. See [1].
>
> I would recommend to either register the BatchTableSource directly
> (tEnv.registerTableSource()) or get a DataSet (via env.createSource()) and
> register the DataSet as a Table (tEnv.registerDataSet()).
>
> Best, Fabian
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-release-1.7/dev/table/sourceSinks.html#define-a-tablefactory
>
>
> Am Mo., 11. Feb. 2019 um 21:09 Uhr schrieb françois lacombe <
> francois.laco...@dcbrain.com>:
>
>> Hi Fabian,
>>
>> I've got issues for a custom InputFormat implementation with my existing
>> code.
>>
>> Is this can be used in combination with a BatchTableSource custom source?
>> As I understand your solution, I should move my source to implementations
>> like :
>>
>> tableEnvironment
>>   .connect(...)
>>   .withFormat(...)
>>   .withSchema(...)
>>   .inAppendMode()
>>   .registerTableSource("MyTable")
>>
>> right?
>>
>> I currently have a BatchTableSource class which produce a DataSet<Row>
>> from a single geojson file.
>> This doesn't sound compatible with a custom InputFormat, don't you?
>>
>> Thanks in advance for any addition hint, all the best
>>
>> François
>>
>> Le lun. 4 févr. 2019 à 12:10, Fabian Hueske <fhue...@gmail.com> a écrit :
>>
>>> Hi,
>>>
>>> The files will be read in a streaming fashion.
>>> Typically files are broken down into processing splits that are
>>> distributed to tasks for reading.
>>> How a task reads a file split depends on the implementation, but usually
>>> the format reads the split as a stream and does not read the split as a
>>> whole before emitting records.
>>>
>>> Best,
>>> Fabian
>>>
>>> Am Mo., 4. Feb. 2019 um 12:06 Uhr schrieb françois lacombe <
>>> francois.laco...@dcbrain.com>:
>>>
>>>> Hi Fabian,
>>>>
>>>> Thank you for this input.
>>>> This is interesting.
>>>>
>>>> With such an input format, will all the file will be loaded in memory
>>>> before to be processed or will all be streamed?
>>>>
>>>> All the best
>>>> François
>>>>
>>>> Le mar. 29 janv. 2019 à 22:20, Fabian Hueske <fhue...@gmail.com> a
>>>> écrit :
>>>>
>>>>> Hi,
>>>>>
>>>>> You can point a file-based input format to a directory and the input
>>>>> format should read all files in that directory.
>>>>> That works as well for TableSources that are internally use file-based
>>>>> input formats.
>>>>> Is that what you are looking for?
>>>>>
>>>>> Best, Fabian
>>>>>
>>>>> Am Mo., 28. Jan. 2019 um 17:22 Uhr schrieb françois lacombe <
>>>>> francois.laco...@dcbrain.com>:
>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>> I'm wondering if it's possible and what's the best way to achieve the
>>>>>> loading of multiple files with a Json source to a JDBC sink ?
>>>>>> I'm running Flink 1.7.0
>>>>>>
>>>>>> Let's say I have about 1500 files with the same structure (same
>>>>>> format, schema, everything) and I want to load them with a *batch* job
>>>>>> Can Flink handle the loading of one and each file in a single source
>>>>>> and send data to my JDBC sink?
>>>>>> I wish I can provide the URL of the directory containing my thousand
>>>>>> files to the batch source to make it load all of them sequentially.
>>>>>> My sources and sinks are currently available for BatchTableSource, I
>>>>>> guess the cost to make them available for streaming would be quite
>>>>>> expensive for me for the moment.
>>>>>>
>>>>>> Have someone ever done this?
>>>>>> Am I wrong to expect doing so with a batch job?
>>>>>>
>>>>>> All the best
>>>>>>
>>>>>> François Lacombe
>>>>>>
>>>>>>
>>>>>> <http://www.dcbrain.com/>
>>>>>> <https://twitter.com/dcbrain_feed?lang=fr>
>>>>>> <https://www.linkedin.com/company/dcbrain>
>>>>>> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>>>>>>
>>>>>> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que
>>>>>> si nécessaire
>>>>>>
>>>>>
>>>>
>>>> <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>
>>>>    <https://www.linkedin.com/company/dcbrain>
>>>> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>>>>
>>>> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que si
>>>> nécessaire
>>>>
>>>
>>
>> <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>
>> <https://www.linkedin.com/company/dcbrain>
>> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>>
>> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que si
>> nécessaire
>>
>

-- 

 <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>   
<https://www.linkedin.com/company/dcbrain>   
<https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>


 Pensez à la 
planète, imprimer ce papier que si nécessaire 

Reply via email to