Hi Samarth,

it definitely looks a bug to me.
I'm not sure there's a workaround for this problem though.

Jihoon

On Mon, Apr 15, 2019 at 1:39 PM Samarth Jain <samarth.j...@gmail.com> wrote:

> Hi,
>
> We are building out a realtime ingestion pipeline using Kafka Indexing
> service for Druid. In order to achieve better rollup, I was trying out the
> hadoop based reingestion job
> http://druid.io/docs/latest/ingestion/update-existing-data.html  which
> basically uses the datasource itself as the input.
>
> When I ran the job, it failed because it was trying to read segment
> metadata from druid_segments table and not from the table,
> customprefix_segments, I specified in the metadataUpdateSpec.
>
> "metadataUpdateSpec": {
>       "connectURI": "jdbc:mysql...",
>       "password": "XXXXXXX",
>       "segmentTable": "customprefix_segments",
>       "type": "mysql",
>       "user": "XXXXXXXX"
> },
>
> Looking at the code, I see that the segmentTable specified in the spec is
> actually passed in as pending_segments table (3rd param is for
> pending_segments and 4th param is for segments table)
>
> https://github.com/apache/incubator-druid/blob/master/indexing-hadoop/src/main/java/org/apache/druid/indexer/updater/MetadataStorageUpdaterJobSpec.java#L92
>
> As a result, the re-ingestion job tries to read from the default segments
> table named DRUID_SEGMENTS which isn't present.
>
> Is this intentional or a bug?
>
> Is there a way to configure the segments table name for this kind of
> re-ingestion job?
>

Reply via email to