Have you looked at SPARK-19888 ?

Please give the full stack trace of the exception you saw.

Cheers

On Mon, Feb 12, 2018 at 12:38 PM, Mina Aslani <aslanim...@gmail.com> wrote:

> Hi Matthias,
> Are you referring to https://issues.apache.org/jira/browse/SPARK-19976?
> Doesn't look like that the jira was not fixed. (e.g. Resolution: "Not a
> Problem").
> So, is there any suggested workaround?
>
> Regards,
> Mina
>
> On Mon, Feb 12, 2018 at 3:03 PM, Matthias J. Sax <matth...@confluent.io>
> wrote:
>
> > AFAIK, Spark does not pass this config to the consumer on purpose...
> > It's not a Kafka issues -- IIRC, there is Spark JIRA ticket for this.
> >
> > -Matthias
> >
> > On 2/12/18 11:04 AM, Mina Aslani wrote:
> > > Hi,
> > >
> > > I am getting below error
> > > Caused by: org.apache.kafka.clients.consumer.
> OffsetOutOfRangeException:
> > > Offsets out of range with no configured reset policy for partitions:
> > > {topic1-0=304337}
> > > as soon as I submit a spark app to my cluster.
> > >
> > > I am using below dependency
> > > name: 'spark-streaming-kafka-0-10_2.11', version: '2.2.0' And setting
> > the
> > > consumer's reset config(e.g. AUTO_OFFSET_RESET_CONFIG) to "earliest".
> > > As per https://kafka.apache.org/0110/documentation.html the exception
> > > should be thrown only when the consumer's reset config has not been set
> > > (e.g. default=none).
> > > Wondering what is the cause and how to fix.
> > >
> > > Best regards,
> > > Mina
> > >
> >
> >
>

Reply via email to