Hi Community,

Hive Version: 3.1.2

We are working on building a Hive Avro table on a few Avro files. I am able
to succesfully create the table and query it when I have the Avro schema
definition(avsc) file on hdfs with no issues.

However, when trying to load the same schema from a rest API(as mentioned
in
https://cwiki.apache.org/confluence/display/Hive/AvroSerDe#AvroSerDe-CreatingAvro-backedHivetables),
hive throws an exception and fails to create the table.

*Sample table:*
CREATE EXTERNAL TABLE ed_avro_1
STORED AS AVRO
LOCATION '/tmp/sample/yyyymmdd=20200206'
TBLPROPERTIES ('avro.schema.literal'='http://
<api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing');

When launched hive in INFO mode below is the trace of the problem which
looks like Hive is interpreting the URL as a file name and throwing out a
FileNotFoundException.

I have tried using avro.schema.literal instead of avro.schema.url, however
it turns out that hive is interpreting URL as a String and throwing a
jackson parsing error.

Can anyone help look into this? Is this a bug in Hive-3.1.2? Any details
will be of great help.

Thanks,
Ravi


StackTrace:

> 2020-06-26T00:06:03,283 INFO [main] org.apache.hadoop.hive.conf.HiveConf -
>> Using the default value passed in for log id:
>> 646da35b-84b0-43aa-9b68-5d668ebbfc36
>
> 2020-06-26T00:06:03,283 INFO [main]
>> org.apache.hadoop.hive.ql.session.SessionState - Updating thread name to
>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 main
>
> 2020-06-26T00:06:03,286  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Compiling
>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>> CREATE EXTERNAL TABLE ed_avro_1
>
> STORED AS AVRO
>
> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>
> TBLPROPERTIES ('avro.schema.url'='http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>
> 2020-06-26T00:06:03,630  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Concurrency mode is disabled, not creating a lock manager
>
> 2020-06-26T00:06:03,638  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> parse.CalcitePlanner: Starting Semantic Analysis
>
> 2020-06-26T00:06:03,669  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> sqlstd.SQLStdHiveAccessController: Created SQLStdHiveAccessController for
>> session context : HiveAuthzSessionContext
>> [sessionString=646da35b-84b0-43aa-9b68-5d668ebbfc36, clientType=HIVECLI]
>
> 2020-06-26T00:06:03,673 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> org.apache.hadoop.hive.ql.session.SessionState - METASTORE_FILTER_HOOK will
>> be ignored, since hive.security.authorization.manager is set to instance of
>> HiveAuthorizerFactory.
>
> 2020-06-26T00:06:03,673  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStoreClient: Mestastore configuration
>> metastore.filter.hook changed from
>> org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to
>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>
> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>
> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread
>> local RawStore...
>
> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>
> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up
>> thread local RawStore
>
> 2020-06-26T00:06:03,680  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: Opening raw store with implementation
>> class:org.apache.hadoop.hive.metastore.ObjectStore
>
> 2020-06-26T00:06:03,680  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>> unsupported value null . Setting it to value: ignored
>
> 2020-06-26T00:06:03,681  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: ObjectStore, initialize called
>
> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL
>
> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: Initialized ObjectStore
>
> 2020-06-26T00:06:03,692  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class
>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs
>> (auth:SIMPLE) retries=1 delay=1 lifetime=0
>
> 2020-06-26T00:06:03,704  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> parse.CalcitePlanner: Creating table scratch.ed_avro_1 position=22
>
> 2020-06-26T00:06:03,719  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>> unsupported value null . Setting it to value: ignored
>
> 2020-06-26T00:06:03,719  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: ObjectStore, initialize called
>
> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL
>
> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.ObjectStore: Initialized ObjectStore
>
> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class
>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs
>> (auth:SIMPLE) retries=1 delay=1 lifetime=0
>
> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: get_database: @hive#scratch
>
> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=get_database:
>> @hive#scratch
>
> 2020-06-26T00:06:03,754  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Semantic Analysis Completed (retrial = false)
>
> 2020-06-26T00:06:03,762  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Returning Hive schema: Schema(fieldSchemas:null, properties:null)
>
> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Completed compiling
>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2);
>> Time taken: 0.477 seconds
>
> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> reexec.ReExecDriver: Execution #1 of query
>
> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Concurrency mode is disabled, not creating a lock manager
>
> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Executing
>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>> CREATE EXTERNAL TABLE ed_avro_1
>
> STORED AS AVRO
>
> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>
> TBLPROPERTIES ('avro.schema.url'='http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>
> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> ql.Driver: Starting task [Stage-0:DDL] in serial mode
>
> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStoreClient: Mestastore configuration
>> metastore.filter.hook changed from
>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>> to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl
>
> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>
> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread
>> local RawStore...
>
> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>
> 2020-06-26T00:06:03,766  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up
>> thread local RawStore
>
> 2020-06-26T00:06:03,790  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>> avro.schema.literal == null
>
> 2020-06-26T00:06:03,809  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>> Returning signal schema to indicate problem
>
> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>> schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> ~[?:1.8.0_252]
>
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>
> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>> ~[hadoop-common-3.1.2.jar:?]
>
> Caused by: java.io.FileNotFoundException: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>> ~[?:1.8.0_252]
>
> at
>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>> ~[?:1.8.0_252]
>
> at
>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> ... 35 more
>
> 2020-06-26T00:06:03,871 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> hive.log: error in initSerDe: org.apache.hadoop.hive.serde2.SerDeException
>> Encountered AvroSerdeException determining schema. Returning signal schema
>> to indicate problem: Unable to read schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> org.apache.hadoop.hive.serde2.SerDeException: Encountered
>> AvroSerdeException determining schema. Returning signal schema to indicate
>> problem: Unable to read schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:543)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> ~[?:1.8.0_252]
>
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>
> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>> ~[hadoop-common-3.1.2.jar:?]
>
> 2020-06-26T00:06:03,874 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> metadata.Table: Unable to get field from serde:
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe
>
> java.lang.RuntimeException:
>> MetaException(message:org.apache.hadoop.hive.serde2.SerDeException
>> Encountered AvroSerdeException determining schema. Returning signal schema
>> to indicate problem: Unable to read schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing)
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:291)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> ~[?:1.8.0_252]
>
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>
> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>> ~[hadoop-common-3.1.2.jar:?]
>
> Caused by: org.apache.hadoop.hive.metastore.api.MetaException:
>> org.apache.hadoop.hive.serde2.SerDeException Encountered AvroSerdeException
>> determining schema. Returning signal schema to indicate problem: Unable to
>> read schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:98)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> ... 28 more
>
> 2020-06-26T00:06:03,874  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>> avro.schema.literal == null
>
> 2020-06-26T00:06:03,878  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>> avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>> Returning signal schema to indicate problem
>
> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>> schema from given path: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:900)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>> ~[hive-cli-3.1.2.jar:3.1.2]
>
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> ~[?:1.8.0_252]
>
> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> ~[?:1.8.0_252]
>
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>
> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>> ~[hadoop-common-3.1.2.jar:?]
>
> Caused by: java.io.FileNotFoundException: http://
>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>
> at
>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>> ~[?:1.8.0_252]
>
> at
>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>> ~[?:1.8.0_252]
>
> at
>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>> ~[hadoop-common-3.1.2.jar:?]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> at
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>> ~[hive-exec-3.1.2.jar:3.1.2]
>
> ... 33 more
>
> Thanks,
Ravi

Reply via email to