Hi Jagat,

Yes, curling or using wget will retrieve the contents of the Avro schema.
Just to be specific not the avsc file but its contents.

Thanks,
Ravi


On Sat, Jun 27, 2020 at 2:06 AM Jagat Singh <jagatsi...@gmail.com> wrote:

> Hello Ravi,
>
> When you wget this url
>
> wget http://<api_server>:9091/schema?name=ed&store=parquet&
> isMutated=true&table=ed&secbypass=testing'
>
> Do you get avsc file?
>
> Regards,
>
> Jagat Singh
>
> On Sat, 27 Jun 2020, 7:01 am ravi kanth, <ravi....@gmail.com> wrote:
>
>> Just want to follow up on the below email.
>>
>> Thanks,
>> Ravi
>>
>>
>> On Thu, Jun 25, 2020 at 5:37 PM ravi kanth <ravi....@gmail.com> wrote:
>>
>>> Hi Community,
>>>
>>> Hive Version: 3.1.2
>>>
>>> We are working on building a Hive Avro table on a few Avro files. I am
>>> able to succesfully create the table and query it when I have the Avro
>>> schema definition(avsc) file on hdfs with no issues.
>>>
>>> However, when trying to load the same schema from a rest API(as
>>> mentioned in
>>> https://cwiki.apache.org/confluence/display/Hive/AvroSerDe#AvroSerDe-CreatingAvro-backedHivetables),
>>> hive throws an exception and fails to create the table.
>>>
>>> *Sample table:*
>>> CREATE EXTERNAL TABLE ed_avro_1
>>> STORED AS AVRO
>>> LOCATION '/tmp/sample/yyyymmdd=20200206'
>>> TBLPROPERTIES ('avro.schema.literal'='http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing');
>>>
>>> When launched hive in INFO mode below is the trace of the problem which
>>> looks like Hive is interpreting the URL as a file name and throwing out a
>>> FileNotFoundException.
>>>
>>> I have tried using avro.schema.literal instead of avro.schema.url,
>>> however it turns out that hive is interpreting URL as a String and throwing
>>> a jackson parsing error.
>>>
>>> Can anyone help look into this? Is this a bug in Hive-3.1.2? Any details
>>> will be of great help.
>>>
>>> Thanks,
>>> Ravi
>>>
>>>
>>> StackTrace:
>>>
>>>> 2020-06-26T00:06:03,283 INFO [main]
>>>>> org.apache.hadoop.hive.conf.HiveConf - Using the default value passed in
>>>>> for log id: 646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>
>>>> 2020-06-26T00:06:03,283 INFO [main]
>>>>> org.apache.hadoop.hive.ql.session.SessionState - Updating thread name to
>>>>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 main
>>>>
>>>> 2020-06-26T00:06:03,286  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Compiling
>>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>>>>> CREATE EXTERNAL TABLE ed_avro_1
>>>>
>>>> STORED AS AVRO
>>>>
>>>> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>>>>
>>>> TBLPROPERTIES ('avro.schema.url'='http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>>>>
>>>> 2020-06-26T00:06:03,630  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Concurrency mode is disabled, not creating a lock manager
>>>>
>>>> 2020-06-26T00:06:03,638  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] parse.CalcitePlanner: Starting Semantic Analysis
>>>>
>>>> 2020-06-26T00:06:03,669  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] sqlstd.SQLStdHiveAccessController: Created 
>>>>> SQLStdHiveAccessController
>>>>> for session context : HiveAuthzSessionContext
>>>>> [sessionString=646da35b-84b0-43aa-9b68-5d668ebbfc36, clientType=HIVECLI]
>>>>
>>>> 2020-06-26T00:06:03,673 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] org.apache.hadoop.hive.ql.session.SessionState -
>>>>> METASTORE_FILTER_HOOK will be ignored, since
>>>>> hive.security.authorization.manager is set to instance of
>>>>> HiveAuthorizerFactory.
>>>>
>>>> 2020-06-26T00:06:03,673  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStoreClient: Mestastore configuration
>>>>> metastore.filter.hook changed from
>>>>> org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to
>>>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>>>>
>>>> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>>>>
>>>> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning
>>>>> up thread local RawStore...
>>>>
>>>> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>>>>
>>>> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done
>>>>> cleaning up thread local RawStore
>>>>
>>>> 2020-06-26T00:06:03,680  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: Opening raw store with implementation
>>>>> class:org.apache.hadoop.hive.metastore.ObjectStore
>>>>
>>>> 2020-06-26T00:06:03,680  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>>>>> unsupported value null . Setting it to value: ignored
>>>>
>>>> 2020-06-26T00:06:03,681  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: ObjectStore, initialize called
>>>>
>>>> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is 
>>>>> MYSQL
>>>>
>>>> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: Initialized ObjectStore
>>>>
>>>> 2020-06-26T00:06:03,692  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.RetryingMetaStoreClient: RetryingMetaStoreClient
>>>>> proxy=class org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
>>>>> ugi=hdfs (auth:SIMPLE) retries=1 delay=1 lifetime=0
>>>>
>>>> 2020-06-26T00:06:03,704  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] parse.CalcitePlanner: Creating table scratch.ed_avro_1 position=22
>>>>
>>>> 2020-06-26T00:06:03,719  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>>>>> unsupported value null . Setting it to value: ignored
>>>>
>>>> 2020-06-26T00:06:03,719  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: ObjectStore, initialize called
>>>>
>>>> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is 
>>>>> MYSQL
>>>>
>>>> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.ObjectStore: Initialized ObjectStore
>>>>
>>>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.RetryingMetaStoreClient: RetryingMetaStoreClient
>>>>> proxy=class org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
>>>>> ugi=hdfs (auth:SIMPLE) retries=1 delay=1 lifetime=0
>>>>
>>>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: get_database: @hive#scratch
>>>>
>>>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=get_database:
>>>>> @hive#scratch
>>>>
>>>> 2020-06-26T00:06:03,754  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Semantic Analysis Completed (retrial = false)
>>>>
>>>> 2020-06-26T00:06:03,762  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Returning Hive schema: Schema(fieldSchemas:null,
>>>>> properties:null)
>>>>
>>>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Completed compiling
>>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2);
>>>>> Time taken: 0.477 seconds
>>>>
>>>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] reexec.ReExecDriver: Execution #1 of query
>>>>
>>>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Concurrency mode is disabled, not creating a lock manager
>>>>
>>>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Executing
>>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>>>>> CREATE EXTERNAL TABLE ed_avro_1
>>>>
>>>> STORED AS AVRO
>>>>
>>>> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>>>>
>>>> TBLPROPERTIES ('avro.schema.url'='http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>>>>
>>>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] ql.Driver: Starting task [Stage-0:DDL] in serial mode
>>>>
>>>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStoreClient: Mestastore configuration
>>>>> metastore.filter.hook changed from
>>>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>>>>> to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl
>>>>
>>>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>>>>
>>>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning
>>>>> up thread local RawStore...
>>>>
>>>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>>>>
>>>> 2020-06-26T00:06:03,766  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done
>>>>> cleaning up thread local RawStore
>>>>
>>>> 2020-06-26T00:06:03,790  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>>>>> avro.schema.literal == null
>>>>
>>>> 2020-06-26T00:06:03,809  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>>>>> Returning signal schema to indicate problem
>>>>
>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>>>>> schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>>>
>>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> Caused by: java.io.FileNotFoundException: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at
>>>>> org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> ... 35 more
>>>>
>>>> 2020-06-26T00:06:03,871 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] hive.log: error in initSerDe:
>>>>> org.apache.hadoop.hive.serde2.SerDeException Encountered 
>>>>> AvroSerdeException
>>>>> determining schema. Returning signal schema to indicate problem: Unable to
>>>>> read schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> org.apache.hadoop.hive.serde2.SerDeException: Encountered
>>>>> AvroSerdeException determining schema. Returning signal schema to indicate
>>>>> problem: Unable to read schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:543)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>>>
>>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> 2020-06-26T00:06:03,874 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] metadata.Table: Unable to get field from serde:
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe
>>>>
>>>> java.lang.RuntimeException:
>>>>> MetaException(message:org.apache.hadoop.hive.serde2.SerDeException
>>>>> Encountered AvroSerdeException determining schema. Returning signal schema
>>>>> to indicate problem: Unable to read schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing)
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:291)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>>>
>>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> Caused by: org.apache.hadoop.hive.metastore.api.MetaException:
>>>>> org.apache.hadoop.hive.serde2.SerDeException Encountered 
>>>>> AvroSerdeException
>>>>> determining schema. Returning signal schema to indicate problem: Unable to
>>>>> read schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:98)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> ... 28 more
>>>>
>>>> 2020-06-26T00:06:03,874  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>>>>> avro.schema.literal == null
>>>>
>>>> 2020-06-26T00:06:03,878  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36
>>>>> main] avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>>>>> Returning signal schema to indicate problem
>>>>
>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>>>>> schema from given path: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:900)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>>>
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>>>
>>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> Caused by: java.io.FileNotFoundException: http://
>>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>>>
>>>> at
>>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>>>>> ~[?:1.8.0_252]
>>>>
>>>> at
>>>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at
>>>>> org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>>>>> ~[hadoop-common-3.1.2.jar:?]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> at
>>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>>>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>>>
>>>> ... 33 more
>>>>
>>>> Thanks,
>>> Ravi
>>>
>>

Reply via email to