Hi Jagat, Yes, curling or using wget will retrieve the contents of the Avro schema. Just to be specific not the avsc file but its contents.
Thanks, Ravi On Sat, Jun 27, 2020 at 2:06 AM Jagat Singh <jagatsi...@gmail.com> wrote: > Hello Ravi, > > When you wget this url > > wget http://<api_server>:9091/schema?name=ed&store=parquet& > isMutated=true&table=ed&secbypass=testing' > > Do you get avsc file? > > Regards, > > Jagat Singh > > On Sat, 27 Jun 2020, 7:01 am ravi kanth, <ravi....@gmail.com> wrote: > >> Just want to follow up on the below email. >> >> Thanks, >> Ravi >> >> >> On Thu, Jun 25, 2020 at 5:37 PM ravi kanth <ravi....@gmail.com> wrote: >> >>> Hi Community, >>> >>> Hive Version: 3.1.2 >>> >>> We are working on building a Hive Avro table on a few Avro files. I am >>> able to succesfully create the table and query it when I have the Avro >>> schema definition(avsc) file on hdfs with no issues. >>> >>> However, when trying to load the same schema from a rest API(as >>> mentioned in >>> https://cwiki.apache.org/confluence/display/Hive/AvroSerDe#AvroSerDe-CreatingAvro-backedHivetables), >>> hive throws an exception and fails to create the table. >>> >>> *Sample table:* >>> CREATE EXTERNAL TABLE ed_avro_1 >>> STORED AS AVRO >>> LOCATION '/tmp/sample/yyyymmdd=20200206' >>> TBLPROPERTIES ('avro.schema.literal'='http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing'); >>> >>> When launched hive in INFO mode below is the trace of the problem which >>> looks like Hive is interpreting the URL as a file name and throwing out a >>> FileNotFoundException. >>> >>> I have tried using avro.schema.literal instead of avro.schema.url, >>> however it turns out that hive is interpreting URL as a String and throwing >>> a jackson parsing error. >>> >>> Can anyone help look into this? Is this a bug in Hive-3.1.2? Any details >>> will be of great help. >>> >>> Thanks, >>> Ravi >>> >>> >>> StackTrace: >>> >>>> 2020-06-26T00:06:03,283 INFO [main] >>>>> org.apache.hadoop.hive.conf.HiveConf - Using the default value passed in >>>>> for log id: 646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>> >>>> 2020-06-26T00:06:03,283 INFO [main] >>>>> org.apache.hadoop.hive.ql.session.SessionState - Updating thread name to >>>>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 main >>>> >>>> 2020-06-26T00:06:03,286 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Compiling >>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2): >>>>> CREATE EXTERNAL TABLE ed_avro_1 >>>> >>>> STORED AS AVRO >>>> >>>> LOCATION '/tmp/event_detail/yyyymmdd=20200206' >>>> >>>> TBLPROPERTIES ('avro.schema.url'='http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing') >>>> >>>> 2020-06-26T00:06:03,630 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Concurrency mode is disabled, not creating a lock manager >>>> >>>> 2020-06-26T00:06:03,638 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] parse.CalcitePlanner: Starting Semantic Analysis >>>> >>>> 2020-06-26T00:06:03,669 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] sqlstd.SQLStdHiveAccessController: Created >>>>> SQLStdHiveAccessController >>>>> for session context : HiveAuthzSessionContext >>>>> [sessionString=646da35b-84b0-43aa-9b68-5d668ebbfc36, clientType=HIVECLI] >>>> >>>> 2020-06-26T00:06:03,673 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] org.apache.hadoop.hive.ql.session.SessionState - >>>>> METASTORE_FILTER_HOOK will be ignored, since >>>>> hive.security.authorization.manager is set to instance of >>>>> HiveAuthorizerFactory. >>>> >>>> 2020-06-26T00:06:03,673 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStoreClient: Mestastore configuration >>>>> metastore.filter.hook changed from >>>>> org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to >>>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook >>>> >>>> 2020-06-26T00:06:03,675 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: Cleaning up thread local RawStore... >>>> >>>> 2020-06-26T00:06:03,675 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning >>>>> up thread local RawStore... >>>> >>>> 2020-06-26T00:06:03,676 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore >>>> >>>> 2020-06-26T00:06:03,676 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done >>>>> cleaning up thread local RawStore >>>> >>>> 2020-06-26T00:06:03,680 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: Opening raw store with implementation >>>>> class:org.apache.hadoop.hive.metastore.ObjectStore >>>> >>>> 2020-06-26T00:06:03,680 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to >>>>> unsupported value null . Setting it to value: ignored >>>> >>>> 2020-06-26T00:06:03,681 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: ObjectStore, initialize called >>>> >>>> 2020-06-26T00:06:03,691 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is >>>>> MYSQL >>>> >>>> 2020-06-26T00:06:03,691 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: Initialized ObjectStore >>>> >>>> 2020-06-26T00:06:03,692 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.RetryingMetaStoreClient: RetryingMetaStoreClient >>>>> proxy=class org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient >>>>> ugi=hdfs (auth:SIMPLE) retries=1 delay=1 lifetime=0 >>>> >>>> 2020-06-26T00:06:03,704 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] parse.CalcitePlanner: Creating table scratch.ed_avro_1 position=22 >>>> >>>> 2020-06-26T00:06:03,719 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to >>>>> unsupported value null . Setting it to value: ignored >>>> >>>> 2020-06-26T00:06:03,719 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: ObjectStore, initialize called >>>> >>>> 2020-06-26T00:06:03,730 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is >>>>> MYSQL >>>> >>>> 2020-06-26T00:06:03,730 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.ObjectStore: Initialized ObjectStore >>>> >>>> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.RetryingMetaStoreClient: RetryingMetaStoreClient >>>>> proxy=class org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient >>>>> ugi=hdfs (auth:SIMPLE) retries=1 delay=1 lifetime=0 >>>> >>>> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: get_database: @hive#scratch >>>> >>>> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=get_database: >>>>> @hive#scratch >>>> >>>> 2020-06-26T00:06:03,754 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Semantic Analysis Completed (retrial = false) >>>> >>>> 2020-06-26T00:06:03,762 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Returning Hive schema: Schema(fieldSchemas:null, >>>>> properties:null) >>>> >>>> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Completed compiling >>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2); >>>>> Time taken: 0.477 seconds >>>> >>>> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] reexec.ReExecDriver: Execution #1 of query >>>> >>>> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Concurrency mode is disabled, not creating a lock manager >>>> >>>> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Executing >>>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2): >>>>> CREATE EXTERNAL TABLE ed_avro_1 >>>> >>>> STORED AS AVRO >>>> >>>> LOCATION '/tmp/event_detail/yyyymmdd=20200206' >>>> >>>> TBLPROPERTIES ('avro.schema.url'='http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing') >>>> >>>> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] ql.Driver: Starting task [Stage-0:DDL] in serial mode >>>> >>>> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStoreClient: Mestastore configuration >>>>> metastore.filter.hook changed from >>>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook >>>>> to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl >>>> >>>> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: Cleaning up thread local RawStore... >>>> >>>> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning >>>>> up thread local RawStore... >>>> >>>> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore >>>> >>>> 2020-06-26T00:06:03,766 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done >>>>> cleaning up thread local RawStore >>>> >>>> 2020-06-26T00:06:03,790 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] avro.AvroSerDe: AvroSerde::initialize(): Preset value of >>>>> avro.schema.literal == null >>>> >>>> 2020-06-26T00:06:03,809 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] avro.AvroSerDe: Encountered AvroSerdeException determining schema. >>>>> Returning signal schema to indicate problem >>>> >>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read >>>>> schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> ~[?:1.8.0_252] >>>> >>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >>>> >>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> Caused by: java.io.FileNotFoundException: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at >>>>> org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> ... 35 more >>>> >>>> 2020-06-26T00:06:03,871 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] hive.log: error in initSerDe: >>>>> org.apache.hadoop.hive.serde2.SerDeException Encountered >>>>> AvroSerdeException >>>>> determining schema. Returning signal schema to indicate problem: Unable to >>>>> read schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> org.apache.hadoop.hive.serde2.SerDeException: Encountered >>>>> AvroSerdeException determining schema. Returning signal schema to indicate >>>>> problem: Unable to read schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:543) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> ~[?:1.8.0_252] >>>> >>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >>>> >>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> 2020-06-26T00:06:03,874 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] metadata.Table: Unable to get field from serde: >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe >>>> >>>> java.lang.RuntimeException: >>>>> MetaException(message:org.apache.hadoop.hive.serde2.SerDeException >>>>> Encountered AvroSerdeException determining schema. Returning signal schema >>>>> to indicate problem: Unable to read schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing) >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:291) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> ~[?:1.8.0_252] >>>> >>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >>>> >>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> Caused by: org.apache.hadoop.hive.metastore.api.MetaException: >>>>> org.apache.hadoop.hive.serde2.SerDeException Encountered >>>>> AvroSerdeException >>>>> determining schema. Returning signal schema to indicate problem: Unable to >>>>> read schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:98) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> ... 28 more >>>> >>>> 2020-06-26T00:06:03,874 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] avro.AvroSerDe: AvroSerde::initialize(): Preset value of >>>>> avro.schema.literal == null >>>> >>>> 2020-06-26T00:06:03,878 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 >>>>> main] avro.AvroSerDe: Encountered AvroSerdeException determining schema. >>>>> Returning signal schema to indicate problem >>>> >>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read >>>>> schema from given path: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:900) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>>>> ~[hive-cli-3.1.2.jar:3.1.2] >>>> >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> ~[?:1.8.0_252] >>>> >>>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >>>> >>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> Caused by: java.io.FileNotFoundException: http:// >>>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >>>> >>>> at >>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498) >>>>> ~[?:1.8.0_252] >>>> >>>> at >>>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at >>>>> org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899) >>>>> ~[hadoop-common-3.1.2.jar:?] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> at >>>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139) >>>>> ~[hive-exec-3.1.2.jar:3.1.2] >>>> >>>> ... 33 more >>>> >>>> Thanks, >>> Ravi >>> >>