804e opened a new issue, #13069:
URL: https://github.com/apache/hudi/issues/13069

   
   **Describe the problem you faced**
   
   I used flink sql to create table hudi_table_hive, inserted data, flink query 
is normal. However, an exception was throw during hive beeline query
   
   **To Reproduce**
   
   Steps to reproduce the behavior:
   
   1. flink sql client create table and insert data
   ```
   CREATE TABLE hudi_table_hive(
       ts BIGINT,
       uuid VARCHAR(40) PRIMARY KEY NOT ENFORCED,
       rider VARCHAR(20),
       driver VARCHAR(20),
       fare DOUBLE,
       city VARCHAR(20)
   )
   PARTITIONED BY (`city`)
   WITH (
     'connector' = 'hudi',
     'path' = 'hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive',
     'table.type' = 'COPY_ON_WRITE',
     'hive_sync.enabled' = 'true',
     'hive_sync.db' = 'hive_db',
     'hive_sync.mode' = 'hms',
     'hive_sync.metastore.uris' = 'thrift://hudi02.jafron.com:9083',
     'hive_sync.conf.dir' = '/opt/hadoop/conf'
   );
   
   -- insert data using values
   INSERT INTO hudi_table_hive VALUES
   
(1695159649087,'334e26e9-8355-45cc-97c6-c31daf0df330','rider-A','driver-K',19.10,'san_francisco'),
   
(1695091554788,'e96c4396-3fad-413a-a942-4cb36106d721','rider-C','driver-M',27.70
 ,'san_francisco'),
   
(1695046462179,'9909a8b1-2d15-4d3d-8ec9-efc48c536a00','rider-D','driver-L',33.90
 ,'san_francisco'),
   
(1695332066204,'1dced545-862b-4ceb-8b43-d2a568f6616b','rider-E','driver-O',93.50,'san_francisco'),
   
(1695516137016,'e3cf430c-889d-4015-bc98-59bdce1e530c','rider-F','driver-P',34.15,'sao_paulo'),
   
(1695376420876,'7a84095f-737f-40bc-b62f-6b69664712d2','rider-G','driver-Q',43.40
 ,'sao_paulo'),
   
(1695173887231,'3eeb61f7-c2b0-4636-99bd-5d7a5a1d2c04','rider-I','driver-S',41.06
 ,'chennai'),
   
(1695115999911,'c8abbe79-8d89-47ea-b4ce-4d224bae5bfa','rider-J','driver-T',17.85,'chennai');
   ```
   2. beeline can show table structure but cannot query
   ```
   0: jdbc:hive2://hudi02.jafron.com:2181,hudi01> show create table 
hudi_table_hive;
   INFO  : Compiling 
command(queryId=hive_20250402112526_4e945f4a-4a57-4086-8153-3530a66bbb40): show 
create table hudi_table_hive
   INFO  : Semantic Analysis Completed (retrial = false)
   INFO  : Returning Hive schema: 
Schema(fieldSchemas:[FieldSchema(name:createtab_stmt, type:string, comment:from 
deserializer)], properties:null)
   INFO  : Completed compiling 
command(queryId=hive_20250402112526_4e945f4a-4a57-4086-8153-3530a66bbb40); Time 
taken: 0.092 seconds
   INFO  : Executing 
command(queryId=hive_20250402112526_4e945f4a-4a57-4086-8153-3530a66bbb40): show 
create table hudi_table_hive
   INFO  : Starting task [Stage-0:DDL] in serial mode
   INFO  : Completed executing 
command(queryId=hive_20250402112526_4e945f4a-4a57-4086-8153-3530a66bbb40); Time 
taken: 0.054 seconds
   INFO  : OK
   +----------------------------------------------------+
   |                   createtab_stmt                   |
   +----------------------------------------------------+
   | CREATE EXTERNAL TABLE `hudi_table_hive`(           |
   |   `_hoodie_commit_time` string COMMENT '',         |
   |   `_hoodie_commit_seqno` string COMMENT '',        |
   |   `_hoodie_record_key` string COMMENT '',          |
   |   `_hoodie_partition_path` string COMMENT '',      |
   |   `_hoodie_file_name` string COMMENT '',           |
   |   `ts` bigint COMMENT '',                          |
   |   `uuid` string COMMENT '',                        |
   |   `rider` string COMMENT '',                       |
   |   `driver` string COMMENT '',                      |
   |   `fare` double COMMENT '')                        |
   | PARTITIONED BY (                                   |
   |   `city` string COMMENT '')                        |
   | ROW FORMAT SERDE                                   |
   |   'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'  |
   | WITH SERDEPROPERTIES (                             |
   |   'hoodie.query.as.ro.table'='false',              |
   |   'path'='hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive')  |
   | STORED AS INPUTFORMAT                              |
   |   'org.apache.hudi.hadoop.HoodieParquetInputFormat'  |
   | OUTPUTFORMAT                                       |
   |   'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' |
   | LOCATION                                           |
   |   'hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive' |
   | TBLPROPERTIES (                                    |
   |   'last_commit_completion_time_sync'='20250402112500650',  |
   |   'last_commit_time_sync'='20250402112458330',     |
   |   'spark.sql.sources.provider'='hudi',             |
   |   'spark.sql.sources.schema.numPartCols'='1',      |
   |   'spark.sql.sources.schema.numParts'='1',         |
   |   
'spark.sql.sources.schema.part.0'='{"type":"struct","fields":[{"name":"_hoodie_commit_time","type":"string","nullable":true,"metadata":{}},{"name":"_hoodie_commit_seqno","type":"string","nullable":true,"metadata":{}},{"name":"_hoodie_record_key","type":"string","nullable":true,"metadata":{}},{"name":"_hoodie_partition_path","type":"string","nullable":true,"metadata":{}},{"name":"_hoodie_file_name","type":"string","nullable":true,"metadata":{}},{"name":"ts","type":"long","nullable":true,"metadata":{}},{"name":"uuid","type":"string","nullable":false,"metadata":{}},{"name":"rider","type":"string","nullable":true,"metadata":{}},{"name":"driver","type":"string","nullable":true,"metadata":{}},{"name":"fare","type":"double","nullable":true,"metadata":{}},{"name":"city","type":"string","nullable":true,"metadata":{}}]}',
  |
   |   'spark.sql.sources.schema.partCol.0'='city',     |
   |   'transient_lastDdlTime'='1743558852')            |
   +----------------------------------------------------+
   33 rows selected (0.213 seconds)
   0: jdbc:hive2://hudi02.jafron.com:2181,hudi01> select * from hudi_table_hive;
   INFO  : Compiling 
command(queryId=hive_20250402112530_de8fbc6e-dbd1-4725-84b6-68ba18971c2a): 
select * from hudi_table_hive
   INFO  : Semantic Analysis Completed (retrial = false)
   INFO  : Returning Hive schema: 
Schema(fieldSchemas:[FieldSchema(name:hudi_table_hive._hoodie_commit_time, 
type:string, comment:null), 
FieldSchema(name:hudi_table_hive._hoodie_commit_seqno, type:string, 
comment:null), FieldSchema(name:hudi_table_hive._hoodie_record_key, 
type:string, comment:null), 
FieldSchema(name:hudi_table_hive._hoodie_partition_path, type:string, 
comment:null), FieldSchema(name:hudi_table_hive._hoodie_file_name, type:string, 
comment:null), FieldSchema(name:hudi_table_hive.ts, type:bigint, comment:null), 
FieldSchema(name:hudi_table_hive.uuid, type:string, comment:null), 
FieldSchema(name:hudi_table_hive.rider, type:string, comment:null), 
FieldSchema(name:hudi_table_hive.driver, type:string, comment:null), 
FieldSchema(name:hudi_table_hive.fare, type:double, comment:null), 
FieldSchema(name:hudi_table_hive.city, type:string, comment:null)], 
properties:null)
   INFO  : Completed compiling 
command(queryId=hive_20250402112530_de8fbc6e-dbd1-4725-84b6-68ba18971c2a); Time 
taken: 0.325 seconds
   INFO  : Executing 
command(queryId=hive_20250402112530_de8fbc6e-dbd1-4725-84b6-68ba18971c2a): 
select * from hudi_table_hive
   INFO  : Completed executing 
command(queryId=hive_20250402112530_de8fbc6e-dbd1-4725-84b6-68ba18971c2a); Time 
taken: 0.0 seconds
   INFO  : OK
   Error: java.io.IOException: java.lang.IllegalArgumentException 
(state=,code=0)
   ```
   hive server error log
   ```
   2025-04-02T11:14:57,719 INFO  [HiveServer2-Handler-Pool: Thread-60]: 
session.SessionState (:()) - Updating thread name to 
f199a7c9-401f-491a-a90e-fa01f9f6f43e HiveServer2-Handler-Pool: Thread-60
   2025-04-02T11:14:57,741 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: utils.HoodieInputFormatUtils (:()) - 
Reading hoodie metadata from path 
hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive
   2025-04-02T11:14:57,742 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: table.HoodieTableMetaClient (:()) - 
Loading HoodieTableMetaClient from 
hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive
   2025-04-02T11:14:57,746 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: table.HoodieTableConfig (:()) - Loading 
table properties from 
hdfs://hudi01.jafron.com:8020/datalake/hudi_table_hive/.hoodie/hoodie.properties
   2025-04-02T11:14:57,752 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: conf.HiveConf 
(HiveConf.java:getLogIdVar(5037)) - Using the default value passed in for log 
id: f199a7c9-401f-491a-a90e-fa01f9f6f43e
   2025-04-02T11:14:57,752 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: session.SessionState (:()) - Resetting 
thread name to  HiveServer2-Handler-Pool: Thread-60
   2025-04-02T11:14:57,755 INFO  [HiveServer2-Handler-Pool: Thread-60]: 
conf.HiveConf (HiveConf.java:getLogIdVar(5037)) - Using the default value 
passed in for log id: f199a7c9-401f-491a-a90e-fa01f9f6f43e
   2025-04-02T11:14:57,755 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: conf.HiveConf 
(HiveConf.java:getLogIdVar(5037)) - Using the default value passed in for log 
id: f199a7c9-401f-491a-a90e-fa01f9f6f43e
   2025-04-02T11:14:57,756 INFO  [HiveServer2-Handler-Pool: Thread-60]: 
conf.HiveConf (HiveConf.java:getLogIdVar(5037)) - Using the default value 
passed in for log id: f199a7c9-401f-491a-a90e-fa01f9f6f43e
   2025-04-02T11:14:57,752 WARN  [HiveServer2-Handler-Pool: Thread-60]: 
thrift.ThriftCLIService (:()) - Error fetching results: 
   org.apache.hive.service.cli.HiveSQLException: java.io.IOException: 
java.lang.IllegalArgumentException
        at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:465)
 ~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:309)
 ~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.session.HiveSessionImpl.fetchResults(HiveSessionImpl.java:895)
 ~[hive-service-3.1.3.jar:3.1.3]
        at sun.reflect.GeneratedMethodAccessor23.invoke(Unknown Source) ~[?:?]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_412]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_412]
        at 
org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
 ~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
 ~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
 ~[hive-service-3.1.3.jar:3.1.3]
        at java.security.AccessController.doPrivileged(Native Method) 
~[?:1.8.0_412]
        at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_412]
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878)
 ~[hadoop-common-3.3.4.jar:?]
        at 
org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
 ~[hive-service-3.1.3.jar:3.1.3]
        at com.sun.proxy.$Proxy44.fetchResults(Unknown Source) ~[?:?]
        at 
org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:561) 
~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:786)
 ~[hive-service-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1837)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1822)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) 
~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge.java:647)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
~[?:1.8.0_412]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
~[?:1.8.0_412]
        at java.lang.Thread.run(Thread.java:750) [?:1.8.0_412]
   Caused by: java.io.IOException: java.lang.IllegalArgumentException
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:602) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:509) 
~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:146) 
~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2691) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.reexec.ReExecDriver.getResults(ReExecDriver.java:229) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:460)
 ~[hive-service-3.1.3.jar:3.1.3]
        ... 24 more
   Caused by: java.lang.IllegalArgumentException
        at 
org.apache.hudi.common.util.ValidationUtils.checkArgument(ValidationUtils.java:33)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.timeline.versioning.TimelineLayoutVersion.<init>(TimelineLayoutVersion.java:40)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.HoodieTableConfig.getTimelineLayoutVersion(HoodieTableConfig.java:522)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.HoodieTableMetaClient.<init>(HoodieTableMetaClient.java:143)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.HoodieTableMetaClient.newMetaClient(HoodieTableMetaClient.java:692)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.HoodieTableMetaClient.access$000(HoodieTableMetaClient.java:85)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.common.table.HoodieTableMetaClient$Builder.build(HoodieTableMetaClient.java:774)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.hadoop.utils.HoodieInputFormatUtils.getTableMetaClientForBasePathUnchecked(HoodieInputFormatUtils.java:374)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.hadoop.InputPathHandler.parseInputPaths(InputPathHandler.java:110)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.hadoop.InputPathHandler.<init>(InputPathHandler.java:72) 
~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hudi.hadoop.HoodieCopyOnWriteTableInputFormat.listStatus(HoodieCopyOnWriteTableInputFormat.java:110)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:332) 
~[hadoop-mapreduce-client-core-3.3.4.jar:?]
        at 
org.apache.hudi.hadoop.HoodieParquetInputFormatBase.getSplits(HoodieParquetInputFormatBase.java:68)
 ~[hudi-hadoop-mr-bundle-0.14.1.jar:0.14.1]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.generateWrappedSplits(FetchOperator.java:425)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextSplits(FetchOperator.java:395)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:314)
 ~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:540) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:509) 
~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:146) 
~[hive-exec-3.1.3.jar:3.1.3]
        at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2691) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hadoop.hive.ql.reexec.ReExecDriver.getResults(ReExecDriver.java:229) 
~[hive-exec-3.1.3.jar:3.1.3]
        at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:460)
 ~[hive-service-3.1.3.jar:3.1.3]
        ... 24 more
   2025-04-02T11:14:57,755 INFO  [HiveServer2-Handler-Pool: Thread-60]: 
session.SessionState (:()) - Updating thread name to 
f199a7c9-401f-491a-a90e-fa01f9f6f43e HiveServer2-Handler-Pool: Thread-60
   2025-04-02T11:14:57,755 INFO  [f199a7c9-401f-491a-a90e-fa01f9f6f43e 
HiveServer2-Handler-Pool: Thread-60]: session.SessionState (:()) - Resetting 
thread name to  HiveServer2-Handler-Pool: Thread-60
   
   ```
   check soure code in `TimelineLayoutVersion.java:40`, found property 
`hoodie.timeline.layout.version` must be in  [0,1], but it was 2 in 
`/datalake/hudi_table_hive/.hoodie/hoodie.properties`
   ```
   #Updated at 2025-04-02T03:24:57.110923Z
   #Wed Apr 02 11:24:57 HKT 2025
   hoodie.table.keygenerator.type=SIMPLE_AVRO
   hoodie.table.type=COPY_ON_WRITE
   hoodie.table.precombine.field=ts
   
hoodie.table.create.schema={"type"\:"record","name"\:"hudi_table_hive_record","namespace"\:"hoodie.hudi_table_hive","fields"\:[{"name"\:"ts","type"\:["null","long"],"default"\:null},{"name"\:"uuid","type"\:"string"},{"name"\:"rider","type"\:["null","string"],"default"\:null},{"name"\:"driver","type"\:["null","string"],"default"\:null},{"name"\:"fare","type"\:["null","double"],"default"\:null},{"name"\:"city","type"\:["null","string"],"default"\:null}]}
   hoodie.timeline.layout.version=2
   hoodie.timeline.history.path=history
   hoodie.table.checksum=860864952
   hoodie.datasource.write.drop.partition.columns=false
   hoodie.record.merge.strategy.id=00000000-0000-0000-0000-000000000000
   hoodie.datasource.write.hive_style_partitioning=false
   hoodie.table.metadata.partitions.inflight=
   hoodie.database.name=default_database
   hoodie.datasource.write.partitionpath.urlencode=false
   hoodie.record.merge.mode=CUSTOM
   hoodie.table.version=8
   
hoodie.compaction.payload.class=org.apache.hudi.common.model.EventTimeAvroPayload
   hoodie.table.initial.version=8
   hoodie.table.metadata.partitions=column_stats,files,partition_stats
   hoodie.table.partition.fields=city
   hoodie.table.cdc.enabled=false
   hoodie.archivelog.folder=history
   hoodie.table.name=hudi_table_hive
   hoodie.table.recordkey.fields=uuid
   hoodie.timeline.path=timeline
   hoodie.populate.meta.fields=true
   ```
   
   **Environment Description**
   
   * Hudi version : 0.14.1
   
   * Hive version : 3.13
   
   * Hadoop version : 3.3.4
   
   * Storage (HDFS/S3/GCS..) : HDFS
   
   
   
   
   
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to