sivabalan narayanan created HUDI-6696:
-----------------------------------------
Summary: Clustering fails w/ spark3.3 bundle 0.12.3
Key: HUDI-6696
URL: https://issues.apache.org/jira/browse/HUDI-6696
Project: Apache Hudi
Issue Type: Bug
Components: clustering
Reporter: sivabalan narayanan
I was trying our spark streaming ingestion to hudi and ran into clustering
issue w/ 0.12.3 bundle.
{code:java}
import org.apache.hudi.QuickstartUtils._
import scala.collection.JavaConversions._
import org.apache.spark.sql.SaveMode._
import org.apache.hudi.DataSourceReadOptions._
import org.apache.hudi.DataSourceWriteOptions._
import org.apache.hudi.config.HoodieWriteConfig._
import org.apache.hudi.common.model.HoodieRecord
import java.time.LocalDateTime
import org.apache.spark.sql.streaming.Trigger{code}
{code:java}
val sourceBasePath = SOURCE_PATH
val targetPath = TARGET_PATH{code}
{code:java}
val df = spark.readStream.format("hudi").load(sourceBasePath)val query =
df.writeStream.foreachBatch { (batchDF: org.apache.spark.sql.DataFrame, _:
Long) => {
println(LocalDateTime.now() + " Start writing cow table")
batchDF.drop("_hoodie_commit_time").drop("_hoodie_commit_seqno").drop("_hoodie_record_key").drop("_hoodie_partition_path").drop("_hoodie_file_name").
filter("id%2 ==1").
withColumn("date_col",substring(col("created_at"),0,10)).
write.format("hudi").
option(TABLE_TYPE.key, "COPY_ON_WRITE").
option(PRECOMBINE_FIELD.key, "created_at").
option(RECORDKEY_FIELD.key, "id").
option(PARTITIONPATH_FIELD.key, "date_col").
option("hoodie.datasource.write.operation","insert").
option("hoodie.insert.shuffle.parallelism","10").
option("hoodie.parquet.small.file.limit","0").
option("hoodie.clustering.inline", "true").
option("hoodie.clustering.inline.max.commits", "10").
option("hoodie.clustering.plan.strategy.target.file.max.bytes",
"104857600").
option("hoodie.clustering.plan.strategy.small.file.limit",
"52428800").
option("hoodie.table.name","hudi_tbl").
option("hoodie.write.concurrency.mode","optimistic_concurrency_control").
option("hoodie.cleaner.policy.failed.writes","LAZY").
option("hoodie.write.lock.provider","org.apache.hudi.client.transaction.lock.ZookeeperBasedLockProvider").
option("hoodie.write.lock.zookeeper.url","ABC").
option("hoodie.write.lock.zookeeper.port","2181").
option("hoodie.write.lock.zookeeper.lock_key","tbl1").
option("hoodie.write.lock.zookeeper.base_path","/tmp/locks").
mode(Append).
save(targetPath)
println(LocalDateTime.now() + " finish")
}
}.option("checkpointLocation", TARGET_CHECKPOINT).
trigger(Trigger.ProcessingTime("10
minutes")).start()query.awaitTermination()
{code}
stacktrace:
{code:java}
scala> query.awaitTermination()
2023-08-15T03:00:05.583925 Start writing cow table
23/08/15 03:00:05 WARN DFSPropertiesConfiguration: Cannot find HUDI_CONF_DIR,
please set it as the dir of hudi-defaults.conf
23/08/15 03:00:05 WARN DFSPropertiesConfiguration: Properties file
file:/etc/hudi/conf/hudi-defaults.conf not found. Ignoring to load props file
# WARNING: Unable to get Instrumentation. Dynamic Attach failed. You may add
this JAR as -javaagent manually, or supply -Djdk.attach.allowAttachSelf
# WARNING: Unable to attach Serviceability Agent. Unable to attach even with
module exceptions: [org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException:
Sense failed., org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException: Sense
failed., org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException: Sense
failed.]
23/08/15 03:01:14 WARN BaseRollbackActionExecutor: Rollback finished without
deleting inflight instant file.
Instant=[==>20230814185344428__replacecommit__INFLIGHT]
23/08/15 03:01:16 ERROR MicroBatchExecution: Query [id =
fd43b951-f1fd-479e-bd87-1aae2e9e396d, runId =
70797db4-fbcd-40f0-b09d-540013da6717] terminated with error
java.util.concurrent.CompletionException: java.lang.ClassNotFoundException:
Failed to find data source: hoodie-parquet. Please find packages at
https://spark.apache.org/third-party-projects.html
at
java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
~[?:?]
at
java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
~[?:?]
at
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
~[?:?]
at
java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1692)
~[?:?]
at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
~[?:?]
at
java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
~[?:?]
at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) ~[?:?]
at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
~[?:?]
at
java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183)
~[?:?]
Caused by: java.lang.ClassNotFoundException:
Failed to find data source: hoodie-parquet. Please find packages at
https://spark.apache.org/third-party-projects.html
at
org.apache.spark.sql.errors.QueryExecutionErrors$.failedToFindDataSourceError(QueryExecutionErrors.scala:573)
~[spark-catalyst_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:675)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
~[?:?]
... 6 more
Caused by: java.lang.ClassNotFoundException: hoodie-parquet.DefaultSource
at
jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
~[?:?]
at
jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
~[?:?]
at java.lang.ClassLoader.loadClass(ClassLoader.java:527) ~[?:?]
at
org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$5(DataSource.scala:661)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at scala.util.Try$.apply(Try.scala:213) ~[scala-library-2.12.14.jar:?]
at
org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$4(DataSource.scala:661)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at scala.util.Failure.orElse(Try.scala:224)
~[scala-library-2.12.14.jar:?]
at
org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:661)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
~[spark-sql_2.12-3.3.0.jar:3.3.0]
at
org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
at
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
~[?:?]
... 6 more
org.apache.spark.sql.streaming.StreamingQueryException: Query [id =
fd43b951-f1fd-479e-bd87-1aae2e9e396d, runId =
70797db4-fbcd-40f0-b09d-540013da6717] terminated with exception:
java.lang.ClassNotFoundException:
Failed to find data source: hoodie-parquet. Please find packages at
https://spark.apache.org/third-party-projects.html at
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:330)
at
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:208)
Caused by: java.util.concurrent.CompletionException:
java.lang.ClassNotFoundException:
Failed to find data source: hoodie-parquet. Please find packages at
https://spark.apache.org/third-party-projects.html at
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
at
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
at
java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
at
java.base/java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1692)
at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
at
java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
at
java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
at
java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183)
Caused by: java.lang.ClassNotFoundException:
Failed to find data source: hoodie-parquet. Please find packages at
https://spark.apache.org/third-party-projects.html at
org.apache.spark.sql.errors.QueryExecutionErrors$.failedToFindDataSourceError(QueryExecutionErrors.scala:573)
at
org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:675)
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
at
org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
at
org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
at
org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
at
org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
at
org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
at org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
at
org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
at
org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
at
java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
... 6 more
Caused by: java.lang.ClassNotFoundException: hoodie-parquet.DefaultSource
at
java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
at
java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:527)
at
org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$5(DataSource.scala:661)
at scala.util.Try$.apply(Try.scala:213)
at
org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$4(DataSource.scala:661)
at scala.util.Failure.orElse(Try.scala:224)
at
org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:661)
... 17 morescala>
{code}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)