This is an automated email from the ASF dual-hosted git repository.
xxyu pushed a commit to branch kylin5
in repository https://gitbox.apache.org/repos/asf/kylin.git
The following commit(s) were added to refs/heads/kylin5 by this push:
new cfe3b8e06b KYLIN-5217, fix ut in kylin-engine-spark module
cfe3b8e06b is described below
commit cfe3b8e06b4d3b65ed1436cb24cb21b130023bae
Author: Mukvin <[email protected]>
AuthorDate: Wed Aug 3 16:31:02 2022 +0800
KYLIN-5217, fix ut in kylin-engine-spark module
---
.../spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java | 4 ++--
.../org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala | 2 +-
2 files changed, 3 insertions(+), 3 deletions(-)
diff --git
a/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java
b/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java
index 16da811964..0e3ce0e763 100644
---
a/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java
+++
b/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java
@@ -60,7 +60,7 @@ public class NSparkCubingSourceInputBySparkDataSourceTest
extends NLocalWithSpar
KylinBuildEnv.clean();
KylinBuildEnv kylinBuildEnv =
KylinBuildEnv.getOrCreate(getTestConfig());
getTestConfig().setProperty("kylin.source.provider.9",
- "NSparkDataSource");
+ "org.apache.kylin.engine.spark.source.NSparkDataSource");
getTestConfig().setProperty("kylin.build.resource.read-transactional-table-enabled",
"true");
NTableMetadataManager tableMgr =
NTableMetadataManager.getInstance(getTestConfig(), "ssb");
TableDesc fact = tableMgr.getTableDesc("SSB.P_LINEORDER");
@@ -80,7 +80,7 @@ public class NSparkCubingSourceInputBySparkDataSourceTest
extends NLocalWithSpar
KylinBuildEnv.clean();
KylinBuildEnv kylinBuildEnv =
KylinBuildEnv.getOrCreate(getTestConfig());
getTestConfig().setProperty("kylin.source.provider.9",
- "NSparkDataSource");
+ "org.apache.kylin.engine.spark.source.NSparkDataSource");
getTestConfig().setProperty("kylin.build.resource.read-transactional-table-enabled",
"true");
NTableMetadataManager tableMgr =
NTableMetadataManager.getInstance(getTestConfig(), "ssb");
TableDesc fact = tableMgr.getTableDesc("SSB.P_LINEORDER");
diff --git
a/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala
b/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala
index 41706bdefd..84ed8635a2 100644
---
a/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala
+++
b/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala
@@ -64,7 +64,7 @@ trait WithKylinExternalCatalog extends SparkFunSuite with
BeforeAndAfterAll {
metaStore.createTestMetadata(metadata: _*)
metaStore.overwriteSystemProp("kylin.use.external.catalog",
externalCatalog)
metaStore.overwriteSystemProp("kylin.NSparkDataSource.data.dir",
s"${kylinConf.getMetadataUrlPrefix}/../data")
- metaStore.overwriteSystemProp("kylin.source.provider.9",
"NSparkDataSource")
+ metaStore.overwriteSystemProp("kylin.source.provider.9",
"org.apache.kylin.engine.spark.source.NSparkDataSource")
metaStore.overwriteSystemProp("kylin.query.engine.sparder-additional-files",
fitPathForUT(additional))
metaStore.overwriteSystemProp("kylin.source.jdbc.adaptor", "Set By
WithKylinExternalCatalog")
metaStore.overwriteSystemProp("kylin.source.jdbc.driver", "Set By
WithKylinExternalCatalog")