LeonYoah commented on issue #9638:
URL: https://github.com/apache/seatunnel/issues/9638#issuecomment-3218098216

   > env { parallelism = 1 job.mode = "BATCH" job.name = 
"dws_user_vehicle_score_merge_result" read_limit.rows_per_second=100 
execution.checkpoint.interval= 60000 }
   > 
   > source { Hive { table_name = "dws.dws_test" metastore_uri = 
"thrift://ip-172-30-xxxxx.cn-north-1.compute.internal:9083" 
hive.hadoop.conf-path = "/home/ec2-user/hadoop-conf" hive.hadoop.conf = { 
bucket="s3a://xxxxxxx" fs.s3a.endpoint="s3.cn-north-1.amazonaws.com.cn" 
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
 } read_partitions = ["dt=2025-05-18"] } }
   > 
   > sink { jdbc { url = 
"jdbc:mysql://mysql-xxxx:3306/test_dev?useUnicode=true&characterEncoding=UTF-8&rewriteBatchedStatements=true"
 driver = "com.mysql.cj.jdbc.Driver" user = "xxxxx" password = "xxxxx" 
schema_save_mode = "CREATE_SCHEMA_WHEN_NOT_EXIST" data_save_mode="DROP_DATA" 
enable_upsert = false query = "insert into test_dev(union_id,vin_score,dt) 
values(?,?,?)" batch_size = 1000 batch_interval = 500 } }
   > 
   > [@LeonYoah](https://github.com/LeonYoah) I checked database, i can `SELECT 
SCHEMA_NAME FROM information_schema.schemata WHERE SCHEMA_NAME = 'test_dev'`, 和 
I have been grant all permission for user. But data_save_mode="DROP_DATA" is 
not effect. The job can seccess run after I try to manual truncate table.
   
   The problem may be in the [query], you put the indication in the insert 
statement, the sink side cannot parse the table name, please use, 
generate_sink_sql=true and table="test_dev"


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to