[ https://issues.apache.org/jira/browse/HIVE-19155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16434912#comment-16434912 ]
Hive QA commented on HIVE-19155: -------------------------------- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 1s{color} | {color:blue} Findbugs executables are not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 47s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 19s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 32s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 55s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 18s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 9s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 13s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 37s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 12s{color} | {color:red} druid-handler: The patch generated 2 new + 6 unchanged - 0 fixed = 8 total (was 6) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 18m 45s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-10148/dev-support/hive-personality.sh | | git revision | master / 244ca8e | | Default Java | 1.8.0_111 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-10148/yetus/diff-checkstyle-druid-handler.txt | | modules | C: druid-handler ql U: . | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-10148/yetus.txt | | Powered by | Apache Yetus http://yetus.apache.org | This message was automatically generated. > Day time saving cause Druid inserts to fail with > org.apache.hive.druid.io.druid.java.util.common.UOE: Cannot add overlapping > segments > ------------------------------------------------------------------------------------------------------------------------------------- > > Key: HIVE-19155 > URL: https://issues.apache.org/jira/browse/HIVE-19155 > Project: Hive > Issue Type: Bug > Components: Druid integration > Reporter: slim bouguerra > Assignee: slim bouguerra > Priority: Major > Attachments: HIVE-19155.patch > > > If you try to insert data around the daylight saving time hour the query > fails with following exception > {code} > 2018-04-10T11:24:58,836 ERROR [065fdaa2-85f9-4e49-adaf-3dc14d51be90 main] > exec.DDLTask: Failed > org.apache.hadoop.hive.ql.metadata.HiveException: > org.apache.hive.druid.io.druid.java.util.common.UOE: Cannot add overlapping > segments [2015-03-08T05:00:00.000Z/2015-03-09T05:00:00.000Z and > 2015-03-09T04:00:00.000Z/2015-03-10T04:00:00.000Z] with the same version > [2018-04-10T11:24:48.388-07:00] > at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:914) > ~[hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:919) > ~[hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4831) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:394) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2443) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2114) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1797) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1538) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1532) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:204) > [hive-exec-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) > [hive-cli-3.1.0-SNAPSHOT.jar:?] > at > org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) > [hive-cli-3.1.0-SNAPSHOT.jar:?] > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) > [hive-cli-3.1.0-SNAPSHOT.jar:?] > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:335) > [hive-cli-3.1.0-SNAPSHOT.jar:?] > at > org.apache.hadoop.hive.ql.QTestUtil.executeClientInternal(QTestUtil.java:1455) > [hive-it-util-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:1429) > [hive-it-util-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.cli.control.CoreCliDriver.runTest(CoreCliDriver.java:177) > [hive-it-util-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:104) > [hive-it-util-3.1.0-SNAPSHOT.jar:3.1.0-SNAPSHOT] > at > org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver(TestMiniDruidCliDriver.java:59) > [test-classes/:?] > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[?:1.8.0_92] > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:1.8.0_92] > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:1.8.0_92] > {code} > You can reproduce this using the following DDL > {code} > create database druid_test; > use druid_test; > create table test_table(`timecolumn` timestamp, `userid` string, `num_l` > float); > insert into test_table values ('2015-03-08 00:00:00', 'i1-start', 4); > insert into test_table values ('2015-03-08 23:59:59', 'i1-end', 1); > insert into test_table values ('2015-03-09 00:00:00', 'i2-start', 4); > insert into test_table values ('2015-03-09 23:59:59', 'i2-end', 1); > insert into test_table values ('2015-03-10 00:00:00', 'i3-start', 2); > insert into test_table values ('2015-03-10 23:59:59', 'i3-end', 2); > CREATE TABLE druid_table > STORED BY 'org.apache.hadoop.hive.druid.DruidStorageHandler' > TBLPROPERTIES ("druid.segment.granularity" = "DAY") > AS > select cast(`timecolumn` as timestamp with local time zone) as `__time`, > `userid`, `num_l` FROM test_table; > {code} > The fix is to always adjust the Druid segments identifiers to UTC. -- This message was sent by Atlassian JIRA (v7.6.3#76005)