[ 
https://issues.apache.org/jira/browse/HIVE-11438?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14651515#comment-14651515
 ] 

Hive QA commented on HIVE-11438:
--------------------------------



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12748398/HIVE-11438.1.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4797/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4797/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4797/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-4797/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 8b2cd2a HIVE-11380: NPE when FileSinkOperator is not initialized 
(Yongzhi Chen, reviewed by Sergio Pena)
+ git clean -f -d
+ git checkout master
Already on 'master'
+ git reset --hard origin/master
HEAD is now at 8b2cd2a HIVE-11380: NPE when FileSinkOperator is not initialized 
(Yongzhi Chen, reviewed by Sergio Pena)
+ git merge --ff-only origin/master
Already up-to-date.
+ git gc
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12748398 - PreCommit-HIVE-TRUNK-Build

> Join a ACID table with non-ACID table fail with MR on 1.0.0
> -----------------------------------------------------------
>
>                 Key: HIVE-11438
>                 URL: https://issues.apache.org/jira/browse/HIVE-11438
>             Project: Hive
>          Issue Type: Bug
>          Components: Query Processor, Transactions
>    Affects Versions: 1.0.0
>            Reporter: Daniel Dai
>            Assignee: Daniel Dai
>             Fix For: 1.0.1
>
>         Attachments: HIVE-11438.1.patch
>
>
> The following script fail on MR mode:
> Preparation:
> {code}
> CREATE TABLE orc_update_table (k1 INT, f1 STRING, op_code STRING) 
> CLUSTERED BY (k1) INTO 2 BUCKETS 
> STORED AS ORC TBLPROPERTIES("transactional"="true"); 
> INSERT INTO TABLE orc_update_table VALUES (1, 'a', 'I');
> CREATE TABLE orc_table (k1 INT, f1 STRING) 
> CLUSTERED BY (k1) SORTED BY (k1) INTO 2 BUCKETS 
> STORED AS ORC; 
> INSERT OVERWRITE TABLE orc_table VALUES (1, 'x');
> {code}
> Then run the following script:
> {code}
> SET hive.execution.engine=mr; 
> SET hive.auto.convert.join=false; 
> SET hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
> SELECT t1.*, t2.* FROM orc_table t1 
> JOIN orc_update_table t2 ON t1.k1=t2.k1 ORDER BY t1.k1;
> {code}
> Stack:
> {code}
> java.lang.NullPointerException
>       at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:265)
>       at 
> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getCombineSplits(CombineHiveInputFormat.java:272)
>       at 
> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:509)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:624)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:616)
>       at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:492)
>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1296)
>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1293)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1293)
>       at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:585)
>       at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:580)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>       at 
> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:580)
>       at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:571)
>       at 
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:429)
>       at 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:137)
>       at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>       at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>       at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1606)
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1367)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1179)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1006)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:996)
>       at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:247)
>       at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:199)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:410)
>       at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:783)
>       at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:677)
>       at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:616)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>       at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> Job Submission failed with exception 'java.lang.NullPointerException(null)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> {code}
> Note the query is the same as HIVE-11422. But in 1.0.0 for this Jira, it 
> throw a different exeception.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to