[ 
https://issues.apache.org/jira/browse/HIVE-4561?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13780727#comment-13780727
 ] 

Hive QA commented on HIVE-4561:
-------------------------------



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12586100/HIVE-4561.3.patch

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/951/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/951/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests failed with: NonZeroExitCodeException: Command 'bash 
/data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and 
output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-951/source-prep.txt
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1527155.

At revision 1527155.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0 to p2
+ exit 1
'
{noformat}

This message is automatically generated.
                
> Column stats :  LOW_VALUE (or HIGH_VALUE) will always be 0.0000 ,if all the 
> column values larger than 0.0 (or if all column values smaller than 0.0)
> ----------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HIVE-4561
>                 URL: https://issues.apache.org/jira/browse/HIVE-4561
>             Project: Hive
>          Issue Type: Bug
>          Components: Statistics
>    Affects Versions: 0.12.0
>            Reporter: caofangkun
>            Assignee: caofangkun
>         Attachments: HIVE-4561.1.patch, HIVE-4561.2.patch, HIVE-4561.3.patch
>
>
> if all column values larger than 0.0  DOUBLE_LOW_VALUE always will be 0.0 
> or  if all column values less than 0.0,  DOUBLE_HIGH_VALUE will always be 
> hive (default)> create table src_test (price double);
> hive (default)> load data local inpath './test.txt' into table src_test;
> hive (default)> select * from src_test;
> OK
> 1.0
> 2.0
> 3.0
> Time taken: 0.313 seconds, Fetched: 3 row(s)
> hive (default)> analyze table src_test compute statistics for columns price;
> mysql> select * from TAB_COL_STATS \G;
>                  CS_ID: 16
>                DB_NAME: default
>             TABLE_NAME: src_test
>            COLUMN_NAME: price
>            COLUMN_TYPE: double
>                 TBL_ID: 2586
>         LONG_LOW_VALUE: 0
>        LONG_HIGH_VALUE: 0
>       DOUBLE_LOW_VALUE: 0.0000   # Wrong Result ! Expected is 1.0000
>      DOUBLE_HIGH_VALUE: 3.0000
>  BIG_DECIMAL_LOW_VALUE: NULL
> BIG_DECIMAL_HIGH_VALUE: NULL
>              NUM_NULLS: 0
>          NUM_DISTINCTS: 1
>            AVG_COL_LEN: 0.0000
>            MAX_COL_LEN: 0
>              NUM_TRUES: 0
>             NUM_FALSES: 0
>          LAST_ANALYZED: 1368596151
> 2 rows in set (0.00 sec)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to