[ https://issues.apache.org/jira/browse/HIVE-12222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15523301#comment-15523301 ]
Hive QA commented on HIVE-12222: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12830337/HIVE-12222.2.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/1305/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/1305/console Test logs: http://ec2-204-236-174-241.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-Build-1305/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2016-09-26 14:56:47.851 + [[ -n /usr/java/jdk1.8.0_25 ]] + export JAVA_HOME=/usr/java/jdk1.8.0_25 + JAVA_HOME=/usr/java/jdk1.8.0_25 + export PATH=/usr/java/jdk1.8.0_25/bin/:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin + PATH=/usr/java/jdk1.8.0_25/bin/:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + cd /data/hive-ptest/working/ + tee /data/hive-ptest/logs/PreCommit-HIVE-Build-1305/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2016-09-26 14:56:47.853 + cd apache-github-source-source + git fetch origin + git reset --hard HEAD warning: unable to access '/home/sseth/.config/git/attributes': Permission denied warning: unable to access '/home/sseth/.config/git/attributes': Permission denied HEAD is now at 4ce5fe1 HIVE-14831: Missing Druid dependencies at runtime (Jesus Camacho Rodriguez, reviewed by Ashutosh Chauhan) + git clean -f -d warning: unable to access '/home/sseth/.config/git/ignore': Permission denied Removing java/ Removing src/ + git checkout master warning: unable to access '/home/sseth/.config/git/ignore': Permission denied Already on 'master' Your branch is up-to-date with 'origin/master'. + git reset --hard origin/master HEAD is now at 4ce5fe1 HIVE-14831: Missing Druid dependencies at runtime (Jesus Camacho Rodriguez, reviewed by Ashutosh Chauhan) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2016-09-26 14:56:48.786 + patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hive-ptest/working/scratch/build.patch + [[ -f /data/hive-ptest/working/scratch/build.patch ]] + chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh + /data/hive-ptest/working/scratch/smart-apply-patch.sh /data/hive-ptest/working/scratch/build.patch warning: unable to access '/home/sseth/.config/git/attributes': Permission denied error: a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java: No such file or directory error: a/spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java: No such file or directory error: a/spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcServer.java: No such file or directory error: a/spark-client/src/test/java/org/apache/hive/spark/client/rpc/TestRpc.java: No such file or directory The patch does not appear to apply with p0, p1, or p2 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12830337 - PreCommit-HIVE-Build > Define port range in property for RPCServer > ------------------------------------------- > > Key: HIVE-12222 > URL: https://issues.apache.org/jira/browse/HIVE-12222 > Project: Hive > Issue Type: Improvement > Components: CLI, Spark > Affects Versions: 1.2.1 > Environment: Apache Hadoop 2.7.0 > Apache Hive 1.2.1 > Apache Spark 1.5.1 > Reporter: Andrew Lee > Assignee: Aihua Xu > Attachments: HIVE-12222.1.patch, HIVE-12222.2.patch > > > Creating this JIRA after discussin with Xuefu on the dev mailing list. Would > need some help to review and update the fields in this JIRA ticket, thanks. > I notice that in > ./spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcServer.java > The port number is assigned with 0 which means it will be a random port every > time when the RPC Server is created to talk to Spark in the same session. > Because of this, this is causing problems to configure firewall between the > HiveCLI RPC Server and Spark due to unpredictable port numbers here. In other > word, users need to open all hive ports range > from Data Node => HiveCLI (edge node). > {code} > this.channel = new ServerBootstrap() > .group(group) > .channel(NioServerSocketChannel.class) > .childHandler(new ChannelInitializer<SocketChannel>() { > @Override > public void initChannel(SocketChannel ch) throws Exception { > SaslServerHandler saslHandler = new SaslServerHandler(config); > final Rpc newRpc = Rpc.createServer(saslHandler, config, ch, > group); > saslHandler.rpc = newRpc; > Runnable cancelTask = new Runnable() { > @Override > public void run() { > LOG.warn("Timed out waiting for hello from client."); > newRpc.close(); > } > }; > saslHandler.cancelTask = group.schedule(cancelTask, > RpcServer.this.config.getServerConnectTimeoutMs(), > TimeUnit.MILLISECONDS); > } > }) > {code} > 2 Main reasons. > - Most users (what I see and encounter) use HiveCLI as a command line tool, > and in order to use that, they need to login to the edge node (via SSH). Now, > here comes the interesting part. > Could be true or not, but this is what I observe and encounter from time to > time. Most users will abuse the resource on that edge node (increasing > HADOOP_HEAPSIZE, dumping output to local disk, running huge python workflow, > etc), this may cause the HS2 process to run into OOME, choke and die, etc. > various resource issues including others like login, etc. > - Analyst connects to Hive via HS2 + ODBC. So HS2 needs to be highly > available. This makes sense to run it on the gateway node or a service node > and separated from the HiveCLI. > The logs are located in different location, monitoring and auditing is easier > to run HS2 with a daemon user account, etc. so we don't want users to run > HiveCLI where HS2 is running. > It's better to isolate the resource this way to avoid any memory, file > handlers, disk space, issues. > From a security standpoint, > - Since users can login to edge node (via SSH), the security on the edge node > needs to be fortified and enhanced. Therefore, all the FW comes in and > auditing. > - Regulation/compliance for auditing is another requirement to monitor all > traffic, specifying ports and locking down the ports makes it easier since we > can focus > on a range to monitor and audit. -- This message was sent by Atlassian JIRA (v6.3.4#6332)