mich, Do you want this ========================================================================================== [running]mqq@10.205.3.29:/data/home/hive/conf$ ps aux | grep SparkPi mqq 20070 3.6 0.8 10445048 267028 pts/16 Sl+ 13:09 0:11 /data/home/jdk/bin/java -Dlog4j.configuration=file:///data/home/spark/conf/log4j.properties -cp /data/home/spark/lib/*:/data/home/hadoop/share/hadoop/common/*:/data/home/hadoop/share/hadoop/common/lib/*:/data/home/hadoop/share/hadoop/yarn/*:/data/home/hadoop/share/hadoop/yarn/lib/*:/data/home/hadoop/share/hadoop/hdfs/*:/data/home/hadoop/share/hadoop/hdfs/lib/*:/data/home/hadoop/share/hadoop/tools/*:/data/home/hadoop/share/hadoop/mapreduce/*:/data/home/spark/conf/:/data/home/spark/lib/spark-assembly-1.4.1-hadoop2.5.1_150903.jar:/data/home/spark/lib/datanucleus-api-jdo-3.2.6.jar:/data/home/spark/lib/datanucleus-core-3.2.10.jar:/data/home/spark/lib/datanucleus-rdbms-3.2.9.jar:/data/home/hadoop/conf/:/data/home/hadoop/conf/:/data/home/spark/lib/*:/data/home/hadoop/share/hadoop/common/*:/data/home/hadoop/share/hadoop/common/lib/*:/data/home/hadoop/share/hadoop/yarn/*:/data/home/hadoop/share/hadoop/yarn/lib/*:/data/home/hadoop/share/hadoop/hdfs/*:/data/home/hadoop/share/hadoop/hdfs/lib/*:/data/home/hadoop/share/hadoop/tools/*:/data/home/hadoop/share/hadoop/mapreduce/* -XX:MaxPermSize=256m org.apache.spark.deploy.SparkSubmit --master yarn-cluster --class org.apache.spark.examples.SparkPi --queue spark --num-executors 4 /data/home/spark/lib/spark-examples-1.4.1-hadoop2.5.1.jar 100000000 mqq 22410 0.0 0.0 110600 1004 pts/8 S+ 13:14 0:00 grep SparkPi [running]mqq@10.205.3.29:/data/home/hive/conf$ top -p 20070
top - 13:14:48 up 504 days, 19:17, 19 users, load average: 1.41, 1.10, 0.99 Tasks: 1 total, 0 running, 1 sleeping, 0 stopped, 0 zombie Cpu(s): 18.1%us, 2.7%sy, 0.0%ni, 74.4%id, 4.5%wa, 0.0%hi, 0.2%si, 0.0%st Mem: 32740732k total, 31606288k used, 1134444k free, 475908k buffers Swap: 2088952k total, 61076k used, 2027876k free, 27594452k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 20070 mqq 20 0 10.0g 260m 32m S 0.0 0.8 0:11.38 java ========================================================================================== Harsh, physical cpu cores is 1, virtual cpu cores is 4 Thanks. 2016-05-13 13:08 GMT+08:00, Harsh J <ha...@cloudera.com>: > How many CPU cores are on that machine? Read http://qr.ae/8Uv3Xq > > You can also confirm the above by running the pmap utility on your process > and most of the virtual memory would be under 'anon'. > > On Fri, 13 May 2016 09:11 jone, <joyoungzh...@gmail.com> wrote: > >> The virtual memory is 9G When i run org.apache.spark.examples.SparkPi >> under yarn-cluster model,which using default configurations. >> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ >> COMMAND >> >> 4519 mqq 20 0 9041 <2009041>m 248m 26m S 0.3 0.8 0:19.85 >> java >> I am curious why is so high? >> >> Thanks. >> > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org