1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
| [root@com SparkR-pkg] Running /root/spark-1.1.0//bin/spark-submit --class edu.berkeley.cs.amplab.sparkr.SparkRRunner --files ./examples/pi.R /root/xiaolong.yuanxl/SparkR-pkg/lib/SparkR/sparkr-assembly-0.1.jar ./examples/pi.R local[2] -------------------------------------------------> old java home=> /usr/java/jdk1.7.0_55-cloudera/ /etc/hadoop/conf new java home /usr/java/jdk1.6.0_31/
Spark assembly has been built with Hive, including Datanucleus jars on classpath WARNING: ignoring environment value of R_HOME Loading required package: methods [SparkR] Initializing with classpath /root/xiaolong.yuanxl/SparkR-pkg/lib/SparkR/sparkr-assembly-0.1.jar
15/04/23 17:51:52 INFO spark.SecurityManager: Changing view acls to: root, 15/04/23 17:51:52 INFO spark.SecurityManager: Changing modify acls to: root, 15/04/23 17:51:52 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root, ); users with modify permissions: Set(root, ) 15/04/23 17:51:53 INFO slf4j.Slf4jLogger: Slf4jLogger started 15/04/23 17:51:53 INFO Remoting: Starting remoting 15/04/23 17:51:53 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@com.hunantv.logservernode2:47089] 15/04/23 17:51:53 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@com.hunantv.logservernode2:47089] 15/04/23 17:51:53 INFO util.Utils: Successfully started service 'sparkDriver' on port 47089. 15/04/23 17:51:53 INFO spark.SparkEnv: Registering MapOutputTracker 15/04/23 17:51:53 INFO spark.SparkEnv: Registering BlockManagerMaster 15/04/23 17:51:53 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-local-20150423175153-0caa 15/04/23 17:51:54 INFO util.Utils: Successfully started service 'Connection manager for block manager' on port 60352. 15/04/23 17:51:54 INFO network.ConnectionManager: Bound socket to port 60352 with id = ConnectionManagerId(com.hunantv.logservernode2,60352) 15/04/23 17:51:54 INFO storage.MemoryStore: MemoryStore started with capacity 265.0 MB 15/04/23 17:51:54 INFO storage.BlockManagerMaster: Trying to register BlockManager 15/04/23 17:51:54 INFO storage.BlockManagerMasterActor: Registering block manager com.hunantv.logservernode2:60352 with 265.0 MB RAM 15/04/23 17:51:54 INFO storage.BlockManagerMaster: Registered BlockManager 15/04/23 17:51:54 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-92c60115-974e-4178-86c1-d5ce689383f4 15/04/23 17:51:54 INFO spark.HttpServer: Starting HTTP Server 15/04/23 17:51:54 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/04/23 17:51:54 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:33381 15/04/23 17:51:54 INFO util.Utils: Successfully started service 'HTTP file server' on port 33381. 15/04/23 17:51:54 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/04/23 17:51:54 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040 15/04/23 17:51:54 INFO util.Utils: Successfully started service 'SparkUI' on port 4040. 15/04/23 17:51:54 INFO ui.SparkUI: Started SparkUI at http://com.hunantv.logservernode2:4040 15/04/23 17:51:54 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 15/04/23 17:51:55 INFO spark.SparkContext: Added JAR file:///root/xiaolong.yuanxl/SparkR-pkg/lib/SparkR/sparkr-assembly-0.1.jar at http://com.hunantv.logservernode2:33381/jars/sparkr-assembly-0.1.jar with timestamp 1429782715609 15/04/23 17:51:55 INFO util.Utils: Copying /root/xiaolong.yuanxl/SparkR-pkg/./examples/pi.R to /tmp/spark-64f7170f-21ed-4551-8c1d-1843895a8e47/pi.R 15/04/23 17:51:55 INFO spark.SparkContext: Added file file:/root/xiaolong.yuanxl/SparkR-pkg/./examples/pi.R at http://com.hunantv.logservernode2:33381/files/pi.R with timestamp 1429782715611 15/04/23 17:51:55 INFO util.AkkaUtils: Connecting to HeartbeatReceiver: akka.tcp://sparkDriver@com.hunantv.logservernode2:47089/user/HeartbeatReceiver 15/04/23 17:51:56 INFO spark.SparkContext: Starting job: collect at NativeMethodAccessorImpl.java:-2 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Got job 0 (collect at NativeMethodAccessorImpl.java:-2) with 2 output partitions (allowLocal=false) 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Final stage: Stage 0(collect at NativeMethodAccessorImpl.java:-2) 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Parents of final stage: List() 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Missing parents: List() 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Submitting Stage 0 (RRDD[1] at RDD at RRDD.scala:19), which has no missing parents 15/04/23 17:51:56 INFO storage.MemoryStore: ensureFreeSpace(73784) called with curMem=0, maxMem=277842493 15/04/23 17:51:56 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 72.1 KB, free 264.9 MB) 15/04/23 17:51:56 INFO scheduler.DAGScheduler: Submitting 2 missing tasks from Stage 0 (RRDD[1] at RDD at RRDD.scala:19) 15/04/23 17:51:56 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with 2 tasks 15/04/23 17:51:56 WARN scheduler.TaskSetManager: Stage 0 contains a task of very large size (391 KB). The maximum recommended task size is 100 KB. 15/04/23 17:51:56 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, localhost, PROCESS_LOCAL, 401308 bytes) 15/04/23 17:51:56 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 0.0 (TID 1, localhost, PROCESS_LOCAL, 401308 bytes) 15/04/23 17:51:56 INFO executor.Executor: Running task 0.0 in stage 0.0 (TID 0) 15/04/23 17:51:56 INFO executor.Executor: Running task 1.0 in stage 0.0 (TID 1) 15/04/23 17:51:56 INFO executor.Executor: Fetching http://com.hunantv.logservernode2:33381/files/pi.R with timestamp 1429782715611 15/04/23 17:51:56 INFO util.Utils: Fetching http://com.hunantv.logservernode2:33381/files/pi.R to /tmp/fetchFileTemp5460076502660272005.tmp 15/04/23 17:51:56 INFO executor.Executor: Fetching http://com.hunantv.logservernode2:33381/jars/sparkr-assembly-0.1.jar with timestamp 1429782715609 15/04/23 17:51:56 INFO util.Utils: Fetching http://com.hunantv.logservernode2:33381/jars/sparkr-assembly-0.1.jar to /tmp/fetchFileTemp4739048988763487952.tmp 15/04/23 17:51:57 INFO executor.Executor: Adding file:/tmp/spark-64f7170f-21ed-4551-8c1d-1843895a8e47/sparkr-assembly-0.1.jar to class loader WARNING: ignoring environment value of R_HOME 100000 100000 15/04/23 17:51:57 INFO sparkr.RRDD: Times: boot = 0.401 s, init = 0.010 s, broadcast = 0.000 s, read-input = 0.004 s, compute = 0.255 s, write-output = 0.001 s, total = 0.671 s 15/04/23 17:51:57 INFO executor.Executor: Finished task 1.0 in stage 0.0 (TID 1). 622 bytes result sent to driver 15/04/23 17:51:57 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 0.0 (TID 1) in 1324 ms on localhost (1/2) 15/04/23 17:51:57 INFO sparkr.RRDD: Times: boot = 0.396 s, init = 0.006 s, broadcast = 0.003 s, read-input = 0.005 s, compute = 0.302 s, write-output = 0.001 s, total = 0.713 s 15/04/23 17:51:57 INFO executor.Executor: Finished task 0.0 in stage 0.0 (TID 0). 622 bytes result sent to driver 15/04/23 17:51:57 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 1365 ms on localhost (2/2) 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Stage 0 (collect at NativeMethodAccessorImpl.java:-2) finished in 1.386 s 15/04/23 17:51:57 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool 15/04/23 17:51:57 INFO spark.SparkContext: Job finished: collect at NativeMethodAccessorImpl.java:-2, took 1.663123356 s Pi is roughly 3.14104 15/04/23 17:51:57 INFO spark.SparkContext: Starting job: collect at NativeMethodAccessorImpl.java:-2 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Got job 1 (collect at NativeMethodAccessorImpl.java:-2) with 2 output partitions (allowLocal=false) 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Final stage: Stage 1(collect at NativeMethodAccessorImpl.java:-2) 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Parents of final stage: List() 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Missing parents: List() 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Submitting Stage 1 (RRDD[2] at RDD at RRDD.scala:19), which has no missing parents 15/04/23 17:51:57 INFO storage.MemoryStore: ensureFreeSpace(8000) called with curMem=73784, maxMem=277842493 15/04/23 17:51:57 INFO storage.MemoryStore: Block broadcast_1 stored as values in memory (estimated size 7.8 KB, free 264.9 MB) 15/04/23 17:51:57 INFO scheduler.DAGScheduler: Submitting 2 missing tasks from Stage 1 (RRDD[2] at RDD at RRDD.scala:19) 15/04/23 17:51:57 INFO scheduler.TaskSchedulerImpl: Adding task set 1.0 with 2 tasks 15/04/23 17:51:57 WARN scheduler.TaskSetManager: Stage 1 contains a task of very large size (391 KB). The maximum recommended task size is 100 KB. 15/04/23 17:51:57 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 1.0 (TID 2, localhost, PROCESS_LOCAL, 401308 bytes) 15/04/23 17:51:57 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 1.0 (TID 3, localhost, PROCESS_LOCAL, 401308 bytes) 15/04/23 17:51:57 INFO executor.Executor: Running task 1.0 in stage 1.0 (TID 3) 15/04/23 17:51:57 INFO executor.Executor: Running task 0.0 in stage 1.0 (TID 2) 15/04/23 17:51:58 INFO sparkr.RRDD: Times: boot = 0.009 s, init = 0.007 s, broadcast = 0.000 s, read-input = 0.005 s, compute = 0.000 s, write-output = 0.001 s, total = 0.022 s 15/04/23 17:51:58 INFO executor.Executor: Finished task 0.0 in stage 1.0 (TID 2). 618 bytes result sent to driver 15/04/23 17:51:58 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 1.0 (TID 2) in 42 ms on localhost (1/2) 15/04/23 17:51:58 INFO sparkr.RRDD: Times: boot = 0.016 s, init = 0.007 s, broadcast = 0.001 s, read-input = 0.004 s, compute = 0.000 s, write-output = 0.001 s, total = 0.029 s 15/04/23 17:51:58 INFO executor.Executor: Finished task 1.0 in stage 1.0 (TID 3). 618 bytes result sent to driver 15/04/23 17:51:58 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 1.0 (TID 3) in 48 ms on localhost (2/2) 15/04/23 17:51:58 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 1.0, whose tasks have all completed, from pool 15/04/23 17:51:58 INFO scheduler.DAGScheduler: Stage 1 (collect at NativeMethodAccessorImpl.java:-2) finished in 0.054 s 15/04/23 17:51:58 INFO spark.SparkContext: Job finished: collect at NativeMethodAccessorImpl.java:-2, took 0.066888288 s Num elements in RDD 200000
|