Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
0.6.1
-
None
-
None
Description
So if R daemon is required in executor, the R script will fail.
How to reproduce it
df <- createDataFrame(sqlContext, mtcars) showDF(df)
Exception in executor side:
10:16:20,024 INFO org.apache.spark.storage.memory.MemoryStore:54 - Block broadcast_1 stored as values in memory (estimated size 14.2 KB, free 366.3 MB) 10:16:21,018 INFO org.apache.spark.api.r.BufferedStreamThread:54 - Fatal error: cannot open file '/Users/jzhang/Temp/hadoop_tmp/nm-local-dir/usercache/jzhang/appcache/application_1473129941656_0037/container_1473129941656_0037_01_000002/sparkr/SparkR/worker/daemon.R': No such file or directory 10:16:31,023 ERROR org.apache.spark.executor.Executor:91 - Exception in task 0.2 in stage 1.0 (TID 3) java.net.SocketTimeoutException: Accept timed out at java.net.PlainSocketImpl.socketAccept(Native Method) at java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:404) at java.net.ServerSocket.implAccept(ServerSocket.java:545) at java.net.ServerSocket.accept(ServerSocket.java:513) at org.apache.spark.api.r.RRunner$.createRWorker(RRunner.scala:367) at org.apache.spark.api.r.RRunner.compute(RRunner.scala:69) at org.apache.spark.api.r.BaseRRDD.compute(RRDD.scala:49)
Attachments
Issue Links
- links to