Details
-
Question
-
Status: Resolved
-
Major
-
Resolution: Invalid
-
2.1.0
-
None
-
None
-
Hadoop, Spark
Description
I need to start running data mining analysis in secured environment (IP, Port, and database name are given), where Spark runs on hive tables. So I have installed R, SparkR, dplyr, and some other r libraries.
Now I understand that I need to point sparkR to that database(with IP/Port/Name).
What should be my R code?
I start with evoking R,
then SparkR library.
Next I right sc<-sparkR.init()
it tells me immediately that spark-submit command:not found
Do I need to have RHive installed first?
Or should I actually point somehow to spark library and to that database?
I couldn't find any documentation on that.
Thank you