Integrate Spark with R
Integrate Spark with R when you want to run R programs as Spark jobs.
About this task
As of Spark 1.5.2, you can integrate Spark with R.
Procedure
-
Install R 3.2.2 or greater on each node that will submit Spark jobs:
- On Ubuntu:
apt-get install r-base-dev
- On CentOS/RedHat:
yum install R
For more information on installing R, see the R documentation.
- On Ubuntu:
-
To verify the integration, run the following commands as the mapr user or as a
user that mapr impersonates: