Upgrading Spark on YARN
About this task
Procedure
-
Update repositories. MapR's
rpm
anddeb
repositories always contain the Spark version recommended for the MapR core release associated with that repository. You can connect to an internet repository or prepare a local repository with any version of Spark you need. You can also manually download packages. If you plan to install from a repository, complete the following steps each node where Spark is installed: -
Back up any custom configuration files in your Spark environment. These cannot be
upgraded automatically. For example, if Spark SQL is configured to work with Hive, copy
the
/opt/mapr/spark/spark-<version>/conf/hive-site.xml
file to a backup directory. -
Shut down the spark-historyserver services (if the spark-historyserver is running):
maprcli node services -nodes <node-ip> -name spark-historyserver -action stop
-
Install the Spark packages.
NOTE: You only need to upgrade the mapr-spark-historyserver if your previous installation included this package.
- On Ubuntu
-
apt-get install mapr-spark mapr-spark-historyserver
- On RedHat / CentOS
-
yum update mapr-spark mapr-spark-historyserver
-
Run
configure.sh
:/opt/mapr/server/configure.sh -R
-
Migrate Custom Configurations (optional).
Migrate any custom configuration settings into the new default files in the conf directory(/opt/mapr/spark/spark-<version>/conf).
-
Start spark-historyserver services (if installed):
maprcli node services -nodes <node-ip> -name spark-historyserver -action start