Adding Nodes to a Cluster

About this task

Complete the following steps to add nodes to a cluster:

Procedure

  1. Prepare all nodes. Although MapR recommends using a Domain Name System (DNS), if you do not, ping the new node from an existing node and vice versa. Use the host name instead of an IP address. If you do not get a response, and if you rule out a network problem, a possible fix is to edit the /etc/hosts files of all nodes in the cluster. All nodes need to be listed in all /etc/hosts files.
  2. Plan which packages to install based on services you want to run on the new nodes. For example, you can base your plan partially on a MapReduce mode, installing the list of packages by mode.
  3. Install MapR Software: On all new nodes, add the MapR Repository. On each new node, install the planned packages.
  4. Configure all new nodes by running configure.sh, as shown in the example in the next section. If you added a ZooKeeper role to a node, run the following command on all nodes with the new ZooKeeper list:
    configure.sh -no-autostart
  5. On all new nodes, format disks for use by MapR even if you plan to re-use a node from another cluster. Format the disks from a re-used node to remove data from the old cluster.
  6. If you manually modified configuration files on the existing nodes and those changes apply to the new nodes, copy only those changes to the respective files on the new nodes.
  7. Start ZooKeeper on all new nodes that have ZooKeeper installed:
    service mapr-zookeeper start
  8. Start Warden on all new nodes:
    service mapr-warden start
  9. Restart services that you reconfigured. Running configure.sh alone does not reconfigure services, such as ZooKeeper. Reconfigured services also require a restart. For example, restart ZooKeeper on each node, one at a time after running configure.sh. Restart the lead ZooKeeper last. Restarting ZooKeeper adds the new nodes into the existing ZooKeeper quorum. Services that need to connect to CLDB do not always discover a newly added CLDB node without restarting warden.
  10. Set up node topology for the new nodes.
  11. On any new nodes running NFS, set up NFS for HA.