The Quick Installer is deprecated. Use the MapR Installer instead.
MapR's Quick Install method automates the installation process for you. It is designed to get a small-scale cluster up and running quickly, with a minimum of user intervention. When you run the MapR installer, it checks prerequisites for you, asks you questions about the configuration of your cluster, prepares the system, and installs MapR software. In most cases, the Quick Install method is the preferred installation method.
Review the following table to verify that the Quick Install method is right for you:
|Quick Install||Expert Installation Mode|
This method is best suited for:
You should only consider performing a manual (expert mode) installation if you:
See Advanced Installation Topics for more information.
While the Quick Installation Guide provides a high-level view of the installation process, this document provides more detail to help you with your installation. Topics include:
- Planning (setup requirements and cluster planning)
- Installation Tips (suggestions to help your installation succeed)
- Installation Process (what the installer is doing during the process)
- Successful Installation (how to recognize when the installation completes successfully)
- Bringing Up the Cluster (registering the cluster and applying the license)
This section explains how to prepare for the Quick Install process. Note that the installer performs a series of checks automatically (see Installation Process). In addition to these checks, make sure you meet the following requirements:
- Your nodes either have internet access, or have access to a local package repository containing MapR packages.
- All the nodes in your cluster can communicate with each other over the network. The installer uses port 22 for ssh. In addition, MapR software requires connectivity across other ports between the cluster nodes. For a list of all ports used by MapR, refer to Services and Ports Quick Reference.
- Each node meets the requirements outlined in Preparing Each Node.
Understanding Node Types
The MapR installer categorizes nodes as control nodes, data nodes, control-as-data nodes (which combine the functions of control and data nodes), or client nodes. Clusters generally consist of one, three, or five control nodes and an arbitrary number of data or client nodes.
The following table provides the function of each node type with some additional details:
|control node||Manages the cluster and has cluster management services installed. To simplify the installation process, all control nodes have the same services installed on them. In Expert Mode, you can configure each node so these management services are split across nodes. See Advanced Installation Topics for more information.|
|data node||Used for processing data, so they have the FileServer and TaskTracker services installed. If you run MapR-DB or HBase on a data node, the HBase Client service is also installed. Data nodes are used for running YARN applications and MapReduce jobs, and for storing file and table data. These nodes run the FileServer service along with NodeManager (for YARN nodes), TaskTracker (for MapReduce nodes), and HBase client (for MapR-DB and HBase nodes).|
|control-as-data node||Acts as both control and data nodes. They perform both functions and have both sets of services installed. These nodes are appropriate only for small clusters. Control-as-data nodes act as both control and data nodes. For a single-node cluster, designate the node as control-as-data so it will have control node and data node services installed.|
Provides access to the cluster so you can communicate via the command line or the MapR Control System. Client nodes provide access to each node on the cluster so you can submit jobs and retrieve data. A client node can be an edge node of the cluster, your laptop, or any Windows machine. You can install as many client nodes as you want on your cluster. When you specify a client node, you provide the hostname of the initial control node, which establishes communication with the cluster.
You can use the Quick Installer to install the MapR client and the MapR HBase client. The Quick Installer does not install the MapR POSIX client. See MapR POSIX Client for details.
Node Types and Associated Services
The following table shows which services are assigned to each node type. The main services correspond to the core MapR packages, while the additional services are determined by the type of cluster you specify (MapReduce, MapR-DB, HBase, or a combination). See the Installation section of Installing MapR Software under Advanced Installation Topics for more information on these services.
|Node Type||YARN Main Services||Core MapR Services||Additional|
HistoryServer (on one control node)
HBase Region Server
HBase Region Server
|client node||MapR Client||MapR Client||HBase Client|
Cluster Planning Guidelines
To help you plan your cluster, here are some scenarios that illustrate how to allocate different types of nodes in a cluster. You can adjust these guidelines for your particular situation.
For a 5-node cluster, you can configure one node as a control node (or choose node type control-as-data) and the remaining four nodes as data nodes. To provide high availability (HA) in a 5-node cluster, you need three control nodes. In addition, all the nodes should be able to process data. In this scenario, choose three control-as-data nodes and two data nodes.
Total # Nodes
|Number of |
For a 20-node cluster, you still only need three control nodes to manage the cluster. If you need all nodes to process data, the control nodes can double as data nodes, which means you can choose either control or control-as-data for the node type. The remaining nodes can be dedicated data nodes, as shown.
These tips help you successfully complete the installation process. To begin installation, run the
install command and select one of these options:
new: starts a new installation
add: adds nodes to an existing installation
remove: uninstalls MapR packages from an existing installation so you can start a new installation
If you have an installation configuration file, you can supply the name of the file on the command line and skip the interview questions. For example:
Installing a New Cluster
When you install nodes on a new cluster, select new to indicate that this cluster uses a new configuration. The installer then asks you if you have a configuration file. If you answer yes, the installer prompts you for the name of the configuration file. If you answer no, the installer proceeds to the next step, which is to enter the hostnames (or IP addresses) of all control nodes, separated by spaces or commas. Next, enter the hostnames (or IP addresses) of all data nodes. Make sure all nodes are up and running (
ping <hostname>) and their hostnames are valid.
During the interview process, you have an opportunity to change the username and the MapR user password for security purposes.
After you answer all the questions, the installer displays a summary and asks if you want to modify the settings. When you are satisfied with the settings, select (c)ontinue to begin the installation process.
If you want to save the configuration and resume the installation later, select (a)bort. The next time you run the installer, it displays the following message:
To use the saved configuration file, enter
y for yes.
Ensure that all user information matches across all nodes. Each username and password must match on every node, and must have the same UID. Each groupname must match on every node, and must have the same GID.
To install a client node, select
c from the modify menu, then enter the client hostname or IP address.
This section explains what happens when you run the MapR installer. When you use the installer to interactively install and configure the nodes on your cluster, the installation script is launched and it performs these tasks for you:
- Prepares the system :
- Checks for necessary resources
- Checks to see if another version of Hadoop is already installed (if so, you must uninstall this version before you run the installer).
- Installs and configures OS packages
- Installs Java
- Installs MapR software
- Configures the repositories
- Installs the MapR packages
- Configures MapR software
Various information messages are displayed to your output device while the installer is running. The installer verifies system pre-requisites for you, and then checks your system configuration. Next, it launches the interactive question-and-answer process. When you finish the process (and select continue), the installer displays messages about the tasks it is performing.
During the installation process, the installer asks questions about your cluster configuration. When you finish answering all the questions, the installer displays a summary that includes the choices you selected as well as some other default settings. Here is a sample summary:
This summary displays all the settings for the current node. Note that the installer does not ask you for values for every setting. Instead, it assigns default values to some settings, and then it allows you to change any setting.
At this stage, you can continue with the install, modify the settings, or save the current configuration and continue later.
You can modify any of the settings in the installation summary. If you enter
m to modify settings, the installer displays the following menu:
The following table describes the settings and provides information about modifying them:
The installer assigns a default name,
The cluster name cannot contain spaces.
|MapR User Name|| The installer assigns a default 'mapr' user name, |
|MapR User Group Name|| The default MapR user group name is |
|MapR User ID||The default MapR user ID is 2000. To change this value, enter |
|MapR User Group ID||The default MapR user group ID is 2000 (the same as the MapR user ID). To change this value, enter |
|MapR User Password||The default MapR user password is |
|Security Settings||Basic security (authentication and authorization) measures are automatically implemented on every MapR cluster. An additional layer of security (data encryption, known as wire-level security) is available, but is disabled by default. If you want to enable wire-level security, enter |
|Disks to Use||You must specify which disks to use for the MapR file system for each node. The installer automatically runs the |
|Disk Stripe Width||To configure the number of disks in a storage pool (known as the stripe width), enter |
|Force Formatting Disks||If you have disks with previously installed MapR software, they must be reformatted. Enter |
By default, the quick installer does not install client nodes. To install client nodes in your cluster, select c from the modify menu, then provide the IP address or hostname for each client.
The quick installer only supports Linux-based clients running CentOS, RedHat, or Ubuntu.
The quick installer does not install the MapR POSIX client.
|Control Nodes||If you need to assign the role of control node to different hostnames, enter |
|Data Nodes||If you need to assign the role of data node to different hostnames, enter |
|Control Nodes to Function as Data Nodes||To change the functionality of control nodes so they also function as data nodes, select |
|MapR Software Version||The installer always installs the latest available version of MapR software. You can change the version by entering |
|MapReduce1 Setting||By default, all nodes on a cluster are configured to run YARN services and not MapReduce1 (MapReduce for Hadoop 1) services. If you want to run MapReduce1 on your data nodes (instead of YARN or in addition to YARN), enter |
|MapR-DB Setting||The default setting for MapR-DB is |
|HBase Setting||When the MapR-DB setting is |
|YARN Setting||By default, all nodes on a cluster are configured to run YARN services and not MapReduce1 (MapReduce for Hadoop 1) services. If you want to run MapReduce1 on your data nodes (instead of YARN), enter |
|MapR Core Repo URL||By default, the MapR core repository is located at |
|MapR Ecosystem Repo URL||By default, the MapR ecosystem repository is located at |
|MapR Database Schema Information|
To specify the MySQL database parameters for the MapR metrics database, enter one of the following options from the modify menu:
See Setting up the MapR Metrics Database for more information.
A successful installation takes approximately 10-30 minutes, depending on how long it takes to reach a quorum of ZooKeeper services. This section shows the messages that appear when control nodes are installed successfully. Data node installation starts immediately after control nodes are installed.
Once control nodes have installed successfully, the quick installer immediately starts to install data nodes. In the meantime, you can access the cluster through the MapR Control System (MCS) via the URL shown in the message.
For a cluster that is configured with two data nodes, the following message appears and indicates the continuation of the installation process:
Bringing Up the Cluster
When you finish the installation process, the resulting cluster will have a Community Edition license without NFS. You can see the state of your cluster by logging in to the MapR Control System (MCS).
To get your cluster up and running, follow these steps:
- Register the cluster to obtain a full Community Edition license.
- Apply the license.
- Restart the NFS service.
Registering the Cluster
You can register your cluster through the MapR Control System (MCS). Select Manage Licenses from the navigation pane and follow the instructions.
When the License Management dialog box opens, select Add licenses via Web. The next dialog box provides a link to www.mapr.com, where you can register your cluster.
Applying the License
After you register your cluster, click Apply Licenses in the License Management dialog box. For best results, use an Enterprise Edition license (available as a trial license), which entitles you to run NFS on any node on which it is installed. A Community Edition license limits you to one node for NFS, which means you can only have one control node or one control-as-data node (which runs a control node as a data node).
The last step in bringing up the cluster is to restart NFS. Even though the installer loads the NFS service on all control and control-as-data nodes, NFS requires a license in order to run (which you applied in the previous step). You can restart the NFS service from the MCS, See Manage Node Services for information.
Once NFS is running, the cluster appears at the mount point
/mapr in the Linux file system for all control and control-as-data nodes.