/sasdmp
.
/opt
or /usr
.
This new directory is referred to as DMPInstallDir throughout
this section.
DMPInstallDir/sasexe/SASDMPHome DMPInstallDir/sasexe/dmsprkhadp-2.40000-1.sh
DMPInstallDir/sasexe
directory.
DMPInstallDir/sasexe/SASDMPHome/bin DMPInstallDir/sasexe/SASDMPHome/dat DMPInstallDir/sasexe/SASDMPHome/etc DMPInstallDir/sasexe/SASDMPHome/lib DMPInstallDir/sasexe/SASDMPHome/share DMPInstallDir/sasexe/SASDMPHome/var
DMPInstallDir/sasexe/SASDMPHome/bin
directory looks like this.DMPInstallDir/sasexe/SASDMPHome/bin/dfwsvc DMPInstallDir/sasexe/SASDMPHome/bin/dfxver DMPInstallDir/sasexe/SASDMPHome/bin/dfxver.bin DMPInstallDir/sasexe/SASDMPHome/bin/sasdmp_admin.sh DMPInstallDir/sasexe/SASDMPHome/bin/settings.sh DMPInstallDir/sasexe/SASDMPHome/bin/dmpsvc
sudo su - root su - hdfs | hdfs-userid kinit -kt location of keytab file user for which you are requesting a ticket exit
hdfs
.
For MapR distributions, the default MapR superuser is mapr
.
You can specify a different user ID with the -hdfsuser argument when
you run the bin/sasdmp_admin.sh -add
script.
klist Ticket cache: FILE/tmp/krb5cc_493 Default principal: hdfs@HOST.COMPANY.COM Valid starting Expires Service principal 06/20/15 09:51:26 06/27/15 09:51:26 krbtgt/HOST.COMPANY.COM@HOST.COMPANY.COM renew until 06/22/15 09:51:26
cd DMPInstallDir/SASDMPHome/
bin/sasdmp_admin.sh -genconfig
bin/sasdmp_admin.sh -add
-host
<hosts>
option.
-check
option.cd DMPInstallDir/SASDMPHome/bin/
bin/sasdmp_admin.sh -check
hadoop fs -ls /sas/ep/config
/sas/ep/config
directory
is created automatically when you run the install script. If you used
-dmpconfig or -genconfig to specify a non-default location, use that
location to find the dmp-config.xml file.
installs SAS Data Management Accelerator for Spark.
Tip | If at a later time you add nodes to the cluster, you can specify the hosts on which you want to install SAS Data Management Accelerator for Spark by using the -hostfile or -host option. The -hostfile and -host options are mutually exclusive. |
See | -hostfile and -host option |
generates the SAS Data Management Accelerator for Spark configuration file in the specified location.
Default | /sas/ep/config/dmp-config.xml |
Interaction | Use the -dmpconfig argument in conjunction with the -add or -remove argument to specify the HDFS location of the configuration file. Use the -genconfig argument when you upgrade to a new version of your Hadoop distribution. |
Tip | Use the -dmpconfig argument to create the configuration file in a non-default location. |
See | -genconfig config-filename -force |
specifies the maximum number of parallel copies between the master and data nodes.
Default | 10 |
Interaction | Use this argument in conjunction with the -add argument. |
specifies the full path of a file that contains the list of hosts where SAS Data Management Accelerator for Spark is installed or removed.
Default | The sasdmp_admin.sh script discovers the cluster topology and uses the retrieved list of data nodes. |
Interaction | Use the -hostfile argument in conjunction with the -add when new nodes are added to the cluster. |
Tip | You can also assign a host list filename to a UNIX variable, SASEP_HOSTS_FILE . export SASEP_HOSTS_FILE=/etc/hadoop/conf/slaves |
See | -hdfsuser user-id |
Example | -hostfile /etc/hadoop/conf/slaves |
specifies the target host or host list where SAS Data Management Accelerator for Spark is installed or removed.
Default | The sasdmp_admin.sh script discovers the cluster topology and uses the retrieved list of data nodes. |
Requirement | If you specify more than one host, the hosts must be enclosed in double quotation marks and separated by spaces. |
Interaction | Use the -host argument in conjunction with the -add when new nodes are added to the cluster. |
Tip | You can also assign a list of hosts to a UNIX variable, SASEP_HOSTS . export SASEP_HOSTS="server1 server2 server3" |
See | -hdfsuser user-id |
Example | -host "server1 server2 server3" -host bluesvr |
specifies the user ID that has Write access to the HDFS root directory.
Default | hdfs for Cloudera, Hortonworks, Pivotal HD, and IBM BigInsights |
mapr for MapR | |
Interaction | Use the -hdfsuser argument in conjunction with the -add or -remove argument to change or remove the HDFS user ID. |
Note | The user ID is used to copy the SAS Data Management Accelerator for Spark configuration files to HDFS. |
writes the installation output to the specified filename.
Interaction | Use the -log argument in conjunction with the -add or -remove argument to write or remove the installation output file. |
removes SAS Data Management Accelerator for Spark.
Tips | You can specify the hosts for which you want to remove SAS Data Management Accelerator for Spark by using the -hostfile or -host option. The -hostfile or -host options are mutually exclusive. |
This argument removes the generated dmp-config.xml file. Use the -keepconfig argument to retain the existing configuration file. | |
See | -hostfile and -host option |
generates a new SAS Data Management Accelerator for Spark configuration file in the specified location.
Default | /sas/ep/config/dmp-config.xml |
Interaction | Use the -dmpconfig argument in conjunction with the -add or -remove argument to specify the HDFS location of the configuration file. Use the -genconfig argument when you upgrade to a new version of your Hadoop distribution. |
Tip | This argument generates an updated dmp-config.xml file. Use the -force argument to overwrite the existing configuration file. |
See | -dmpconfig config-filename |
checks if SAS Data Management Accelerator for Spark is installed correctly on all data nodes.
displays the Hadoop configuration environment.
displays the Hadoop version information for the cluster.
installs a hotfix on an existing SAS Data Management Accelerator for Spark installation.
displays all live DataNodes on the cluster.
displays the Spark version information for the cluster.
validates the install by executing simple Spark and MapReduce jobs.
displays the version of SAS Data Management Accelerator for Spark that is installed.