SASEP-ADMIN.SH Script

Overview of the SASEP-ADMIN.SH Script

The sasep-admin.sh script enables you to perform the following actions.
  • Install or uninstall the SAS Embedded Process and SAS Hadoop MapReduce JAR files on a single node or a group of nodes.
  • Check if the SAS Embedded Process is installed correctly.
  • Generate a SAS Embedded Process configuration file and write the file to an HDFS location.
  • Create a SAS Hadoop MapReduce JAR file symbolic link in the hadoop/lib directory.
  • Create a HADOOP_JARS.zip file. This ZIP file contains all required client JAR files.
  • Write the installation output to a log file.
  • Display all live data nodes on the cluster.
  • Display the Hadoop configuration environment.
Note: The sasep-admin.sh script must be run from the EPInstallDir/sasexe/SASEPHome/bin directory.
Note: You must have sudo access on the master node only to run the sasep-admin.sh script. You must also have SSH set up in such a way that the master node can passwordless SSH to all data nodes on the cluster where the SAS Embedded Process is installed.

SASEP-ADMIN.SH Syntax

sasep-admin.sh
-add <-link><-epconfig config-filename > <-maxscp number-of-copies>
<-hostfile host-list-filename | -host <">host-list<">>
<-hdfsuser user-id> <-log filename>
sasep-admin.sh
-remove <-epconfig config-filename > <-hostfile host-list-filename | -host <">host-list<">>
<-hdfsuser user-id> <-log filename>
sasep-admin.sh
<-genconfig config-filename <-force>>
<-getjars>
<-linklib | -unlinklib>
<-check> <-hostfile host-list-filename | -host <">host-list<">>
<-env>
<-hadoopversion >
<-log filename>
<-nodelist>
<-version >
<-hotfix >
Arguments

-add

installs the SAS Embedded Process.

Tip If at a later time you add nodes to the cluster, you can specify the hosts on which you want to install the SAS Embedded Process by using the -hostfile or -host option. The -hostfile or -host options are mutually exclusive.
See -hostfile and -host option

-link

forces the creation of SAS Hadoop MapReduce JAR files symbolic links in the hadoop/lib folder during the installation of the SAS Embedded Process.

Restriction This argument should be used only for backward compatibility (that is, when you install the July 2015 release of SAS 9.4 of the SAS Embedded Process on a client that runs the second maintenance release of SAS 9.4).
Requirement If you use this argument, you must restart the MapReduce service, the YARN service, or both after the SAS Embedded Process is installed.
Interactions Use this argument in conjunction with the -add argument to force the creation of the symbolic links.
Use the -linklib argument after the SAS Embedded Process is already installed to create the symbolic links.
See Backward Compatibility
-linklib

-epconfig config-filename

generates the SAS Embedded Process configuration file in the specified location.

Default If the -epconfig argument is not specified, the install script will create the SAS Embedded Process configuration file in the default location /sas/ep/config/ep-config.xml.
Requirement If the -epconfig argument is not specified a configuration file location must be provided. If you choose a non-default location, you must set the sas.ep.config.file property in the mapred-site.xml file that is on your client machine to the non-default location.
Interaction Use the -epconfig argument in conjunction with the -add or -remove argument to specify the HDFS location of the configuration file.
Tip Use the -epconfig argument only if you decide to create the configuration file in a non-default location.
See -genconfig config-filename -force

-maxscp number-of-copies

specifies the maximum number of parallel copies between the master and data nodes.

Default 10
Interaction Use this argument in conjunction with the -add argument.

-hostfile host-list-filename

specifies the full path of a file that contains the list of hosts where the SAS Embedded Process is installed or removed.

Default The sasep-admin.sh script discovers the cluster topology and uses the retrieved list of data nodes.
Interaction Use the -hostfile argument in conjunction with the -add when new nodes are added to the cluster.
Tip You can also assign a host list filename to a UNIX variable, sas_ephosts_file.
export sasep_hosts=/etc/hadoop/conf/slaves
See -hdfsuser user-id
Example
-hostfile /etc/hadoop/conf/slaves

-host <">host-list<">

specifies the target host or host list where the SAS Embedded Process is installed or removed.

Default The sasep-admin.sh script discovers the cluster topology and uses the retrieved list of data nodes.
Requirement If you specify more than one host, the hosts must be enclosed in double quotation marks and separated by spaces.
Interaction Use the -host argument in conjunction with the -add when new nodes are added to the cluster.
Tip You can also assign a list of hosts to a UNIX variable, sas_ephosts.
export sasep_hosts="server1 server2 server3"
See -hdfsuser user-id
Example
-host "server1 server2 server3"
-host bluesvr

-hdfsuser user-id

specifies the user ID that has Write access to HDFS root directory.

Default hdfs for Cloudera, Hortonworks, Pivotal HD, and IBM BigInsights
mapr for MapR
Interaction Use the -hdfsuser argument in conjunction with the -add or -remove argument to change or remove the HDFS user ID.
Note The user ID is used to copy the SAS Embedded Process configuration files to HDFS.

-log filename

writes the installation output to the specified filename.

Interaction Use the -log argument in conjunction with the -add or -remove argument to write or remove the installation output file.

-remove

removes the SAS Embedded Process.

CAUTION:
If you are using SAS Data Loader, you should remove the QKB and the SAS Data Management Accelerator for Spark from the Hadoop nodes before removing the SAS Embedded Process.
For more information, see Removing the QKB or SASDMP_ADMIN.SH Syntax.
Tip You can specify the hosts for which you want to remove the SAS Embedded Process by using the -hostfile or -host option. The -hostfile or -host options are mutually exclusive.
See -hostfile and -host option

-genconfig config-filename <-force>

generates a new SAS Embedded Process configuration file in the specified location.

Requirement There is no default location associated with the -genconfig argument. If you specify the -genconfig argument, you must provide a location. If the provided location already exists, you can overwrite it by specifying the -force argument. The SAS Embedded Process reads its configuration from the default location /sas/ep/config/ep-config.xml. If you decide to generate a new configuration file in a non-default location, you must set the sas.ep.config.file property in the mapred-site.xml file that is on your client machine to the non-default location.
Interaction Use the -genconfig argument to generate a new SAS Embedded Process configuration file when you upgrade to a new version of your Hadoop distribution.
Tip This argument generates an updated ep-config.xml file. Use the -force argument to overwrite the existing configuration file.
See -epconfig config-filename

-getjars

creates a HADOOP_JARS.zip file in the EPInstall dir/SASEPHome/bin directory. This ZIP file contains all required client JAR files.

Restrictions This argument is not supported for MapR distributions.
The -getjars argument is for use only with TKGrid and High-Performance Analytics. It does not gather all of the JAR files that are required for full functionality of SAS software that requires the use of the SAS Embedded Process. Most of the JAR files that are required for full functionality of SAS software are gathered when you install SAS/ACCESS Interface to Hadoop. For more information, see SAS Hadoop Configuration Guide for Base SAS and SAS/ACCESS at http://support.sas.com/resources/thirdpartysupport/v94/hadoop/.
Note In the July 2015 release of SAS 9.4, the SAS_HADOOP_JAR_PATH environment variable has replaced the need for copying the Hadoop JAR files to the client machine with the exception of High-Performance Analytics. The SAS_HADOOP_JAR_PATH environment variable is usually set when you install SAS/ACCESS Interface to Hadoop.
Tip You can move this ZIP file to your client machine and unpack it. If you want to replace the existing JAR files, move it to the same directory where you previously unpacked the existing JAR files.

-linklib

creates SAS Hadoop MapReduce JAR file symbolic links in the hadoop/lib folder.

Restriction This argument should be used only for backward compatibility (that is, when you install the July 2015 release of SAS 9.4 of the SAS Embedded Process on a client that runs the second maintenance release of SAS 9.4).
Requirement If you use this argument, you must restart the MapReduce service, the YARN service, or both after the SAS Embedded Process is installed.
Interaction Use the -linklib argument after the SAS Embedded Process is already installed to create the symbolic links. Use the -link argument in conjunction with the -add argument to force the creation of the symbolic links.
See Backward Compatibility
-link

-unlinklib

removes SAS Hadoop MapReduce JAR file symbolic links in the hadoop/lib folder.

Restriction This argument should be used only for backward compatibility (that is, when you install the July 2015 release of SAS 9.4 of the SAS Embedded Process on a client that runs the second maintenance release of SAS 9.4).
Requirement If you use this argument, you must restart the MapReduce service, the YARN service, or both after the SAS Embedded Process is installed.
See Backward Compatibility

-check

checks if the SAS Embedded Process is installed correctly on all data nodes.

-env

displays the Hadoop configuration environment.

-hadoopversion

displays the Hadoop version information for the cluster.

-nodelist

displays all live DataNodes on the cluster.

-version

displays the version of the SAS Embedded Process that is installed.

-hotfix

distributes a hot fix package.

Requirements Hot fixes must be installed using the same user ID who performed the initial software installation.
Hot fixes should be installed following the installation instructions provided by SAS Technical Support.