The SAS Embedded Process
runs within MapReduce to read and write data. The SAS Embedded Process
runs on your Hadoop system where the data lives.
There are two methods
by which you can deploy the SAS Embedded Process:
-
SAS Deployment Manager
If you are using Cloudera,
Hortonworks, IBM BigInsights, or Pivotal HD, you can use the SAS Deployment
Manager to deploy the SAS Embedded Process to the Hadoop cluster.
-
SAS Embedded Process install script
By default, the SAS
Embedded Process install script (sasep-admin.sh) discovers the cluster
topology and installs the SAS Embedded Process on all DataNode nodes,
including the host node from where you run the script (the Hadoop
master NameNode). This occurs even if a DataNode is not present. If
you want to add the SAS Embedded Process to new nodes at a later
time, you can run the sasep-admin.sh script with the -host
<hosts> option.
For distributions that
are running MapReduce 1, the SAS Hadoop MapReduce JAR files are required
in the hadoop/lib
directory. For distributions
that are running MapReduce 2, the SAS Hadoop MapReduce JAR files are
in the EPInstallDir/SASEPHome/jars/
directory.