The SASHDAT engine
is used with a Hadoop cluster that is configured to use the SAS Plug-ins
for Hadoop, or with machines that use NFS to mount a distributed file
system.
The engine is designed as a write-only engine for transferring data. However, SAS
High-Performance Analytics procedures are designed to read data in parallel from a
co-located data provider. The LASR procedure, and other procedures such as HPREG and HPLOGISTIC, can read
data with the engine. The HPDS2 procedure is designed to read data and write data
in parallel. The HPDS2 procedure can be used with the engine to read data and create
new SASHDAT tables.
Whenever a SAS High-Performance
Analytics procedure is used to create a SASHDAT table, the procedure
creates the data with a default block size of 2 megabytes.
This
size can be overridden with the BLOCKSIZE= data set option.