SetEnv
option
set in the dfs_serv.xml
configuration file.
Here is an example:<OptionSet name="SetEnv"> <Option name="SAS_HADOOP_JAR_PATH">\SAS\Config\Lev1\FederationServer \lib\Hadoop</Option> </OptionSet>
Option
|
Description
|
---|---|
DRIVER
|
DRIVER=HIVE Required. Identifies
the HIVE data source to which you want to connect.
|
SERVER
|
SERVER=‘server-name’ Specifies the host name
of the Hive server. If the server name contains spaces or nonalphanumeric
characters, enclose it in quotation marks.
|
PORT
|
PORT=port_number Specifies the port number
that is used to connect to the specified Hive Server. The default
is 10000.
|
SUBPROTOCOL
|
SUBPROTOCOL=Hive|Hive2 Specifies whether you
are connecting to a Hive service or a HiveServer2 (Hive2) service.
The default is Hive2.
|
CATALOG
|
CATALOG=catalog-name Specifies an arbitrary identifier for an SQL catalog, which groups logically related schemas.
|
SCHEMA
|
SCHEMA=hive-schema-name Specifies a Hive schema
name that is used to define a name other than ‘default’.
Alias: DATABASE, DB
|
AUTH_MODE
|
AUTH_MODE=default
| Kerberos Specifies the authentication mode for the connection. The options are default and
Kerberos. If using Kerberos
for the authentication mode, you must specify the Hive principal host name using the
HIVE_PRINCIPAL connection option.
|
HIVE_PRINCIPAL
|
HIVE_PRINCIPAL=service-principal-hostname Specifies the Hive principal
string in an environment that uses Kerberos(for example,
HIVE_PRINCIPAL=hive/_HOST@UNX.SAS.COM ).
Required with AUTH_MODE=Kerberos.
|
USER_PRINCIPAL
|
USER_PRINCIPAL= Specifies that the HDFS
path and JDBC path use JAAS to perform a doAs for the given Kerberos
user principal. Alias:
auth_mode=Kerberos;uid= |
UID
|
UID=(user-name) Specifies the user name with the necessary permissions to perform Read and Write operations.
UID and PWD are not needed when connecting to a server that uses Kerberos authentication.
Alias: USER
|
PWD
|
PWD=’user-password’ Specifies a password that correlates with the user ID (UID) value. If the password
contains spaces or nonalphanumeric characters, enclose it in quotation marks. UID
and PWD are not needed when connecting to a server that uses Kerberos authentication.
Alias: PASSWORD
|
PROPERTIES (JDBC session
configuration properties)
|
Use the PROPERTIES option
to specify one or more JDBC connection properties to override the
default JDBC connection properties. In a JDBC URL, custom properties
are separated from the default properties by the question mark (?)
character. Multiple properties are separated by the semicolon (;)
character. Here is an example:
PROPERTIES="hive.default.fileformat=ORC;hive.exec.compress.intermediate=true" Site-wide Hive properties
are specified in the
hive-site.xml file in
the Hive configuration directory.
You can use the properties
option to set the default file format for managed and unmanaged tables
respectively. The following example specifies optimized row columnar
(ORC) as the default file format when creating a table:
PROPERTIES=hive.default.fileformat.managed=orc PROPERTIES=hive.default.fileformat=orc Note: The ORC file format is available
beginning with Hive version 0.11.
The following example sets the partition mode to nonstrict, which allows dynamic inserts
against a partitioned table (for example, when a static partition has not been explicitly
defined in the SQL statement).
PROPERTIES=hive.exec.dynamic.partition.mode=nonstrict Note: These Hive–defined
properties can be changed or removed by Hadoop vendors at any time.
|
HD_CONFIG
|
HD_CONFIG=path
to hadoop configuration file Specifies the name and
path for the Hadoop cluster configuration file. This file contains
entries for Hadoop system information, including file system properties
such as fs.defaultFS. The configuration file can be a copy of the
Hadoop core-site.xml file. However, if your Hadoop cluster is running
with HDFS failover enabled, you must create a file that combines the
contents of the Hadoop core-site.xml and hdfs-site.xml files.
Note: The name and path values
for the Hadoop cluster configuration file are normally set in the
SAS_HADOOP_CONFIG_PATH option in the dfs_serv.xml configuration file.
This is the recommended configuration method. Use HD_CONFIG only if
you want to override the settings in the server configuration file.
See "SAS Federation Server Configuration Reference" for details.
|
HDFS_TEMPDIR
|
HDFS_TEMPDIR=’path’ Specifies the path to
the HDFS directory that is used for read and write of temporary data.
The default is
HDFS_TEMPDIR=’/tmp’ |
DBMAX_TEXT
|
DBMAX_TEXT=32767 Specifies the length for a string data type. The maximum length is 2 gigabytes. The default is 32767.
|
LOGIN_TIMEOUT
|
LOGIN_TIMEOUT=number_of_seconds Specifies a login time-out, in seconds, for non-responsive connections. A value of 0 indicates that there is
no time-out and the connection will ‘wait forever’. The default value is 30 seconds.
|
DRIVER TRACE
|
DRIVER_TRACE=’API
| SQL | ALL’; Requests tracing information, which logs transaction records to an external file that
can be used for debugging purposes. The SAS Federation Server driver writes a record
of each command that is sent to the database to the trace log based
on the specified tracing level, which determines the type of tracing information. The tracing levels are as follows:
Default: Tracing
is not activated.
Note: If
you activate tracing, you must also specify the location of the trace
log with DRIVER_TRACEFILE=. Note that DRIVER_TRACEFILE= is resolved
against the TRACEFILEPATH set in ALTER SERVER. TRACEFILEPATH is relative
to the server's content root location.
(Optional) You can control trace
log formatting with DRIVER_TRACEOPTIONS=.
Interaction: You
can specify one trace level, or you can concatenate more than one
by including the | (OR) symbol. For example,
driver_trace='api|sql' generates tracing information for API calls and SQL statements.
|
DRIVER TRACE FILE
|
DRIVER_TRACEFILE=’filename’ Used to specify the name of the
text file for the trace log. Include the filename and extension in
single or double quotation marks. For example:
driver_tracefile='\mytrace.log' Default: The
default TRACEFILE location applies to a relative filename, and it
is placed relative to TRACEFILEPATH.
Requirement: DRIVER_TRACEFILE
is required when activating tracing using DRIVER_TRACE.
Interaction: (Optional)
You can control trace log formatting with DRIVER_TRACEOPTIONS=.
|
DRIVER TRACE OPTIONS
|
DRIVER_TRACEOPTIONS=APPEND
| THREADSTAMP | TIMESTAMP; Specifies options
in order to control formatting and other properties for the trace
log:
Default: The trace log is overwritten with no thread identification or time stamp.
|
Windows
|
The options are located
on the Advanced or Performance tabs
in the ODBC Administrator.
|
UNIX
|
The options are available
when configuring data sources using the
dfdbconf tool.
Values can also be set by editing the odbc.ini file
in which their data sources are defined.
|
STRING
, which is a variable-length character string with a maximum size of 2G. As a result,
this can create very large character fields when processing data. Since Hive’s string
type is comparable to VARCHAR
in
other data sources, you can set the ODBC attribute, Max
Varchar Size
to specify the maximum character string
size. Set the Max Varchar Size
value using
Advanced Options in Windows ODBC Administrator, or in UNIX by editing odbc.ini
in
the specified path or $HOME directory.
CONOPTS
container.
Here is an example: DRIVER=ODBC;DB=hive;UID=dbitest;PWD=dbigrp1;SCHEMA=default;CONOPTS=(MaxVarcharSize=300);CATALOG=FOO;
.