Tasks
|
Description
|
---|---|
Copy data to and from
Hadoop
|
Copy relational databases
and SAS data sets to and from Hadoop via parallel, bulk data transfer.
For more information, see Copy Data to Hadoop, and Copy Data from Hadoop.
Import data from delimited
text files, such as comma-separated value (CSV) files. For more information,
see Import a File.
|
Transform and transpose
data in Hadoop
|
Transform data by filtering
rows, managing columns, and summarizing rows. For more information,
see Transform Data in Hadoop.
Select columns and transpose
or group them. For more information,
see Transpose Data in Hadoop.
|
Cleanse data in Hadoop
|
Standardize, match,
parse, and perform other data quality functions on data in Hadoop. For more information,
see Cleanse Data in Hadoop.
Use rules and expressions
to filter data. For more information,
see About DS2 Expressions and the Advanced Editor.
|
Sort or de-duplicate
data in Hadoop
|
Sort data in an existing
table and remove duplicate rows from the table. For more information,
see Sort and De-Duplicate Data in Hadoop.
|
Query or join data in
Hadoop
|
Query a table or join
multiple tables without knowing SQL. For more information,
see Query or Join Data in Hadoop.
Run aggregations on
selected columns. For more information,
see About the Aggregations in the Summarize Rows Transformation.
Power users can generate
and edit a HiveQL query, or paste and run an existing HiveQL query. For more information,
see Run a Hive Program.
|
Profile data and save
profile reports
|
Analyze source columns
from one or more tables to determine patterns, uniqueness, and completeness. For more information,
see Profile Data.
View data profile reports.
Add notes to a data
profile report to explain a result or ask a question.
|
Run user-written code
|
Use the Run a SAS Program
directive to execute user-written Base SAS code or DS2 code. For more information,
see Run a SAS Program.
Use the Run a Hive Program
directive to execute user-written Hive code. For more information,
see Run a Hive Program.
|
Manage and reuse directives
|
Use directives to guide
you through the process of creating and running jobs in Hadoop.
View the status of current
and previous job executions. For more information,
see Run Status.
Stop and start directives.
Open their logs and generated code files.
Run, view, or edit saved
directives for reuse. For more information,
see Saved Directives.
|
Load data to SAS LASR
Analytic Server
|
Load specified Hadoop
columns in memory onto the SAS LASR Analytic Server for analysis
using SAS Visual Analytics or SAS Visual Statistics (licensed separately). For more information,
see Load Data to LASR.
|
Specify global options
|
Specify server connections,
data sources, global options, and other settings for SAS Data Loader. For more information,
see Set Global Options.
|