Brief description of SAS Data integration Studio Transformation
Introduction to Transformations
You
want to select the right transformation to perform a specific task. The
transformation enables you to include that task in a SAS Data Integration
Studio job flow.
A transformation is a metadata object that specifies how to
extract data, transform data, or load data into data stores. Each
transformation that you specify in a process flow diagram generates or
retrieves SAS code. You can also specify user-written code in the metadata for
any transformation in a process flow diagram.
Overview of the Transformations Tree
The
Transformations tree organizes transformations into a set of folders. You can
drag a transformation from the Transformations tree to the Job Editor, where
you can connect it to source and target tables and update its default metadata.
By updating a transformation with the metadata for actual sources, targets, and
transformations, you can quickly create process flow diagrams for common
scenarios. The following display shows the standard Transformations tree.
This document has an example of the main transformations
used in SAS Data Integration Studio, and the online Help has an example of all
transformations. The following sections describe the contents of the
Transformations tree folders.
Access Folder
The
following table describes the transformations in the Access folder in the
Transformations tree.
Access Folder Transformations
|
|
|
Used to bulk load SAS and most DBMS source tables to a DB2
target table. For more information, see About the DB2 Bulk Table Loader.
|
|
Reads an external file and writes to a SAS or DBMS table.
For more information, see Using an External File
in the Process Flow for a Job.
|
|
Reads a SAS or DBMS table and writes to an external file.
For more information, see Using an External File
in the Process Flow for a Job..
|
|
Generates an output table that lists the tables contained
in an input library. If there is no input library, then the transformation
generates a list of tables from all of the libraries that are allocated on
the SAS Workspace Server. For more information, see Creating a Control
Table.
|
|
Delivers content from a Microsoft MQ message queue to SAS
Data Integration Studio. If the message is being sent into a table, the
message queue content is sent to a table or a SAS Data Integration Studio
transformation. If the message is being sent to a macro variable or file,
then these files or macro variables can be referenced by a later step. For
more information, see Processing a Microsoft
Queue.
|
|
Enables writing files in binary mode, tables, or
structured lines of text to the WebSphere MQ messaging system. The queue and
queue manager objects necessary to get to the messaging system are defined in
SAS Management Console. For more information, see Processing a Microsoft
Queue.
|
|
Enables bulk loading of SAS or Oracle source data into an
Oracle target. For more information, see About the Oracle Bulk
Table Loader Transformation.
|
|
Reads a source and writes to a SAS SPD Server target.
Enables you to specify options that are specific to SAS SPD Server tables.
For more information, see About the SPD Server
Table Loader Transformation.
|
|
Reads a source table and writes to a target table.
Provides more loading options than other transformations that create tables.
For more information, see About the Table Loader
Transformation.
|
|
Enables you to set table options unique to Teradata tables
and supports the pushdown feature that enables you to process relational
database tables directly on the appropriate relational database server. For
more information, see Teradata Table Loader
Transformation
|
|
Delivers content from a WebSphere MQ message queue to SAS
Data Integration Studio. If the message is being sent into a table, the
message queue content is sent to a table or a SAS Data Integration Studio
transformation. If the message is being sent to a macro variable or file,
then these files or macro variables can be referenced by a later step. For
more information, see Processing a WebSphere
Queue.
|
|
Enables writing files in binary mode, tables, or
structured lines of text to the WebSphere MQ messaging system. The queue and
queue manager objects necessary to get to the messaging system are defined in
SAS Management Console. For more information, see Processing a WebSphere
Queue.
|
|
Puts data into an XML table. In a SAS Data Integration
Studio job, if you want to put data into an XML table, you must use an XML
Writer transformation. For example, you cannot use the Table Loader
transformation to load an XML table. For more information, see Converting a SAS or
DBMS Table to an XML Table.
|
The
following table describes the transformations in the Analysis folder in the
Transformations tree.
Analysis Folder Transformations
|
|
|
Creates an output table that contains correlation
statistics. For more information, see Creating a Correlation
Analysis.
|
|
Creates an output table that contains a distribution
analysis. For more information, see Creating a Distribution
Analysis.
|
|
Enables you to run the High-Performance Forecasting
procedure (PROC HPF) against a warehouse data store. PROC HPF provides a
quick and automatic way to generate forecasts for many sets of time series or
transactional data. For more information, see Generating Forecasts.
|
|
Creates an output table that contains frequency
information. For more information, see Frequency of Eye Color
By Hair Color Crosstabulation.
|
|
Creates a one-way output table that contains frequency
information about the relationship between two classification variables. For
more information, see One-Way Frequency of
Eye Color By Region.
|
|
Creates an output table that contains summary statistics.
For more information, see Creating Summary
Statistics for a Table.
|
|
Creates an output table that contains descriptive
statistics in tabular format, using some or all of the variables in a data
set. It computes many of the same statistics that are computed by other
descriptive statistical procedures such as MEANS, FREQ, and REPORT. For more
information, see Creating a Summary
Tables Report from Table Data.
|
Archived Folder
In
order to support backwards compatibility for existing processes and guarantee
that processes run exactly as defined using older transformations, SAS has
developed a methodology for archiving older versions of transformations in the
Process library. The process library continues to surface the archived
transformations for some number of releases. When a job is opened that contains
a newer transformation replacement, a dialog box is displayed and indicates the
name of the old transformation. The dialog box also provides the name and
location of the new transformation in the process library tree.
The following table describes the deprecated and archived
transformations in the Archived Transforms folder in the Transformations tree.
Archived Transforms Folder Transformations
|
|
|
Loads source data into a fact table and translates
business keys into generated keys.
This older transformation is marked with a flag on its
icon. This flag indicates that the transformation is an older version of an
updated transformation. For information about the current version,
|
Change Data Capture Folder
Change
data capture (CDC) is a process that shortens the time required to load data
from relational databases. The CDC loading process is more efficient because
the source table contains changed data only. The changed data table is much
smaller than the relational base table. The following table describes the
transformations in the Change Data Capture folder in the Transformations tree.
Change Folder Transformations
|
|
|
Loads changed data only from Attunity and other selected
databases. For more information, see Working with Change
Data Capture.
|
|
Loads changed data only from DB2 databases. For more
information, see Working with Change Data
Capture.
|
|
Loads changed data only from a wide range of databases.
For more information, see Working with Change
Data Capture.
|
|
Loads changed data only from Oracle databases. For more
information, see Working with Change
Data Capture.
|
Control Folder
The
following table describes the transformations in the Control folder in the
Transformations tree.
Control Folder Transformations
|
|
|
Marks the beginning of the iterative processing sequence
in an iterative job. For more information, see Creating and Running an
Iterative Job.
|
|
Marks the end of the iterative processing sequence in an
iterative job. For more information, see Creating and Running an
Iterative Job.
|
|
Provides status-handling logic at a desired point in the
process flow diagram for a job. Can be inserted between existing transformations
and removed later without affecting the mappings in the original process
flow. For more information, see Perform Actions Based
on the Status of a Transformation.
|
Data Folder
The
following table describes the transformations in the Data Transforms folder in
the Transformations tree.
Data Folder Transformations
|
|
|
Creates a single target table by combining data from
several source tables. For more information, see Creating a Table That
Appends Two or More Source Tables.
|
|
Enables you to detect changes between two tables such as
an update table and a master table and generate a variety of output for matched
and unmatched records. For more information, see Comparing Tables.
|
|
Moves data directly from one machine to another. Direct
data transfer is more efficient than the default transfer mechanism. For more
information, see Moving Data Directly
from One Machine to Another Machine.
|
|
Cleanses data before it is added to a data warehouse or
data mart. For more information, see Validating Product Data.
|
|
Enables change tracking in intersection tables. For more
information, see Tracking Changes in
Source Datetime Values.
|
|
Loads a target with columns taken from a source and from
several lookup tables. For more information, see Loading a Fact Table
Using Dimension Table Lookup.
|
|
Integrates a SAS Enterprise Miner model into a SAS Data
Integration Studio data warehouse. Typically used to create target tables
from a SAS Enterprise Miner model. For more information, see Integrating a SAS
Enterprise Miner Model with Existing SAS Data.
|
|
Ranks one or more numeric column variables in the source
and stores the ranks in the target. For more information, see Create a Table That
Ranks the Contents of a Source.
|
|
Enables you to load a dimension table using type 1
updates. Type 1 updates insert new rows, update existing rows, and generate
surrogate key values in a dimension table without maintaining a history of
data changes. Each business key is represented by a single row in the
dimension table. For more information, see Loading a Dimension
Table with Type 1 Updates.
|
|
Loads source data into a dimension table, detects changes
between source and target rows, updates change tracking columns, and applies
generated key values. This transformation implements slowly changing
dimensions. For more information, see Loading a Dimension
Table with Type 1 and 2 Updates.
|
|
Reads data from a source, sorts it, and writes the sorted
data to a target. For more information, see Creating a Table That
Contains the Sorted Contents of a Source.
|
|
Selects multiple sets of rows from one source and writes
each set of rows to a different target. Typically used to create two or more
subsets of a source. Can also be used to create two or more copies of a
source. For more information, see Create Two Tables That
Are Subsets of a Source.
|
|
Creates an output table that contains data standardized to
a particular number. For more information, see Creating Standardized
Statistics from Table Data.
|
|
Loads a target, adds generated whole number values to a
surrogate key column, and sorts and saves the source based on the values in
the business key column or columns. For more information, see Loading a Table and
Adding a Surrogate Primary Key.
|
|
Creates an output table that contains transposed data. For
more information, see Creating Transposed
Data from Table Data.
|
|
Retrieves a user-written transformation. Can be inserted
between existing transformations and removed later without affecting the
mappings in the original process flow. Can also be used to document the
process flow for the transformation so that you can view and analyze the
metadata for a user-written transformation, similarly to how you can analyze
metadata for other transformations. For more information, see Adding a User Written
Code Transformation to a Job.
|
Data Quality Folder
The
following table describes the transformations in the Data Quality folder in the
Transformations tree. In general, you can use Apply Lookup Standardization,
Create Match Code, and Standardize with Definition for data cleansing
operations. You can use DataFlux Batch Job and DataFlux Data Service to perform
tasks that are a specialty of DataFlux software, such as profiling, monitoring,
or address verification.
|
|
Apply Lookup Standardization
|
Enables you to select and apply DataFlux schemes that
standardize the format, casing, and spelling of character columns in a source
table. For more information, see Standardizing Values with a Standardization
Scheme.
|
|
Enables you to analyze source data and generate match
codes based on common information shared by clusters of records. Comparing
match codes instead of actual data enables you to identify records that are
in fact the same entity, despite minor variations in the data. For more
information, see Using Match Codes to Improve Record Matching.
|
|
Enables you to select and execute a DataFlux job that is
stored on a DataFlux Data Management Server. You can execute DataFlux Data
Management Studio data jobs, process jobs, and profiles. You can also execute
Architect jobs that were created with DataFlux® dfPower® Studio. For more
information, see Executing a DataFlux Job from SAS Data
Integration Studio.
|
|
Enables you to select and execute a data job that has been
configured as a real-time service and deployed to a DataFlux Data Management
Server. For more information, see Using a DataFlux Data Service in a Job.
|
Standardize with Definition
|
Enables you to select and apply DataFlux standardization
definitions to elements within a text string. For example, you might want to
change all instances of “Mister” to “Mr.” but only when “Mister” is used as a
salutation. For more information, see Standardizing Values with a Definition.
|
Hadoop Folder
The
Hadoop folder is experimental.
Output Folder
The
following table describes the transformations in the Output folder in the
Transformations tree.
Output Folder Transformations
|
|
|
Creates an HTML report that contains selected columns from
a source table. For more information, see Creating Reports from
Table Data.
|
Publish Folder
The
following table describes the transformations in the Publish folder in the
Transformations tree.
Publish Folder Transformations
|
|
|
Creates an HTML report and an archive of the report. For
more information, see Creating a Publish to
Archive Report from Table Data.
|
|
Creates an HTML report and e-mails it to a designated
address. For more information, see Creating a Publish to
Email Report from Table Data.
|
|
Creates an HTML report and publishes it to a queue using
MQSeries. For more information, see Creating a Publish to
Queue Report from Table Data.
|
SPD Server Dynamic Cluster Folder
The
following table describes the transformations in the SPD Server Dynamic Cluster
folder in the Transformations tree.
SPD Server Dynamic Cluster Folder Transformations
|
|
Create or Add to a Cluster
|
Creates or updates an SPD Server cluster table. For more
information, see Creating an SPD Server
Cluster Table.
|
|
Lists the contents of an SPD Server cluster table. For
more information, see Maintaining an SPD
Server Cluster.
|
Remove Cluster Definition
|
Deletes an SPD Server cluster table. For more information,
see Maintaining an SPD
Server Cluster.
|
SQL Folder
The
following table describes the transformations in the SQL folder in the
Transformations tree. For more information, see Working with SQL Join Transformations and Working with Other SQL Transformations.
SQL Folder Transformations
|
|
|
Provides a simple SQL interface for creating tables.
|
|
Generates a PROC SQL statement that deletes user-selected
rows in a single target table. Supports delete, truncate, or delete with a
WHERE clause. Also supports implicit and explicit pass-through.
|
|
Enables you to specify custom SQL code to be executed and
provides SQL templates for supported databases.
|
|
Selects multiple sets of rows from a source and writes
those rows to a target. Typically used to create one subset from a source.
Can also be used to create columns in a target that are derived from columns
in a source. For more information, see Extracting Data from a
Source Table.
|
|
Provides a simple SQL interface for inserting rows into a
target table. For more information, see Inserting Rows into a Target Table.
|
|
Selects multiple sets of rows from one or more sources and
writes each set of rows to a single target. Typically used to merge two or
more sources into one target. Can also be used to merge two or more copies of
a single source. For more information, see Creating a Simple SQL
Query.
|
|
Inserts new rows and updates existing rows using the SQL
Merge DML command. The command was officially introduced in the SQL:2008
standard.
|
|
Enables you to use set operators to combine the results of
table-based queries. For more information, see Using the SQL Set
Operators Transformation.
|
|
Updates user-selected columns in a single target table.
The target columns can be updated by case, constant, expression, or subquery.
Handles correlated subqueries.
|
Note: Some functions in the Delete, Execute, Insert Rows,
Merge, and Update transformations might work only when the table comes from a
database management system that provides an implementation of an SQL command
for which a SAS/ACCESS interface is available. One example is sort. You can use
SAS tables and tables from database management systems that do not implement
the SQL command, but these command-specific functions might not work.
Ungrouped Folder
The
Ungrouped folder contains any transformations that have been created with the
Transformation Generator wizard and not assigned to a transformation category.
The folder is displayed only when a generated transformation is present. It is
displayed only to other users when the generated transformations are placed in
the Shared Data folder.
Good for quick revision! Nice work!
ReplyDeleteIndian Cyber Army Provides Best Ethical Hacking Training in India with an opportunity to work with Indian cyber army.Indian Cyber Army credibility in Ethical hacking training & Cybercrime investigation training is acknowledged across nation as we offer hands on practical knowledge and full assistance with basic as well as advanced level ethical hacking & cybercrime investigation courses in online and offline mode both. https://www.ica.in/blog/ethical-hacking-training
ReplyDeletesalt likit
ReplyDeletesalt likit
dr mood likit
big boss likit
dl likit
dark likit
SPİ1NC