Output Settings

In the Output Settings screen, users can see the various components that make up an Output, including tabs for Output Settings, Mappings, and Output History. When initially configuring an Output, this is the only visible tab.
 

 

Initial Parameters

  • Name: The name of the Output. Every output in the DataForge environment must have a unique name.
  • Description: The description of the Output.
  • Active: If set to Active, the Output will be immediately available for use, and any sources connected to the output through a Channel will automatically run Output at the end of all processes.
 

Output Type

It is important to decide which Output Type an Output is. There are three types described below. The parameters available will dynamically change depending on users' selections.
 
Batch sources can be mapped to any output type. Stream sources support mapping to output types of Event, Delta Table, and Virtual.
 
Event:
DataForge will publish data to an Event topic. For more information on setting up an Event output, refer to the User Manual's Kafka Events Connections documentation.
 
File:
DataForge will output a File using a File Connection.
There are five Output File Types: Avro, CSV, JSON, Parquet, or Text. Parameter selections will update dynamically depending on the selection.
 
Table:
DataForge can output and refresh data to a database Table using a Table Connection. The type of table output is defined by the Driver option on the Output Connection used. 
 
Virtual:
Rather than push data out to a separate system, DataForge can also manage a database view on top of the Hub tables within the connected Databricks environment, otherwise known as a Virtual Output.
 

Output Parameters

Asterisks (*) in the Parameter Name column indicate mandatory parameters

Output

Appears Under Parameter Name Default Value Description
Table, File Connection*   Name of the DataForge connection used to write the output to the desired destination
Table Table Name*   Name of output table to be written in the target DB
Table Table Schema*   Name of the schema the output table will be written to in the target DB.
Table Output Source Id Partition TRUE Designates whether output_source_id will be top level partition and only applies to Delta Lake outputs.
Table Manage Table TRUE Decides if you want table to be altered if there are missing columns, or automatically updated if new columns are added
Table
Reset Channels On
Delta Overwrite
TRUE Reset all other output channels when Delta Lake table is overwritten due to schema changes. Only applies to Delta Lake output.
Table Delete Orphan Data TRUE Delete data from output table that was created by a channel that no longer exists. Does not delete rows where s_output_source_id is null
Table Create Cci On Table FALSE Creates clustered columnstore index on destination table.  We recommend CCI on SQL server output tables to improve performance and avoid deadlocks.
Virtual View Name*   Name of the Virtual Table/View that will appear in the Databricks environment once the output is processed.
Virtual View Database Default Name of database for virtual view in Databricks. 
File Type: All File Name*   File name for output file. For all except text, extension will be added automatically.
File Type: Text File Extension*   Extension that will be appended to the file name when DataForge writes text file outputs
File Type: All Single File

TRUE

(FALSE for Parquet)

Toggle for whether or not the output should be written to multiple files or one single file. Multiple files is more performant
File Type: Avro, CSV, JSON, Parquet Limit By Effective Range FALSE Output uses effective range calculations to limit data for Time Series sources
Event Value Schema Type avro_from_registry

Schema type for selected topic. Used to associate data with schema type during output

Event Key Schema Type avro_from_registry

Schema type for selected topic. Used to associate data with schema type during output

Event Key Schema  

Json (avro) or text (json DDL) specified

Event Value Schema  

Json (avro) or text (json DDL) specified

Event Key Subject  

Schema registry subject for key schema

Event Value Subject  

Schema registry subject for value schema

Output Retention

Appears Under Parameter Default Value Description
File Type: CSV, Parquet Archive Files 1 year How long files will remain in archive folder.
File Type: CSV, Parquet Buffer Files 0 Interval to retain files in fast output buffer storage
File Type: CSV, Parquet Temporary Files 14 days
Amount of time temporary files will remain stored

Output Alerts

Appears Under Parameter Default Value Description
All Output Failure Topics   List of AWS Simple Notification Service topic ARNs to be notified upon Output Failure
All Output Success Topics   List of AWS Simple Notification Service topic ARNs to be notified upon Output Success
All Output Failure Emails   List of email addresses to be notified upon Output Failure.  Requires SMTP setup.
All Output Success Emails   List of email addresses to be notified upon Output Success.  Requires SMTP setup.

 

Post Output Commands *

When Custom Notebook value is selected, a Custom Cluster Configuration for the Post-Output step needs to be selected.
 

Updated

Was this article helpful?

0 out of 0 found this helpful