azurermDataFactoryDatasetDelimitedText
Manages an Azure Delimited Text Dataset inside an Azure Data Factory.
Example Usage
/*Provider bindings are generated by running cdktf get.
See https://cdk.tf/provider-generation for more details.*/
import * as azurerm from "./.gen/providers/azurerm";
/*The following providers are missing schema information and might need manual adjustments to synthesize correctly: azurerm.
For a more precise conversion please use the --provider flag in convert.*/
const azurermResourceGroupExample = new azurerm.resourceGroup.ResourceGroup(
this,
"example",
{
location: "West Europe",
name: "example-resources",
}
);
const azurermDataFactoryExample = new azurerm.dataFactory.DataFactory(
this,
"example_1",
{
location: azurermResourceGroupExample.location,
name: "example",
resource_group_name: azurermResourceGroupExample.name,
}
);
/*This allows the Terraform resource name to match the original name. You can remove the call if you don't need them to match.*/
azurermDataFactoryExample.overrideLogicalId("example");
const azurermDataFactoryLinkedServiceWebExample =
new azurerm.dataFactoryLinkedServiceWeb.DataFactoryLinkedServiceWeb(
this,
"example_2",
{
authentication_type: "Anonymous",
data_factory_id: azurermDataFactoryExample.id,
name: "example",
url: "https://www.bing.com",
}
);
/*This allows the Terraform resource name to match the original name. You can remove the call if you don't need them to match.*/
azurermDataFactoryLinkedServiceWebExample.overrideLogicalId("example");
const azurermDataFactoryDatasetDelimitedTextExample =
new azurerm.dataFactoryDatasetDelimitedText.DataFactoryDatasetDelimitedText(
this,
"example_3",
{
column_delimiter: ",",
data_factory_id: azurermDataFactoryExample.id,
encoding: "UTF-8",
escape_character: "f",
first_row_as_header: true,
http_server_location: [
{
filename: "fizz.txt",
path: "foo/bar/",
relative_url: "http://www.bing.com",
},
],
linked_service_name: azurermDataFactoryLinkedServiceWebExample.name,
name: "example",
null_value: "NULL",
quote_character: "x",
row_delimiter: "NEW",
}
);
/*This allows the Terraform resource name to match the original name. You can remove the call if you don't need them to match.*/
azurermDataFactoryDatasetDelimitedTextExample.overrideLogicalId("example");
Argument Reference
The following supported arguments are common across all Azure Data Factory Datasets:
-
name
- (Required) Specifies the name of the Data Factory Dataset. Changing this forces a new resource to be created. Must be globally unique. See the Microsoft documentation for all restrictions. -
dataFactoryId
- (Required) The Data Factory ID in which to associate the Linked Service with. Changing this forces a new resource. -
linkedServiceName
- (Required) The Data Factory Linked Service name in which to associate the Dataset with. -
folder
- (Optional) The folder that this Dataset is in. If not specified, the Dataset will appear at the root level. -
schemaColumn
- (Optional) AschemaColumn
block as defined below. -
description
- (Optional) The description for the Data Factory Dataset. -
annotations
- (Optional) List of tags that can be used for describing the Data Factory Dataset. -
parameters
- (Optional) A map of parameters to associate with the Data Factory Dataset. -
additionalProperties
- (Optional) A map of additional properties to associate with the Data Factory Dataset.
The following supported locations for a Delimited Text Dataset (exactly one of them must be set):
-
azureBlobFsLocation
- (Optional) AnazureBlobFsLocation
block as defined below. -
azureBlobStorageLocation
- (Optional) AnazureBlobStorageLocation
block as defined below. -
httpServerLocation
- (Optional) AhttpServerLocation
block as defined below.
The following supported arguments are specific to Delimited Text Dataset:
-
columnDelimiter
- (Optional) The column delimiter. Defaults to,
. -
rowDelimiter
- (Optional) The row delimiter. Defaults to any of the following values on read:\r\n
,\r
,\n
, and\n
or\r\n
on write by mapping data flow and Copy activity respectively. -
encoding
- (Optional) The encoding format for the file. -
quoteCharacter
- (Optional) The quote character. Defaults to"
. -
escapeCharacter
- (Optional) The escape character. Defaults to\
. -
firstRowAsHeader
- (Optional) When used as input, treat the first row of data as headers. When used as output, write the headers into the output as the first row of data. Defaults tofalse
. -
nullValue
- (Optional) The null value string. Defaults to an empty string. Defaults to""
. -
compressionCodec
- (Optional) The compression codec used to read/write text files. Valid values arenone
,bzip2
,gzip
,deflate
,zipDeflate
,tarGzip
,tar
,snappy
andlz4
. Please note these values are case sensitive. -
compressionLevel
- (Optional) The compression ratio for the Data Factory Dataset. Valid values arefastest
oroptimal
. Please note these values are case sensitive.
A schemaColumn
block supports the following:
-
name
- (Required) The name of the column. -
type
- (Optional) Type of the column. Valid values arebyte
,byte[]
,boolean
,date
,dateTime
,dateTimeOffset
,decimal
,double
,guid
,int16
,int32
,int64
,single
,string
,timeSpan
. Please note these values are case sensitive. -
description
- (Optional) The description of the column.
An azureBlobFsLocation
block supports the following:
-
fileSystem
- (Required) The storage data lake gen2 file system on the Azure Blob Storage Account hosting the file. -
path
- (Optional) The folder path to the file. -
filename
- (Optional) The filename of the file.
An azureBlobStorageLocation
block supports the following:
-
container
- (Required) The container on the Azure Blob Storage Account hosting the file. -
path
- (Optional) The folder path to the file. This can be an empty string. -
filename
- (Optional) The filename of the file. -
dynamicContainerEnabled
- (Optional) Is thecontainer
using dynamic expression, function or system variables? Defaults tofalse
. -
dynamicPathEnabled
- (Optional) Is thepath
using dynamic expression, function or system variables? Defaults tofalse
. -
dynamicFilenameEnabled
- (Optional) Is thefilename
using dynamic expression, function or system variables? Defaults tofalse
.
A httpServerLocation
block supports the following:
-
relativeUrl
- (Required) The base URL to the web server hosting the file. -
path
- (Required) The folder path to the file on the web server. -
filename
- (Required) The filename of the file on the web server. -
dynamicPathEnabled
- (Optional) Is thepath
using dynamic expression, function or system variables? Defaults tofalse
. -
dynamicFilenameEnabled
- (Optional) Is thefilename
using dynamic expression, function or system variables? Defaults tofalse
.
Attributes Reference
The following attributes are exported:
id
- The ID of the Data Factory Dataset.
Timeouts
The timeouts
block allows you to specify timeouts for certain actions:
create
- (Defaults to 30 minutes) Used when creating the Data Factory Dataset.update
- (Defaults to 30 minutes) Used when updating the Data Factory Dataset.read
- (Defaults to 5 minutes) Used when retrieving the Data Factory Dataset.delete
- (Defaults to 30 minutes) Used when deleting the Data Factory Dataset.
Import
Data Factory Datasets can be imported using the resourceId
, e.g.