Cinchy Platform Documentation
Cinchy v5.6
Search
K

Delimited File

1. Overview

A delimited file is a sequential file with column delimiters. Each delimited file is a stream of records, which consists of fields that are ordered by column. Each record contains fields for one row. Within each row, individual fields are separated by column delimiters.
Example Use Case: You have a delimited file that contains your Employee information. You want to use a batch sync to pull this info into a Cinchy table and liberate your data.
The Delimited File source supports batch syncs.
The Delimited File source does not support Geometry, Geography, or Binary data types.

2. Info Tab

You can review the parameters that can be found in the info tab below (Image 1).

Values

Parameter
Description
Example
Title
Mandatory. Input a name for your data sync
Employee Sync
Version
Mandatory. This is a pre-populated field containing a version number for your data sync. You can override it if you wish.
1.0.0
Parameters
Optional. Review our documentation on Parameters here for more information about this field.
@Filepath
Image 1: The Info Tab

3. Source Tab

The following table outlines the mandatory and optional parameters you will find on the Source tab (Image 2).
Source Details
Schema
Filter
The following parameters will help to define your data sync source and how it functions.
Parameter
Description
Example
(Sync) Source
Mandatory. Select your source from the drop down menu.
Delimited File
Source
The location of the source file. Either a Local upload, Amazon S3, or Azure Blob Storage The following authentication methods are supported per source: Amazon S3: Access Key ID/Secret Access Key Azure Blob Storage: Connection String
Local
Delimiter
Mandatory. The delimiter character used to separate the. text strings. Use U+#### syntax (e.g. U+0001) for unicode characters.
,
Text Qualifier
Mandatory. The text qualifier character, which is used in the event that the delimiter is contained within the row cell. Typically, the text qualifier is a double quote.
"
Header Rows to Ignore
Mandatory. The number of records from the top of the file to ignore before the data starts (includes column header). If you use both useHeaderRecord="true" and HeaderRowsToIgnore = 1, two rows will be ignored. Refer to the below to ensure you are receiving the results you want: One row as headers: useHeaderRecord="true" and HeaderRowsToIgnore = 0
Two rows as headers: useHeaderRecord="true" and HeaderRowsToIgnore = 1 Three rows as headers: useHeaderRecord="true" and HeaderRowsToIgnore = 2
1
Encoding
Optional. The encoding of the file. This default to UTF8, however also supports: UTF8_BOM, UTF16, ASCII.
Use Header Record
Optional. Check this box to use the Header record to match schema. If set to true, fields not present inthe record will default to null.
Path
Mandatory. The path to the source file to load. To upload a local file, you must first insert a Parameter in the Info tab of the connection (ex: filepath). Then, you would reference that same value in this location (Ex: @Filepath). This will then trigger a File Upload option to import your file.
@Filepath
AuthType
This field defines the authentication type for your data sync. Cinchy supports "Access Key" and "IAM" role. When selecting "Access Key", you must provide the key and key secret. When selecting "IAM role", a new field will appear for you to paste in the role's Amazon Resource Name (ARN). You also must ensure that:
The Schema section is where you define which source columns you want to sync in your connection. You can repeat the values for multiple columns.
Parameter
Description
Example
Name
Mandatory. The name of your column as it appears in the source.
Name
Alias
Optional. You may choose to use an alias on your column so that it has a different name in the data sync.
Data Type
Mandatory. The data type of the column values.
Text
Description
Optional. You may choose to add a description to your column.
There are other options available for the Schema section if you click on Show Advanced.
Parameter
Description
Example
Mandatory
  • If both Mandatory and Validated are checked on a column, then rows where the column is empty are rejected
  • If just Mandatory is checked on a column, then all rows are synced with the execution log status of failed, and the source error of "Mandatory Rule Violation"
  • If just Validated is checked on a column, then all rows are synced.
Validate Data
  • If both Mandatory and Validated are checked on a column, then rows where the column is empty are rejected
  • If just Validated is checked on a column, then all rows are synced.
Trim Whitespace
Optional if data type = text. If your data type was chosen as "text", you can choose whether to trim the whitespace (that is, spaces and other non-printing characters).
Max Length
Optional if data type = text. You can input a numerical value in this field that represents the maximum length of the data that can be synced in your column. If the value is exceeded, the row will be rejected (you can find this error in the Execution Log).
You can choose to add in a Transformation > String Replacement by inputting the following:
Parameter
Description
Example
Pattern
Mandatory if using a Transformation. The pattern for your string replacement, i.e. the string that will be searched and replaced.
Replacement
What you want to replace your pattern with.
Note that you can have more than one String Replacement
You have the option to add a source filter to your data sync. Please review the documentation here for more information on source filters.
Image 2: Define your Source

4. Next Steps

Last modified 3mo ago