PostgreSQL, often referred to as Postgres, is an object-relational database management system.
Postgres uses and extends the SQL language combined with many other features that safely store and scales the most complicated data workloads.
The native data types that are supported include:
Primitives: Integer, Numeric, String, Boolean
Structured: Date/Time, Array, Range, UUID
Document: JSON/JSONB, XML, Key-value (Hstore)
Geometry: Point, Line, Circle, Polygon
Customizations: Composite, Custom Types
Learn more: about Postgres from PostgreSQL tutorials.
The Postgres loader is used to load data into a table in the PostgreSQL database.
Credentials of the database you would like to connect to:
- Name of the database, schema
If you wish to connect to a database outside of the Meiro Integrations environment, you may need a front-end tool like PgAdmin4, Postico, DBeaver or any other front-end (GUI) tool available online. To learn more about these tools, please refer to their respective documentation online. Connecting via a GUI tool is not mandatory, but it may be useful if you want to access the database outside the Meiro Integrations environment.
Postgres components work only with CSV tables. Other data formats need to be converted to CSV format by processors.
- One file and multi-file tables,
- Header and no header,
- Custom delimiter, escape and quotation characters,
- Automatic creation of the destination table (all data types as a text for columns) and manual override if the user wants to specify the data type for a column.
Data In/Data Out
Data from the previous component should be saved as a CSV table in the folder
Or without header
Learn more: about the folder structure here.
|DB Table Name (required)||Table’s name in the database.|
|Auto Create(true/ false)||If true, auto-creates columns by headers in CSV files, the type is text, nullable true and default is an empty string. Columns definitions will be ignored.|
|Delimiter (required)||Customer delimiter.|
|Quotation (required)||Custom quotation.|
|Escape(required)||Custom escape character.|
|Headers (required)||If true, headers are specified in the CSV files, if false columns will be populated in order in which they are entered.|
|Path to File/ Folder (required)||
Path to the folder with data folders with specifications on the name of the table. For example,
For loading all CSV files to the database, folder path
For loading one specific CSV file to the table in a database use a file path without
|Load method (required)||
|Full_load method||Method deletes given tables if exists. Re-creates them and inserts data from file(s).|
|Insert method||The method inserts data into the tables and ignores conflicts.|
|Upsert method||Method upsert data to the tables. In case of conflicts, data will be updated.|
|Primary key (required)||List of primary keys (from columns specified further).|
|Column Name (required)||Refers to the column in the input CSV file.|
|DB Column Name (required)||Refers to how this column will be named in the destination database.|
Data Type (required)
|The data type used in the column.|
|The maximum number of digits used by the data type of the column or parameter.|
If checked, then the data with empty strings are loaded as NULL values in the destination database (in this case is_nullable = YES).
If unchecked and data contains empty strings, then loader returns an error message. For example: “Cannot insert the value NULL into column XY, column does not allow nulls. INSERT fails.”
Warning: All columns listed as primary keys should have unchecked Nullable (is_nullable = NO), otherwise the loader throws an error “Cannot define PRIMARY KEY constraint on nullable column”.
|Default Value (optional)||
The default value is inserted for empty values in the column. For example,
|Host (required)||Name of the server that hosts the database.|
|Port (required)||The port number you want to access.|
|Schema (required)||Organization unit in the database.|
|Database (required)||Name of the database.|
|User (required)||Account name.|
|Password (required)||Account password.|
Remember: You should be able to get all the credentials from the database administrator.