ePrivacy and GPDR Cookie Consent by Cookie Consent Skip to main content

Loader Postgres

PostgreSQL, often referred to as Postgres, is an object-relational database management system (RDBMS) known for its extensibility, robustness, and compliance with SQL standards. Postgres uses and extends the SQL language combined with many other features that safely store and scale the most complicated data workloads. 

Business value in CDP

Postgres can be an efficient loader to export data due to its ability to handle diverse data types, support complex queries, and maintain data integrity. Its extensibility allows integration with other tools and systems, ensuring seamless data transfer and transformation from the CDP to Postgres. This capability makes Postgres a valuable component in building a reliable and scalable data infrastructure, enabling organizations to store and analyze customer data effectively.


Setting up the loader in MI

The Postgres loader is used to load data into a table in the PostgreSQL database.

The native data types that are supported include: 

  • Primitives: Integer, Numeric, String, Boolean
  • Structured: Date/Time, Array, Range, UUID 
  • Document: JSON/JSONB, XML, Key-value (Hstore) 
  • Geometry: Point, Line, Circle, Polygon 
  • Customizations: Composite, Custom Types

Learn more: about Postgres from PostgreSQL tutorials

Requirements

Credentials of the database you would like to connect to:

  • Host
  • Port
  • Name of the database, schema
  • Username
  • Password

If you wish to connect to a database outside of the Meiro  Integrations environment, you may need a front-end tool like PgAdmin4, Postico, DBeaver or any other front-end (GUI) tool available online. To learn more about these tools, please refer to their respective documentation online. Connecting via a GUI tool is not mandatory, but it may be useful if you want to access the database outside the Meiro  Integrations environment.

Postgres components work only with CSV tables. Other data formats need to be converted to CSV format by processors.

Features 

  • One file and multi-file tables,
  • Header and no header,
  • Custom delimiter, escape and quotation characters,
  • Automatic creation of the destination table (all data types as a text for columns) and manual override if the user wants to specify the data type for a column.

Data In/Data Out

Data In

Data from the previous component should be saved as a CSV table in the folder /data/in/tables/folder_name.


File structure:

ID,CAR,COST,DATE_OF_SALE

1,BWM,400000,2019-09-10

2,Volvo,3000,2019-01-10


Or without header

 1,BWM,400000,2019-09-10

2,Volvo,3000,2019-01-10

Warning:  json data should be escaped in the input file.


Data Out N/A

Learn more: about the folder structure here.

Parameters

Table
DB Table Name (required) Table’s name in the database.
Auto Create(true/ false) If true, auto-creates columns by headers in CSV files, the type is text, nullable true and default is an empty string. Columns definitions will be ignored.
Delimiter (required) Customer delimiter.
Quotation (required) Custom quotation.
Escape(required) Custom escape character.
Headers (required) If true, headers are specified in the CSV files, if false columns will be populated in order in which they are entered.
Path to File/ Folder (required)

Path to the folder with data folders with specifications on the name of the table.  For example,
./data/in/tables/Saving_Product_MarFebJanDec


For loading all CSV files to the database, folder path /data/in/tables  or /data/in/tables/folder_name  is sufficient enough. A component goes through all CSV files in the folder and load all of them into a table in a database.


For loading one specific CSV file to the table in a database use a file path without  .csv  suffix --> /data/in/tables/file_name

Load method (required)

full_load, insert, upsert

Full_load method

Method deletes given tables if exists. Re-creates them and inserts data from file(s).


Note that when using this method, it is important to disable auto create.

Insert method The method inserts data into the tables and ignores conflicts.
Upsert method Method upsert data to the tables. In case of conflicts, data will be updated. 
Primary key (required) List of primary keys (from columns specified further).


Column
Column Name (required) Refers to the column in the input CSV file.
DB Column Name (required) Refers to how this column will be named in the destination database.

Data Type (required)

The data type used in the column.

Size (optional)

The maximum number of digits used by the data type of the column or parameter.

Nullable (optional)

If checked, then the data with empty strings are loaded as NULL values in the destination database (in this case is_nullable = YES).


If unchecked and data contains empty strings, then loader returns an error message. For example: “Cannot insert the value NULL into column XY, column does not allow nulls. INSERT fails.”


Warning: All columns listed as primary keys should have unchecked Nullable (is_nullable = NO), otherwise the loader throws an error “Cannot define PRIMARY KEY constraint on nullable column”.


Default Value (optional)

The default value is inserted for empty values in the column. For example,  inserts NULLs in the destination columns database for empty strings in csv file. - inserts - for empty strings.

 
DB
Host (required) Name of the server that hosts the database.
Port (required) The port number you want to access. 
Schema (required) Organization unit in the database.
Database (required) Name of the database.
User (required) Account name. 
Password (required) Account password. 

Remember: You should be able to get all the credentials from the database administrator.