Copyright © 2007-2024 JumpMind, Inc

Version 3.15.11

Permission to use, copy, modify, and distribute this SymmetricDS User Guide for any purpose and without fee is hereby granted in perpetuity, provided that the above copyright notice and this paragraph appear in all copies.

Preface

This user guide introduces SymmetricDS and its features for data synchronization. It is intended for users, developers, and administrators who want to install the software, configure synchronization, and manage its operation. Thank you to all the members of the open source community whose feedback and contributions helped us build better software and documentation. This version of the guide was generated on 2024-12-05.

1. Introduction

SymmetricDS is open source software for database and file synchronization, with support for multi-master replication, filtered synchronization, and transformation. It uses web and database technologies to replicate change data as a scheduled or near real-time operation, and it includes an initial load feature for full data loads. The software was designed to scale for a large number of nodes, work across low-bandwidth connections, and withstand periods of network outage.

1.1. System Requirements

SymmetricDS is written in Java and requires a Java Runtime Environment (JRE) Standard Edition (SE) or Java Development Kit (JDK) Standard Edition (SE). Most major operating systems and databases are supported. See the list of supported databases in the Database Compatibility section. The minimum operating system requirements are:

  • Java SE Runtime Environment 17 or newer

  • Memory - 64 (MB) available

  • Disk - 256 (MB) available

The memory, disk, and CPU requirements increase with the number of connected clients and the amount of data being synchronized. The best way to size a server is to simulate synchronization in a lower environment and benchmark data loading. However, a rule of thumb for servers is one server-class CPU with 2 GB of memory for every 500 MB/hour of data transfer and 350 clients. Multiple servers can be used as a cluster behind a load balancer to achieve better performance and availability.

SymmetricDS Pro is accessed from a web console, which requires one of the following supported web browsers:

  • Chrome - Evergreen

  • Firefox - Evergreen, ESR

  • Safari - 15 or newer

  • Edge - Evergreen, Chromium

1.2. Overview

A node is responsible for synchronizing the data from a database or file system with other nodes in the network using HTTP. Nodes are assigned to one of the node Groups that are configured together as a unit. The node groups are linked together with Group Links to define either a push or pull communication. A pull causes one node to connect with other nodes and request changes that are waiting, while a push causes one node to connect with other nodes when it has changes to send.

Each node is connected to a database with a Java Database Connectivity (JDBC) driver using a connection URL, username, and password. While nodes can be separated across wide area networks, the database a node is connected to should be located nearby on a local area network for the best performance. Using its database connection, a node creates tables as a Data Model for configuration settings and runtime operations. The user populates configuration tables to define the synchronization and the runtime tables capture changes and track activity. The tables to sync can be located in any Catalog and Schema that are accessible from the connection, while the files to sync can be located in any directory that is accessible on the local server.

overview

At startup, SymmetricDS looks for Node Properties Files and starts a node for each file it finds, which allows multiple nodes to run in the same instance and share resources. The property file for a node contains its external ID, node group, registration server URL, and database connection information. The external ID is the name for a node used to identify it from other nodes. One node is configured as the registration server where the master configuration is stored. When a node is started for the first time, it contacts the registration server using a registration process that sends its external ID and node group. In response, the node receives its configuration and a node password that must be sent as authentication during synchronization with other nodes.

1.3. Architecture

Each subsystem in the node is responsible for part of the data movement and is controlled through configuration. Data flows through the system in the following steps:

  1. Capture into a runtime table at the source database

  2. Route for delivery to target nodes and group into batches

  3. Extract and transform into the rows, columns, and values needed for the outgoing batch

  4. Send the outgoing batch to target nodes

  5. Receive the incoming batch at the target node

  6. Transform into the rows, columns, and values needed for the incoming batch

  7. Load data and return an acknowledgment to the source node

architecture
Capture

Change Data Capture (CDC) for tables uses database triggers that fire and record changes as comma-separated values into a runtime table called DATA. For file sync, a similar mechanism is used, except changes to the metadata about files are captured. The changes are recorded as insert, update, and delete event types. The subsystem installs and maintains triggers on tables based on the configuration provided by the user, and it can automatically detect schema changes on tables and regenerate triggers.

Route

Routers run across new changes to determine which target nodes will receive the data. The user configures which routers to use and what criteria is used to match data, creating subsets of rows if needed. Changes are grouped into batches and assigned to target nodes in the DATA_EVENT and OUTGOING_BATCH tables.

Extract

Changes are extracted from the runtime tables and prepared to be sent as an outgoing batch. If large objects are configured for streaming instead of capture, they are queried from the table. Special event types like "reload" for Initial Loads are also processed.

Transform

If transformations are configured, they operate on the change data either during the extract phase at the source node or the load phase at the target node. The node’s database can be queried to enhance the data. Data is transformed into the tables, rows, columns, and values needed for either the outgoing or incoming batch.

Outgoing

The synchronization sends batches to target nodes to be loaded. Multiple batches can be configured to send during a single synchronization. The status of the batch is updated on the OUTGOING_BATCH table as it processes. An acknowledgment is received from target nodes and recorded on the batch.

Incoming

The synchronization receives batches from remote nodes and the data is loaded. The status of the batch is updated on the INCOMING_BATCH table as it processes. The resulting status of the batch is returned to the source node in an acknowledgment.

1.4. Features

SymmetricDS offers a rich set of features with flexible configuration for large scale deployment in a mixed environment with multiple systems.

  • Web UI - The web console provides easy configuration, management, and troubleshooting.

  • Data Synchronization - Change data capture for relational databases and file synchronization for file systems can be periodic or near real-time, with an initial load feature to fully populate a node.

  • Central Management - Configure, monitor, and troubleshoot synchronization from a central location where conflicts and errors can be investigated and resolved.

  • Automatic Recovery - Data delivery is durable and low maintenance, withstanding periods of downtime and automatically recovering from a network outage.

  • Secure and Efficient - Communication uses a data protocol designed for low bandwidth networks and streamed over HTTPS for encrypted transfer.

  • Transformation - Manipulate data at multiple points to filter, subset, translate, merge, and enrich the data.

  • Conflict Management - Enforce consistency of two-way synchronization by configuring rules for automatic and manual resolution.

  • Extendable - Scripts and Java code can be configured to handle events, transform data, and create customized behavior.

  • Deployment Options - The software can be installed as a self-contained server that stands alone, deployed to a web application server, or embedded within an application.

1.5. Why SymmetricDS?

SymmetricDS is a feature-rich data synchronization solution that focuses on ease of use, openness, and flexibility. The software encourages interoperability and accessibility for users and developers with the availability of source code, an application programming interface (API), and a data model supported by documentation. Configuration includes a powerful set of options to define node topology, communication direction, transformation of data, and integration with external systems. Through scripts and Java code, the user can also extend functionality with custom behavior. With a central database for setup and runtime information, the user has one place to configure, manage, and troubleshoot synchronization, with changes taking immediate effect across the network.

The trigger-based data capture system is easy to understand and widely supported by database systems. Table synchronization can be setup by users and application developers without requiring a database administrator to modify the server. Triggers are database objects written in a procedural language, so they are open for examination, and include flexible configuration options for conditions and customization. Some overhead is associated with triggers, but they perform well for applications of online transaction processing, and their benefits of flexibility and maintenance outweigh the cost for most scenarios.

Using an architecture based on web server technology, many simultaneous requests can be handled at a central server, with proven deployments in production supporting more than ten thousand client nodes. Large networks of nodes can be grouped into tiers for more control and efficiency, with each group synchronizing data to the next tier. Data loading is durable and reliable by tracking batches in transactions and retrying of faults for automatic recovery, making it a low maintenance system.

1.6. License

SymmetricDS Pro is commercial software that is licensed, not sold. It is subject to the terms of the End User License Agreement (EULA) and any accompanying JumpMind Support Contract. See the standard SymmetricDS Pro license for reference, but your agreement with JumpMind may be different.

2. Installation

SymmetricDS at its core is a web application. A SymmetricDS instance runs within the context of a web application container like Jetty or Tomcat, and uses web based protocols like HTTP to communicate with other instances.

An instance has one of the following installation options:

  1. Standalone Installation - SymmetricDS is installed and run as a standalone process using the built-in Jetty web server. This is the simplest and recommended way to install an instance.

  2. Web Archive (WAR) - A SymmetricDS web archive (WAR) file is deployed to an existing web application container that is separately installed, maintained and run.

  3. Embedded - SymmetricDS is embedded within an existing application. In this option, a custom wrapper program is written that calls the SymmetricDS API to synchronize data.

See also:

2.1. Standalone Installation

The SymmetricDS Pro setup program is an executable JAR file that can run on any system with a Java Runtime Environment (JRE). See System Requirements for prerequisites. Download the setup program from SymmetricDS Pro Downloads.

Run the setup program:

  • From a desktop environment, double click the symmetric-pro-<version>-setup.jar file

  • If double clicking doesn’t work, use a command prompt to run: java -jar symmetric-pro-<version>-setup.jar

  • From a text-based environment, use a terminal to run: java -jar symmetric-pro-<version>-setup.jar -console

install1

The first screen shows the SymmetricDS Pro software version. The setup program will ask a series of questions before writing files to disk.

To begin selecting options, click Next.

install2

Carefully read the SymmetricDS Pro License Agreement.

If you accept, select I accept the terms of this license agreement and click Next.

install3

Specify Install new software to install a new version of SymmetricDS for the first time.

For upgrading an existing installation of SymmetricDS, see Upgrading.

Click Next to continue.

install4

Choose the installation path where SymmetricDS will either be installed or upgraded. If the directory does not already exist, it will be created for you. Make sure your user has permission to write to the file system.

After entering the directory path, click Next.

install5

Select the packages you want to install and verify disk space requirements are met. By default, all packages are selected. Drivers for popular databases are included, but they can be unselected if you don’t plan to use them.

After selecting packages, click Next.

install6

SymmetricDS can either be run automatically by the system or manually by the user. Select the Install service to run automatically checkbox to install a Windows service or Unix daemon that will start SymmetricDS when the computer is restarted. The service can installed or uninstalled later using the Control Center or command line (see Running as a Service).

Select the Run server after installing checkbox to also run SymmetricDS after installation so it can be used immediately.

After selecting options, click Next.

install7

HTTPS and HTTPS/2 protocols are recommended for protecting data security. For testing without security or encryption, the HTTP protocol can be enabled. Choose an available port number to listen on, which will be validated.

Java Management eXtension (JMX) is an optional way to manage the server from third party tools like JConsole. Most installations leave it disabled and use the web console for management.

Click Next to continue.

install8

Specify how much memory to use for sending and receive data changes. More memory is needed to communicate with multiple clients and when data contains large objects (LOB). Estimate an extra 5 MB of memory for each client and each 500 MB/hour of data transfer.

Click Next to continue.

install9

Specify disk space options for temporarily staging incoming and outgoing data changes. Using staging helps the overall performance of the system and minimizes use of the database. The default location is the "tmp" sub-directory of the installation directory. For Clustering, specify a common network share.

Click Next to continue.

install10

Confirm your installation settings look correct.

Click Next to begin installing files.

install11

The packages you selected are installed to disk.

After it finishes, click Next.

install12

During the finish step, it will install the service and start the service if you selected those options.

After it finishes, click Next.

install13

The installation is now complete. Choose if you want to open the SymmetricDS Pro Control Center where you can view the server status and open a web console.

Click Done to exit the setup program.

install14

From the SymmetricDS Pro Control Center, you can start/stop the server, open the web console, and install/uninstall the service.

To begin configuration of SymmetricDS, check that the server is running, and then click Open Web Console.

To continue setup and configuration of SymmetricDS, refer to the Setup section.

2.2. Running as a Service

SymmetricDS can be configured to start automatically when the system boots, running as a Windows service or Linux/Unix daemon. A wrapper process starts SymmetricDS and monitors it, so it can be restarted if it runs out of memory or exits unexpectedly. The wrapper writes standard output and standard error to the logs/wrapper.log file.

For SymmetricDS Pro, you may have already installed as a service, so this section will show you how to manually install the service from command line.

2.2.1. Running as a Windows Service

To install the service, run the following command as Administrator:

bin\sym_service.bat install

Most configuration changes do not require the service to be re-installed. To uninstall the service, run the following command as Administrator:

bin\sym_service.bat uninstall

To start and stop the service manually, run the following commands as Administrator:

bin\sym_service.bat start
bin\sym_service.bat stop

2.2.2. Running as a Linux/Unix daemon

An init script is written to the system /etc/init.d directory. Symbolic links are created for starting on run levels 2, 3, and 5 and stopping on run levels 0, 1, and 6. To install the script, running the following command as root:

bin/sym_service install

Most configuration changes do not require the service to be re-installed. To uninstall the service, run the following command as root:

bin/sym_service uninstall

To start and stop the service manually, run the following commands:

bin/sym_service start
bin/sym_service stop

2.3. Clustering

A single SymmetricDS node can be deployed across a series of servers to cooperate as a cluster. A node can be clustered to provide load balancing and high availability.

Each node in the cluster shares the same database. A separate hardware or software load balancer is required to receive incoming requests and direct them to one of the backend nodes. Use the following steps to setup a cluster:

  1. Set the cluster.lock.enabled property to true

  2. Optionally, set the cluster.server.id property to a unique name, otherwise the hostname will be used

  3. Set the sync.url property to the URL of the load balancer

  4. Set the initial.load.use.extract.job.enabled property to false if using local staging

  5. Copy the engine properties, security/keystore, and conf/sym_service.conf files to each installation

  6. Configure the load balancer for sticky sessions

  7. If the load balancer requires cookies, then client nodes need to set the server.http.cookies.enabled property to true in conf/symmetric-server.properties

With the cluster.lock.enabled property set to true, jobs will acquire an entry in the LOCK table to ensure that only one instance of the job runs across the cluster. When a lock is acquired, a row is updated in the lock table with the time of the lock and the server ID of the locking job. The locking server ID defaults to the host name, but it can specified with the cluster.server.id property if nodes are running on the same server. Another instance of the job cannot acquire a lock until the locking instance releases the lock and sets the lock time back to null. If an instance is terminated while the lock is still held, an instance with the same server ID is allowed to re-acquire the lock. If the locking instance remains down, the lock can be broken after it expires, specified by the cluster.lock.timeout.ms property. Jobs refresh their lock periodically as they run, which prevents a lock from expiring due to a long run time.

The load balancer should be configured to use sticky sessions if the cluster will receive push synchronization. Push connections first request a reservation from the target node and then connect again using the reservation to push changes. Sticky sessions ensures that the push request is sent to the same server where the reservation is held.

Staging is writing batches to disk before sending over the network, which can use local disk or a shared network drive. Staging can improve performance by reducing the time that resources are held open in the database and by extracting batches before they are served. To use local staging in a cluster, disable the initial.load.use.extract.job.enabled property so the initial load will extract batches on the node serving the request, rather than extracting in the background on a different node. To use shared staging in a cluster, set the staging.dir property to the directory path of the network drive and enable the cluster.staging.enabled property so files are locked during use. With shared staging, the initial load extracts in the background on one node, but batches can be served from any of the nodes in the cluster, which can improve performance.

When deploying nodes in a cluster to an application server like Tomcat or JBoss, the application server does NOT need any clustering of sessions configured.

2.4. Other Deployment Options

It is recommended that SymmetricDS is installed as a standalone service, however there are two other deployment options.

2.4.1. Web Archive (WAR)

This option means packaging a WAR file and deploying to your favorite web server, like Apache Tomcat. It’s a little more work, but you can configure the web server to do whatever you need. SymmetricDS can also be embedded in an existing web application, if desired. As a web application archive, a WAR is deployed to an application server, such as Tomcat, Jetty, or JBoss. The structure of the archive will have a web.xml file in the WEB-INF folder, an appropriately configured symmetric.properties file in the WEB-INF/classes folder, and the required JAR files in the WEB-INF/lib folder.

symmetric war
Figure 1. War

A war file can be generated using the standalone installation’s symadmin utility and the create-war subcommand. The command requires the name of the war file to generate. It essentially packages up the web directory, the conf directory and includes an optional properties file. Note that if a properties file is included, it will be copied to WEB-INF/classes/symmetric.properties. This is the same location conf/symmetric.properties would have been copied to. The generated war distribution uses the same web.xml as the standalone deployment.

bin/symadmin -p my-symmetric-ds.properties create-war /some/path/to/symmetric-ds.war

2.4.2. Embedded

This option means you must write a wrapper Java program that runs SymmetricDS. You would probably use Jetty web server, which is also embeddable. You could bring up an embedded database like Derby or H2. You could configure the web server, database, or SymmetricDS to do whatever you needed, but it’s also the most work of the three options discussed thus far.

The deployment model you choose depends on how much flexibility you need versus how easy you want it to be. Both Jetty and Tomcat are excellent, scalable web servers that compete with each other and have great performance. Most people choose either the Standalone or Web Archive with Tomcat 5.5 or 6. Deploying to Tomcat is a good middle-of-the-road decision that requires a little more work for more flexibility.

A Java application with the SymmetricDS Java Archive (JAR) library on its classpath can use the SymmetricWebServer to start the server.

import org.jumpmind.symmetric.SymmetricWebServer;

public class StartSymmetricEngine {

    public static void main(String[] args) throws Exception {

        SymmetricWebServer node = new SymmetricWebServer(
                                   "classpath://my-application.properties", "conf/web_dir");

        // this will create the database, sync triggers, start jobs running
        node.start(8080);

        // this will stop the node
        node.stop();
    }

This example starts the SymmetricDS server on port 8080. The configuration properties file, my-application.properties, is packaged in the application to provide properties that override the SymmetricDS default values. The second parameter to the constructor points to the web directory. The default location is web. In this example the web directory is located at conf/web_dir. The web.xml is expected to be found at conf/web_dir/WEB-INF/web.xml.

2.4.3. Client Mode

This option runs the SymmetricDS engine without a web server, so it can initiate push and pull requests, but not receive them. Without the web server, there are no open ports listening for sync requests, which can help with security requirements. Be aware that this also means losing access to the web console at this node and any enhanced troubleshooting provided by remote status.

The conf/sym_service.conf file has a parameter to start the service in client mode:

wrapper.app.parameter.3=--client

2.5. Upgrade Considerations

Upgrading an existing SymmetricDS instance depends on the deployment option used originally.

  1. Standalone Installation: The SymmetricDS Pro setup program will automatically detect an older version and upgrade it.

  2. Clustered Installation has an added challenge of coordinating the database schema upgrade and synchronizing the keystore. See Upgrade SymmetricDS Cluster for more details.

  3. Web Archive (WAR): Use commands described in Web Archive (WAR) to generate a new WAR file and install that as before.

  4. Embedded Installation: As noted in the Embedded section, you would need to swap all SymmetricDS Java Archive (JAR) library on the classpath (used for your Java application) for their new version.

  5. Rollback: a combination of restoring database from a prior backup and the The SymmetricDS Pro setup program will overwrite current installation with an older version. See Downgrade Considerations for more details and options.

2.5.1. Upgrade SymmetricDS Cluster

Upgrading a cluster with multiple SymmetricDS installations requires careful coordination. Below are the most important steps. However, it is recommended to test the entire upgrade process in a non-Production environment and to measure time each step took.

In order to minimize risk of downtime, some IT departments require a verified rollback plan. To ensure your rollback plan works, practice it in a non-Production environment.

Contact JumpStart to request Jumpmind experts help you plan and implement a complicated upgrade in the shortest amount of time possible.

  1. Delete any offline nodes, which are not absolutely necessary.

  2. Run the SymmetricDS purge jobs on-demand: Purge Outgoing, Purge Incoming.

  3. Deactivate/disable all servers with SymmetricDS engines in the load balancer.

  4. Stop all SymmetricDS engines (on all servers in the cluster);

  5. Hold any external application jobs or batch processes.

  6. Database backup:

    • Disconnect all users from the database (use a single-user DBA mode, if available).

    • Take a full database backup (to support your rollback plan).

    • Exit single user DBA mode after database backup is complete.

  7. Install SymmetricDS on all new servers (upgrade in place is possible, but slows down rollback plan).

  8. Migrate SymmetricDS configuration from an old server:

    • Parameters: copy the engine .properties file from the old SymmetricDS server to a new one.

    • Logging: edit the log4j2.xml file on the new SymmetricDS installation to include features from the old one log4j file.

  9. First start and initialization:

    • Start only one SymmetricDS engine

    • Allow SymmetricDS the time necessary to execute any database change (schema upgrade scripts). Duration of this step varies as it depends on the version number you are upgrading from.

    • Enable the first SymmetricDS server in the load balancer;

    • Test the first server using the /ping URL.

    • Log into the SymmetricDS console and look for alerts.

    • Review symmetricds.log for any errors. If an error is linked to the SSL certificate CA(unrecognised issuing authority), use the [Import URL] button in the console (Manage→Security→Authorities tab). This will update the keystore file.

  10. Clone SymmetricDS cluster configuration to new servers

    • Copy the engine .properties file from the old SymmetricDS server to a new one.

    • Edit the log4j2.xml file on the new SymmetricDS installation and manually migrate all features from the older log4j configuration.

    • Copy the entire security sub-directory (contains keystore, which must be kept in-sync for all nodes in this cluster!).

    • Start SymmetricDS engine on additional servers in the cluster and monitor errors in the log;

    • Enable additional SymmetricDS servers in the load balancer;

    • Test additional SymmetricDS servers by using the /ping URL;

  11. Test the cluster - one server at a time:

    • Stop the SymmetricDS engine;

    • Reboot the server;

    • Ensure load balancer detects changes (server is listed as offline);

    • Confirm that the SymmetricDS engine runs automatically at system start-up.

  12. Enable any external application jobs or batch processes.

2.6. Downgrade Considerations

There are two common ways to downgrade an existing SymmetricDS instance depending on the availability or the full database backup.

  1. When very recent full database backup (from before the upgrade) is available:

    • Shut down SymmetricDS engine;

    • Restore full database backup;

    • If old server is still available, start the old server with an old installation of SymmetricDS;

    • Or run the SymmetricDS installer for an older version - matching the old installation.

    • Start the SymmetricDS engine and monitor logs for errors.

  2. When database backup is not available or is out of date (missing new / unprocessed data):

    • Shut down SymmetricDS engine;

    • Manually execute command to modify database version marker to match older version:
      UPDATE sym_node SET schema_version='your_prior_symmetric_version' WHERE node_id='your_cluster_nodes';

    • Run the SymmetricDS installer for an older version - matching the old installation.

    • Start the SymmetricDS engine and monitor logs for errors.

3. Setup

Once the SymmetricDS software is installed on a computer and an instance of it is running, the next step in setting up the synchronization scenario is to set up a SymmetricDS node within that running SymmetricDS instance. As a reminder, a SymmetricDS node is connected to a database or file system and is responsible for synchronizing that database’s data to other SymmetricDS nodes within the node network.

3.1. Node Type

When opening the web console, if there are no nodes defined within the running instance, the Connect Database Wizard will be displayed to guide you through the process of creating one. There are two types of nodes:

  1. Setup New Replication - The primary node is typically the first node set up when creating a new replication scenario, and it serves as the central place where configuration is done. All configuration is stored in the database that this node is attached to.

  2. Join Existing Replication - All other nodes join existing replication by registering with the primary node, where they receive a copy of the configuration and learn about other nodes.

A third option will perform the Setup Demo installation, which does the following:

  1. Creates two example databases

  2. Creates a primary node attached to the server database

  3. Creates a secondary node attached to the client database

  4. Configures synchronization of the two databases

  5. Randomly executes insert, update, and delete statements on the database to show activity

wizard node setup
Figure 2. Node selection.

Select whether the node will setup a new replication, join an existing replication, or run a demonstration.

3.2. Primary Node Setup

3.2.1. Connect to the database

wizard connect database

The first step in setting up a primary node is to connect it to the database it will be responsible for synchronizing. This database will also be used to store the SymmetricDS configuration.

  • Specify the database type

  • Choose the replication mode (not all options available for every database)

    • Trigger-based - Triggers installed and maintained by SymmetricDS to capture changes.

    • Log-based - Database providers transaction logs are read to capture changes.

    • Time-based - A timestamp coumn or a single column numeric value is used to capture changes.

    • Load Only - A node that will only act as a target and will not capture changes.

    • Extract Only - Deprecated and should use time-based moving forward.

  • Specify the jdbc url. An example will be provided based on the database type selected. Be sure to replace host name, port and database name if necessary (varies by database type)

  • Specify the id to be used to login to the database

  • Specify the password to be used to login to the database

3.2.2. Table permissions

wizard connect permissions

This screen shows the permissions required to install the Symmetric System tables for the provided user and database.

3.2.3. Select a configuration

SymmetricDS has several predefined configuration options that can assist you with setting up a synchronization scenario. These options create a base configuration and allow you to tweak that configuration after initial creation. The following configuration options are provided:

Two Tier

This pattern is a central database being synchronized with many child databases of the same type. Typical examples include a central office database synchronizing to field databases (central office to stores, central office to field offices, etc.)

Source to Target

Similar to Two Tier, except the node groups are named source and target

Multi Master

Multi Master synchronization. Typically two master databases / nodes used in failover fashion

Secondary

A primary node group and a secondary node group synchronize by pushing data to each other.

On Prem to Cloud

An on prem node group that includes the registration server synchronizes data to and from a cloud node group.

Multi Source

Many source databases synchronizing to a target database. A typical example for this pattern is operational databases to a data warehouse or data mart.

The other options aren’t predefined configurations, but options for making your own configuration.

Custom configuration

This option allows you to configure everything from scratch

Import existing configuration

Allows you to import an existing SymmetricDS configuration file for this node

wizard connect config

Select the option and click Next.

3.2.4. Communication Settings

wizard connect host

Provide the URL that other SymmetricDS nodes will use to connect to this primary node. The default option provided should be used unless the primary node will be accessed through a proxy, load balancer, or different host name than shown in the default option.

You can also select HTTPS or HTTP to be used for transferring data.

3.2.5. Console User

wizard connect user

By default SymmetricDS uses an internal security scheme for protecting the SymmetricDS application. The admin user is the administrative user for the SymmetricDS web console. Provide a password for this user and click Next.

Additional users can be setup in the web console after the wizard completes. For more information, see Users.

SymmetricDS can also use LDAP or SAML to secure the SymmetricDS application. For more information, see Single Sign-On.

3.2.6. Summary

wizard connect ready

The node is now ready to be installed. Click the Finish button.

3.3. Add Table(s) Wizard

Once the SymmetricDS software is installed, an instance of it is running, and a master node is setup, the synchronization scenario must be configured (i.e. what tables and/or files do I want to synchronize from which source node groups to which target node groups). The Add Table(s) Wizard simplifies this process by walking you through several simple configuration steps.

The Add Table(s) Wizard can be found under the Configure→Overview screen.

configure overview
The add table(s) wizard will show a screen for each router that is in place. The below example shows the Two Tier configuration option with a client and server.
wizard config start

3.3.1. Router: Client to Server

wizard config client server

Select all tables that should be synchronized in the direction stated in the header, in this case from CLIENT to the SERVER

3.3.2. Router: Server to Client

wizard config server client

Next we will select tables to sync from the SERVER to the CLIENT.

The item_selling_price table was selected on both wizard screens so this table will be configured for bi-directional sync.

3.3.3. Configuration Summary

wizard config verify

The summary of the configuration made by the wizard.

Once the add table(s) wizard has completed your setup you can still navigate to the specific components created to provide customizations beyond the default settings.

3.4. Node Setup

Once configuration is complete SymmetricDS will automatically pop up a Connect Database Wizard if there is only one node currently setup in your network.

See Add Node for more information.

3.5. Node Properties File

The connect database wizard generates a properties file that contains all the information a node needs to bootstrap itself.

Each node that is deployed to a server is represented by a properties file that allows it to connect to a database and register with a parent node. Properties are configured in a file named xxxxx.properties. It is placed in the engines directory of the SymmetricDS install. The file is usually named according to the engine.name, but it is not a requirement.

To give a node its identity, the following properties are required. Any other properties found in conf/symmetric.properties can be overridden for a specific engine in an engine’s properties file. If the properties are changed in conf/symmetric.properties they will take effect across all engines deployed to the server.

You can use the variable $(hostName) to represent the host name of the machine when defining these properties (for example, external.id=$(hostName)). You can also access external id, engine name, node group id, sync URL, and registration URL in this manner. (for example, engine.name=$(nodeGroupId)-$(externalId)).
You can also use BSH script for the external id, engine name, node group id, sync URL, and registration URL. Use back ticks to indicate the BSH expression, and note that only one BSH expression is supporter for a given property line. The script can be prefixed or suffixed with fixed text. For example, if you wish to based the external id off of just a part of the hostname (e.g., substring of hostName): external.id=store-`import org.apache.commons.lang.StringUtils; return StringUtils.substring(hostName,2,4);\`
You can also set both environmental variables and system properties directly in the engine file. Environmental variables refer to external variables supplied by the Operating System, and system properties are variables that can be passed into the JVM. For example, if the environment variable you’re using is USERNAME, you would use $(USERNAME) in the engine file. Note, this is case sensitive.
engine.name

This is an arbitrary name that is used to access a specific engine using an HTTP URL. Each node configured in the engines directory must have a unique engine name. The engine name is also used for the domain name of registered JMX beans.

group.id

The node group that this node is a member of. Synchronization is specified between node groups, which means you only need to specify it once for multiple nodes in the same group.

external.id

The external id for this node has meaning to the user and provides integration into the system where it is deployed. For example, it might be a retail store number or a region number. The external id can be used in expressions for conditional and subset data synchronization. Behind the scenes, each node has a unique sequence number for tracking synchronization events. That makes it possible to assign the same external id to multiple nodes, if desired.

sync.url

The URL where this node can be contacted for synchronization. At startup and during each heartbeat, the node updates its entry in the database with this URL. The sync url is of the format: http://{hostname}:{port}/{webcontext}/sync/{engine.name}

The {webcontext} is blank for a standalone deployment. It will typically be the name of the war file for an application server deployment.

The {engine.name} can be left blank if there is only one engine deployed in a SymmetricDS server.

When a new node is first started, it is has no information about synchronizing. It contacts the registration server in order to join the network and receive its configuration. The configuration for all nodes is stored on the registration server, and the URL must be specified in the following property:

registration.url

The URL where this node can connect for registration to receive its configuration. The registration server is part of SymmetricDS and is enabled as part of the deployment. This is typically equal to the value of the sync.url of the registration server.

Note that a registration server node is defined as one whose registration.url is either blank or identical to its sync.url.

For a deployment where the database connection pool should be created using a JDBC driver, set the following properties:

db.driver

The class name of the JDBC driver.

db.url

The JDBC URL used to connect to the database.

db.user

The database username, which is used to login, create, and update SymmetricDS tables.

db.password

The password for the database user.

See Startup Parameters, for additional parameters that can be specified in the engine properties file.

3.6. Load Only Node

Load only nodes can be configured for nodes that are only designed to load data into the database. This prevents any triggers or runtime tables from being installed on this database but still allow data to be replicated to it.

Advantages of Load Only Nodes
  • No SymmetricDS runtime tables (SYM_*) installed inside the target database.

  • No SymmetricDS triggers installed inside the target database.

  • There is still support for bulk loaders if provided (MSSQL, Oracle, Postgres, MySQL for example).

  • Allows SymmetricDS to load data into dialects that are not fully supported yet or may not have full trigger support.

Constraints of Load Only Nodes
  • Can not be set up to capture changes

  • Requires a JDBC driver

  • May require table creation outside of SymmetricDS

  • Requires an additional database to be used for SymmetricDS runtime usage. H2 is used in the setup below but any other database could be used as the runtime database.

3.6.1. Setup Load Only Node

  • To setup a load only node, begin by adding a new node to your configuration (See Add Node for details on adding a new node).

Setup Load Only Node For Supported Databases
  • On the database setup screen of the connect database wizard, select load only after selecting your database type.

manage load only
  • The remaining steps in the connect database wizard are the same as if adding a full node so continue through the remainder of the connect database wizard to complete the configuration.

Setup Load Only Node For Unsupported Databases
  • Click the Add Driver…​ button to setup a new JDBC driver to support a load only node.

manage load only driver
  • Enter the information for the new JDBC driver

    Alias

    Provide a name for the database as it will appear in the future drop down list of available databases.

    Upload

    Upload button to upload a new JDBC driver jar file to the SymmetricDS /lib folder

    Driver Class

    This will be populated by default by a class in the uploaded JDBC jar that implements the java.sql.Driver interface.

    Example Url

    An exmaple jdbc connection url for this driver so that when it is selected in the future a template will provided.

  • Click Save and the database setup screen will now be populated with your new driver information.

  • The remaining steps in the connect database wizard are the same as if adding a full node so continue through the remainder of the connect database wizard to complete the configuration.

Uploading a jar file is only available in the standalone installation of SymmetricDS, NOT with the deployed war file setup. For a web container (war) based deployment you would need to add the driver jar to the web servers classpath and restart. Then provide the driver class and url manually as explained in steps 4 and 5 below.
If the driver is already a supported database by SymmetricDA and the driver jar file was packaged with SymmetricDS (in the /lib folder) or has already been uploaded for another node there is no need to upload anything here.

4. Configuration

Configuring SymmetricDS is the process of setting up your synchronization scenario.

The Configure Overview screen provides a summary of what is currently configured in this master node. The main table on this screen shows configuration elements (Groups, Group Links, Routers, etc.) and the number of items configured for each element. As an example, in the sample screenshot below, there are 2 Groups defined, 2 Group Links defined, 2 Routers defined, 9 Channels, defined, etc. Certain configuration elements are required to configure the synchronization scenario. If these elements are not configured, a warning symbol will be displayed next to the element name.

In addition to using the Add Table(s) Wizard to configure the synchronization solution, you can also configure SymmetricDS configuration elements one by one using the links on the left hand side of the Configure Overview screen. Each configuration element is described in the following sections of this guide.

Configuration can also be imported from and exported to external files by pressing the Import and Export buttons from the Configure Overview screen.

configure overview

4.1. Import

There are two ways a SymmetricDS configuration can be imported. Both ways support importing an export that is in CSV format. In order to import in SQL format you would run the SQL script via database specific tools or import through the web console in Configure → Overview.

  1. During the creation of the master node, while selecting a configuration

    import wizard profile
    Figure 3. Import Existing configuration during the node creation
  2. Import a configuration through the web console

    This will wipe out all existing configuration settings and replace with the newly imported ones.

4.2. Export

Export your configuration at any time through the web console. Use the drop down to select the format you would like to use for the export.

Exporting your configuration might be useful in order to save a configuration in source control or to be used as an import into another environment. Export and Import are frequently used to migrate configuration between test and production environments.

4.3. Groups

In SymmetricDS, configuration rules are applied to groups of nodes versus individual nodes. A group is a categorization of nodes with similar synchronization needs. For example, in a synchronization scenario where a corporate office database is synchronized with field office databases, two node groups would be created, one for the corporate office database (Corporate), and one for the field office databases (Field_office). In the corporate group, there would be a single node and database. In the field_office group, there would be many nodes and databases, one for each field office. Configuration rules/elements are applied to the node group versus the individual nodes in order to simplify the configuration setup (no need to configure each individual field office node, just how the field office nodes sync with the corporate office node).

group
Required Fields
Group ID

Unique identifier for the group.

Advanced Options
Description

Description of the group that is available through the console.

Example 1. Sample Node Groups

Basic Two Tier

Server, Client

Retail

Corp, Store, Register,Handheld

Field Office

Corporate, Field_office

Failover

Master, Failover

Group links define at a high level how data moves throughout your synchronization scenario. The group link defines which node groups will synchronize data to other node groups and within that exchange, which node group will initiate the conversation for that exchange.

group link
Required Fields
Source Group ID

The source group of the communication link.

Link

Defines how the source and target groups will communicate.

Table 1. Options for Group Links

Push [P]

Indicates that nodes in the source node group will initiate communication over an HTTP PUT and push data to nodes in the target node group.

Wait for Pull [W]

Indicates nodes in the source node group will wait for a node in the target node group to connect via an HTTP GET and allow the nodes in the target node group to pull data from the nodes in the source node group.

Route-only [R]

Route-only indicates that the data isn’t exchanged between nodes in the source and nodes in the target node groups via SymmetricDS. This action type might be useful when using an XML publishing router or an audit table changes router.

Target Group ID

The target group of the communication link.

Advanced Options
Sync Configuration

Determines if configuration is also sent through this group link during routing. This affects all SymmetricDS configuration tables except for extract_request, monitor_event, node, node_host, node_security, table_reload_request, and table_reload_status. By default this is checked and configuration will communicate on this path. There are configurations that might cause configuration to continuously loop through the network. As a result, this might need to be unchecked for some links.

Sync SQL/DDL

Determines if SQL events (including DDL) is also sent through this group link during routing. By default this is checked and SQL/DDL will communicate on this path. When using DDL replication, this gives the user control of where to route DDL changes.

Reversible

Allows the communication link to send in the reverse direction if specified on the channel. A push link can be overridden to pull and a pull link can be overridden to push using a setting on the channel.

Example 2. Sample Group Links
Basic

Common two tier corp and store store bi-directional setup

group link basic
Firewall

All communication is initiated from corp group. This is a common way to setup synchronization when the corp group is behind a firewall.

group link firewall
Peer to Peer

Allows a node to sync to all other nodes in the same group.

group link same
Loop

Turn sync configuration off on group link c to a so that configuration does not loop.

group link config
Group Links Tips and Tricks
The link also defines if configuration data will be synchronized on the link. For example, you might not want remote nodes to be able to change configuration and affect other nodes in the network. In this case you would uncheck sync configuration.
Turn off the sync configuration option where appropriate in non hierarchical setup so that configuration changes do not create a loop. See "Loop" example above.

4.5. Routers

Routers ride on top of group links. While a group link specifies that data should be moved from nodes in a source node group to nodes in a target node group, routers define more specifically which captured data from a source node should be sent to which specific nodes in a target node group, all within the context of the node group link.

router
Required Fields
Router Id

Unique description of a specific router

Group Link

The group link used for the source and target node groups of this router

Router Type

The type of router. Standard router types are listed below. Custom routers can be configured as extension points. If a router does not support non-DML data events, then it will send all non-DML data events to the default router instead.

Table 2. Router Types
Type Description Supports Non-DML

default

A router that sends all captured data to all nodes that belong to the target node group defined in the router. See Default Router

column

A router that compares old or new column values in a captured data row to a constant value or the value of a target node’s external id or node id. See Column Match Router

audit

A router that inserts into an automatically created audit table. It records captured changes to tables that it is linked to. See Audit Table Router

java

A router that executes a Java expression in order to select nodes to route to. The script can use the old and new column values. See [Java Router]

lookuptable

A router which can be configured to determine routing based on an existing or ancillary table specifically for the purpose of routing data. See Lookup Table Router

subselect

A router that executes a SQL expression against the database to select nodes to route to. This SQL expression can be passed values of old and new column values. See Subselect Router

convertToReload

When subselect router is too slow and there are too many tables or rows for lookuptable router, a convertToReload router can efficiently sub-set data by converting multiple changes into a reload batch. See ConvertToReload Router

bsh

A router that executes a Bean Shell script expression in order to select nodes to route to. The script can use the old and new column values. See Beanshell Router

csv

A router that sends data from a CSV file to a target table of the target node group defined in the router. See CSV Router

dbf

A router that sends data from a dBase generated DBF file to a target table of the target node group defined in the router. See DBF Router

Router Expression

An expression that is specific to the type of router that is configured in router type. See the documentation for each router for more details.

Use Source Catalog/Schema

If set then the source catalog and source schema are sent to the target to be used to find the target table.

Target Catalog

Optional name of catalog where a target table is located. If this field is unspecified, the catalog will be either the default catalog at the target node or the "source catalog name" from the table trigger, depending on how "use source catalog schema" is set for the router. Variables are substituted for $(sourceNodeId), $(sourceExternalId), $(sourceNodeGroupId), $(targetNodeId), $(targetExternalId), $(targetNodeGroupId), $(sourceCatalogName), and $(sourceSchemaName). Parameter values can be substituted using $(name) syntax. See Variables.

Target Schema

Optional name of schema where a target table is located. If this field is unspecified, the schema will be either the default schema at the target node or the "source schema name" from the table trigger, depending on how "use source catalog schema" is set for the router. Variables are substituted for $(sourceNodeId), $(sourceExternalId), $(sourceNodeGroupId), $(targetNodeId), $(targetExternalId), $(targetNodeGroupId), $(sourceCatalogName), and $(sourceSchemaName). Parameter values can be substituted using $(name) syntax. See Variables.

Advanced Options
Sync on Update

Flag that indicates that this router should send updated rows from nodes in the source node group to nodes in the target node group.

Sync on Insert

Flag that indicates that this router should send inserted rows from nodes in the source node group to nodes in the target node group.

Sync on Delete

Flag that indicates that this router should send deleted rows from nodes in the source node group to nodes in the target node group.

Target Table

Optional name for a target table. Only use this if the target table name is different than the source.

Example 3. Sample Routers
routers examples
  • link break A router that is not linked to any triggers.

  • table A router that is linked to a single trigger.

  • table multiple A router that is linked to multiple triggers.

4.5.1. Router Types

Default Router

The simplest router is a router that sends all the data that is captured by its associated triggers to all the nodes that belong to the target node group defined in the router.

Column Match Router

Sometimes requirements may exist that require data to be routed based on the current value or the old value of a column in the table that is being routed.

The first part of the expression is always the column name. The column name should always be defined in upper case. The upper case column name prefixed by OLD_ can be used for a comparison being done with the old column data value.

The second part of the expression can be a constant value, a token that represents another column, or a token that represents some other SymmetricDS concept. Token values always begin with a colon (:).

  1. Consider a table that needs to be routed to all nodes in the target group only when a status column is set to 'READY TO SEND.'

Router Expression
STATUS=READY TO SEND
  1. Consider a table that needs to be routed to all nodes in the target group only when a status column changes values.

The use of OLD_STATUS, where the OLD_ prefix gives access to the old column value.
Router Expression
STATUS!=:OLD_STATUS
Attributes on a NODE that can be referenced with the following tokens
  • :NODE_ID

  • :SOURCE_NODE_ID

  • :EXTERNAL_ID

  • :SOURCE_EXTERNAL_ID

  • :NODE_GROUP_ID

  • :SOURCE_NODE_GROUP_ID

  • :REDIRECT_NODE

  1. Consider a table that needs to be routed to only nodes in the target group whose STORE_ID column matches the external id of a node.

Router Expression
STORE_ID=:EXTERNAL_ID
  1. Consider a table that needs to be routed to a redirect node defined by its external id in the REGISTRATION_REDIRECT table.

Router Expression
STORE_ID=:REDIRECT_NODE
  1. More than one column may be configured in a router_expression. When more than one column is configured, all matches are added to the list of nodes to route to. The following is an example where the STORE_ID column may contain the STORE_ID to route to or the constant of ALL which indicates that all nodes should receive the update.

Router Expression
STORE_ID=ALL or STORE_ID=:EXTERNAL_ID
  1. The NULL keyword may be used to check if a column is null. If the column is null, then data will be routed to all nodes who qualify for the update. This following is an example where the STORE_ID column is used to route to a set of nodes who have a STORE_ID equal to their EXTERNAL_ID, or to all nodes if the STORE_ID is null.

Router Expression
STORE_ID=NULL or STORE_ID=:EXTERNAL_ID
  1. External data collected as part of the trigger firing (see External Select) can also be used as a virtual column in the router expression as well.

Router Expression
EXTERNAL_DATA=:EXTERNAL_ID
Audit Table Router

This router audits captured data by recording the change in an audit table that the router creates and keeps up to date. The router creates a table named the same as the table for which data was captured with the suffix of _AUDIT. It will contain all of the same columns as the original table with the same data types only each column is nullable with no default values.

The following parameter must be set to true so that the audit table can be created.
auto.config.database=true
Three extra "AUDIT" columns are added to the table:
AUDIT_ID

the primary key of the table.

AUDIT_TIME

the time at which the change occurred.

AUDIT_EVENT

the DML type that happened to the row.

The audit router must be associated with a node group link of type 'R'. The 'R' stands for 'only routes to' (see Group Links).
Lookup Table Router

A lookup table may contain the id of the node where data needs to be routed. This could be an existing table or an ancillary table that is added specifically for the purpose of routing data.

Each of the following configuration parameters are required.
LOOKUP_TABLE

This is the name of the lookup table.

KEY_COLUMN

This is the name of the column on the table that is being routed. It will be used as a key into the lookup table.

LOOKUP_KEY_COLUMN

This is the name of the column that is the key on the lookup table.

EXTERNAL_ID_COLUMN

This is the name of the column that contains the external_id of the node to route to on the lookup table.

ALL_NODES_VALUE

This is an optional parameter that allows you to specify a value for the EXTERNAL_ID_COLUMN that means "send to all nodes". The value of "null" will be interpreted as a null value, not the string "null".

The lookup table will be read into memory and cached for the duration of a routing pass for a single channel.

Consider a table that needs to be routed to a specific store, but the data in the changing table only contains brand information. In this case, the STORE table may be used as a lookup table.

Router Expression

LOOKUP_TABLE=STORE KEY_COLUMN=BRAND_ID LOOKUP_KEY_COLUMN=BRAND_ID EXTERNAL_ID_COLUMN=STORE_ID

Subselect Router

Sometimes routing decisions need to be made based on data that is not in the current row being synchronized. A 'subselect' router can be used in these cases. A 'subselect' is configured with a router expression that is a SQL select statement which returns a result set of the node ids that need routed to. Column tokens can be used in the SQL expression and will be replaced with row column data.

The overhead of using this router type is high because the 'subselect' statement runs for each row that is routed. It should not be used for tables that have a lot of rows that are updated. It also has the disadvantage that if the data being relied on to determine the node id has been deleted before routing takes place, then no results would be returned and routing would not happen.

The router expression you specify is appended to the following SQL statement in order to select the node ids:

select c.node_id
from sym_node c
where c.node_group_id=:NODE_GROUP_ID
        and c.sync_enabled=1 and ...

The SQL statement has access to the following variables that are replaced before running:

Table 3. Variables available to the subselect router

:NODE_GROUP_ID

The target node group ID that is configured for the router.

:EXTERNAL_DATA

The external data for current row, as configured by sym_trigger.external_select.

:DATA_EVENT_TYPE

The event type of either INSERT, UPDATE, or DELETE.

:TABLE_NAME

The table name for the current row.

:COLUMN_NAME

Variables named for each column name (in uppercase), which return the column value for the new row.

:OLD_COLUMN_NAME

Variables named for each column name (in uppercase and prefixed with OLD_), which return the column value for the old row.

Example 4. Sample Use Case for Subselect Router

For an example, consider the case where an Order table and an OrderLineItem table need to be routed to a specific store. The Order table has a column named order_id and STORE_ID. A store node has an external_id that is equal to the STORE_ID on the Order table. OrderLineItem, however, only has a foreign key to its Order of order_id. To route OrderLineItems to the same nodes that the Order will be routed to, we need to reference the master Order record.

There are two possible ways to solve this in SymmetricDS.

  1. Configure a 'subselect' router type (shown below).

  2. Use an external select to capture the data via a trigger for use in a column match router, see External Select.

Router Expression
c.external_id in (select STORE_ID from order where order_id=:ORDER_ID)
In this example that the parent row in Order must still exist at the moment of routing for the child rows (OrderLineItem) to route, since the select statement is run when routing is occurring, not when the change data is first captured.
ConvertToReload Router

This router converts multiple change events into a single reload event, which can be used to sub-set data quickly when the subselect router is too slow or when the lookuptable router can’t handle the amount of lookup data. ConvertToReload inserts the primary key values for each row, along with a unique load ID, into a temporary table. The reload event uses the initial load SQL along with a join to the temporary table to retrieve the changes and sub-set data.

The router expression requires the name of the temporary table to use:

temptable=mytable_router
Example 5. Sample Use Case for ConvertToReload Router
create table mytest (id integer, name varchar(50), primary key(id));
create table mytest_router (id integer, load_id integer, primary key(load_id, id));
insert into SYM_ROUTER (router_id,
        source_node_group_id, target_node_group_id, router_type,
        router_expression, create_time, last_update_time) values
        ('corp-2-store','corp', 'store', 'convertToReload', 'temptable=mytest_router', current_timestamp,
        current_timestamp);
Beanshell Router

When more flexibility is needed in the logic to choose the nodes to route to, then the a scripted router may be used. The currently available scripting language is Bean Shell. Bean Shell is a Java-like scripting language. Documentation for the Bean Shell scripting language can be found at http://www.beanshell.org .

The router type for a Bean Shell scripted router is 'bsh'. The router expression is a valid Bean Shell script that:

Table 4. Variables available to the script

nodes

Collection of org.jumpmind.symmetric.model.Node objects the router would route to normally.

nodeIds

Collection of node ids that the router would route to normally. You can just return this if you want the bsh router to behave like the default router.

targetNodes

Collection of org.jumpmind.symmetric.model.Node objects to be populated and returned.

engine

The instance of org.jumpmind.symmetric.ISymmetricEngine which has access to SymmetricDS services.

Any Data Column

Data column values are bound to the script evaluation as Java object representations of the column data. The columns are bound using the uppercase names of the columns. For example, a table with a store_id column will have a STORE_ID variable name available in Bean Shell script. When using file sync, columns from the FILE_SNAPSHOT table will be available in the script.

Any Old Values

Old Data column values are bound to the script evaluation as Java object representations of the column data. The columns are bound using the uppercase representations that are prefixed with 'OLD_'. For example, a table with a store_id column will have an OLD_STORE_ID variable name available in Bean Shell script representing the old value for the store_id before the change. When using file sync, columns from the FILE_SNAPSHOT table will be available in the script.

Table 5. Return options

targetNodes

Collection of org.jumpmind.symmetric.model.Node objects that will be routed to.

true

All nodes should be routed

false

No nodes should be routed

The last line of a bsh script is always the return value.
Example 6. Use case using a Bean Shell where the node_id is a combination of STORE_ID and WORKSTATION_NUMBER, both of which are columns on the table that is being routed.
targetNodes.add(STORE_ID + "-" + WORKSTATION_NUMBER)

The same could also be accomplished by simply returning the node id.

STORE_ID + "-" + WORKSTATION_NUMBER
Example 7. Use case using a Bean Shell script to synchronize to all nodes if the FLAG column has changed, otherwise no nodes will be synchronized.
FLAG != null && !FLAG.equals(OLD_FLAG)
Here we make use of OLD_, which provides access to the old column value.
Example 8. Use case using a Bean Shell script that iterates over each eligible node and checks to see if the trimmed value of the column named STATION equals the external_id.
for(org.jumpmind.symmetric.model.Node node : nodes) {
        if (STATION != null && node.getExternalId().equals(STATION.trim())) {
                targetNodes.add(node.getNodeId());
        }
}
CSV Router

This router is used to route csv files to a specified target database table.

Consider a csv file that needs to be routed to a target table. To specify the target table, click on "Show Advanced Options" and enter your table name.

csv router

In your router expression you can tell the router to include a transaction ID for the routed data with INCLUDE_TRANSACTION_ID=true|false. Default is false.

The CSV Router routes from file to database so a file trigger must be created to specify the directory or path that the file trigger should watch. (see File Triggers)
DBF Router

This router is used to route dbf files that are generated from a dBase system to a specified target database table.

The DBF Router routes from file to database so a file trigger must be created to specify the directory or path that the file trigger should watch. (see File Triggers)

4.6. Channels

Once group links and routers are defined, configuration must be completed to specify which data (tables, file systems, etc.) should be synchronized over those links and routers. The next step in defining which specific data in the database is moved is to define logical groupings for that data. Channels define those logical groupings. As an example, a set of tables that hold customer data might be logically grouped together in a Customer channel. Sales, returns, tenders, etc. (transaction data) might be logically grouped into a transaction channel. A default channel is automatically created that all tables will fall into unless other channels are created and specified. The default channel is called 'default'.

Channels can be disabled, suspended, or scheduled as needed.

channel
Transactions will NOT be preserved across channels so its important to setup channels to contain all tables that participate in a given transaction.
Required Fields
Channel ID

Identifier used through the system to identify a given channel.

Processing Order

Numeric value to determine the order in which a channel will be processed. Channels will be processed in ascending order.

Batch Algorithm

Batching is the grouping of data, by channel, to be transferred and committed at the client together.

Default

All changes that happen in a transaction are guaranteed to be batched together. Multiple transactions will be batched and committed together until there is no more data to be sent or the max_batch_size is reached. The routing.max.batch.size.exceed.percent parameter is used to keep batch sizes within a percentage over the max_batch_size, or it can be set to zero for no limit.

Transactional

Batches will map directly to database transactions. If there are many small database transactions, then there will be many batches. The max_batch_size column has no effect.

Nontransactional

Multiple transactions will be batched and committed together until there is no more data to be sent or the max_batch_size is reached. The batch will be cut off at the max_batch_size regardless of whether it is in the middle of a transaction.

Max Batch Size

Specifies the maximum number of data events to process within a batch for this channel.

Max Batch To Send

Specifies the maximum number of batches to send for a given channel during a 'synchronization' between two nodes. A 'synchronization' is equivalent to a push or a pull. For example, if there are 12 batches ready to be sent for a channel and max_batch_to_send is equal to 10, then only the first 10 batches will be sent even though 12 batches are ready.

Max Data To Route

Specifies the maximum number of data rows to route for a channel at a time.

Max KB/s

Specifies the maximum network transfer rate in kilobytes per second. Use zero to indicate unlimited. When throttling the channel, make sure the channel is on its own queue or within a queue of channels that are throttled at the same rate. This is currently only implemented when staging is enabled.

Data Loader Types

Determines how data will be loaded into the target tables. These are used during an initial load or a reverse initial load. Data loaders do not always have to load into the target relational database. They can write to a file, a web service, or any other type of non-relational data source. Data loaders can also use other techniques to increase performance of data loads into the target relation database.

default

Performs an insert first and if this fails will fall back to an update to load the data.

ftp_localhost

Sends the data in CSV format to a configured ftp location. These locations are setup in the TODO {SYM_HOME}/conf/ftp-extensions.xml

bulk

Assigns the appropriate bulk loader to this channel. Supported bulk loaders include: Microsoft SQL, PostgreSQL, MySQL and Amazon Redshift over S3.

mongodb

MongoDB data loader.

Tables that should be data loaded should be configured to use this channel. Many times, a reload channel will be set to bulk load to increase the performance of an initial load.
Queue Name

Determines a queue that the channel will sync in. Channels with the same queue name are processed synchronously (one at a time) and channels on different queues are processed asynchronously (in parallel).

Advanced Options
Group Link Direction

For a node group link that is reversible, the channel can specify either "push" or "pull" to override the default group link communication. If this field is empty, the default group link communication is used.

Enabled

Indicates whether the channel is enabled or disabled. If a channel is disabled, data is still captured for changes that occur on the source system, but it will not be routed and sent to the target until the channel is re-enabled.

Reload Channel

Indicates whether a channel is available for initial loads and reverse initial loads.

File Sync Channel

Indicates whether a channel is available for file synchronization.

Use Old Data To Route

Indicates if the old data will be included for routing. Routing can then use this data for processing. Defaults to true.

Use Row Data To Route

Indicates if the current data will be included for routing. Routing can then use this data for processing. Defaults to true.

Use Primary Key (PK) Data to Route

Indicates if the primary key data will be include for routing. For example maybe a store ID is needed to apply logic on before sending to the appropriate target nodes. Defaults to true.

Contains Lob or Wide Row Data

For Oracle, Tibero, Firebird, and Interbase, this setting can be enabled when change data capture exceeds the character limit. Oracle and Tibero have a character limit of 4000, while Firebird and Interbase have a character limit of 20000 for changes and 1000 for primary key values. Change data capture is first attempted to extract as character data for better performance, then it will automatically fall back to extract as a large object (LOB). Enable this setting when most changes captured on the channel need extracted as LOB or when the extraction is receiving a truncation error.

Example 9. Sample Channels
TODO
Channel Tips and Tricks
Increase performance by creating designated channels for tables that use LOB data types. For these channels be sure to check the "Table Contains Big Lobs" to increase performance.

4.7. Table Triggers

The next step in defining the synchronization scenario is to define which tables should be synchronized. The simplest way to do this is by using the Auto Create…​ button on the Configure → Table Triggers screen. The Auto Create functionality introspects the master node’s database and looks for available tables to synchronize.

Note that you will be configuring tables to synchronize for specific Routers. You will get a chance to select tables for each configured Router. Select the channel you desire from the dropdown, select the tables that should be synchronized using that channel, and then click the Next button. A summary will be presented at the end of the configuration. Click Finish to create the triggers.

The Auto Create functionality can be used repeatedly to select different table, router and channel combinations. If you are configuring a new set of tables triggers that should use an independent Trigger definition be sure to enter values for either the Trigger Prefix or the Trigger Suffix fields. These will be used to create a unique Trigger Id.

triggers auto create

For each table selected in the Auto Create wizard, a corresponding Table Trigger configuration element will be added that can be maintained within the Table Triggers list screen. New Table Trigger configuration elements can also be created manually from this screen.

trigger
Required Fields
Trigger Id

Unique identifier for a trigger.

Source Catalog

Optional name for the catalog the configured table is in. If the name includes * then a wildcard match on the table name will be attempted. \ Wildcard names can include a list of names that are comma separated. The ! symbol may be used to indicate a NOT match condition. Parameter values can be substituted using $(name) syntax. See Variables.

Source Schema

Optional name for the schema a configured table is in. If the name includes * then a wildcard match on the table name will be attempted. Wildcard names can include a list of names that are comma separated. The ! symbol may be used to indicate a NOT match condition. Parameter values can be substituted using $(name) syntax. See Variables.

Source Table

The name of the source table that will have a trigger installed to watch for data changes. See Trigger Wildcards for using wildcards to specify multiple source tables. Parameter values can be substituted using $(name) syntax. See Variables.

Channel

The channel_id of the channel that data changes will flow through.

Advanced Options
Sync On Insert

Flag for installing an insert trigger.

Sync On Update

Flag for installing an update trigger.

Sync On Delete

Flag for installing a delete trigger.

Reload Channel Id

The channel_id of the channel that will be used for initial loads.

Sync Conditions

A procedure language expression included in the trigger text to determine whether a change is captured or not. Most platforms include the condition inside an "IF" statement, while SQL-Server includes the condition in a "WHERE" clause. Old and new values of a column can be referenced using "$(oldTriggerValue)" and "$(newTriggerValue)" aliases respectively. See Trigger Variables. For example, if a character column is named "STATUS" and the row should be captured when the value is "2", then the condition would be:

 $(newTriggerValue).status = '2'
Sync On Insert Condition

Conditional expression for the insert trigger to determine if a change is captured or not. See Sync Conditions.

Sync On Update Condition

Conditional expression for the update trigger to determine if a change is captured or not. See Sync Conditions.

Sync On Delete Condition

Conditional expression for the delete trigger to determine if a change is captured or not. See Sync Conditions.

Custom Insert Trigger Text

Specify insert trigger text (SQL) to execute after the SymmetricDS trigger fires. This field is not applicable for H2, HSQLDB 1.x or Apache Derby.

Custom Update Trigger Text

Specify update trigger text (SQL) to execute after the SymmetricDS trigger fires. This field is not applicable for H2, HSQLDB 1.x or Apache Derby.

Custom Delete Trigger Text

Specify delete trigger text (SQL) to execute after the SymmetricDS trigger fires. This field is not applicable for H2, HSQLDB 1.x or Apache Derby.

Sync On Incoming

Whether or not an incoming batch that loads data into this table should cause the triggers to capture changes. Changes are never sent back to the source node, but enabling this setting makes it possible to create a never-ending loop through intermediary nodes, such as A to B to C to A.

Capture Row As LOB

For Oracle, Tibero, SQL Server, and DB2, row data that exceeds the character limit needs converted into a large object (LOB) for capture. Oracle and Tibero have a character limit of 4000, SQL Server has a nvarchar limit of 4000 and a varchar limit of 8000, while DB2 has a character limit of 32767. The trigger will try to automatically capture the row correctly, so only enable this setting if you are getting truncation errors during trigger creation or when changing data.

Stream LOBs

Captures an empty placeholder for large object (LOB) data types when a row is changed, then queries for the LOB value later when the batch is extracted. If normal capturing of LOBs is not working, enabling this setting may work instead. When very large LOB data is involved, this setting can reduce the overhead of making changes in the database, but it usually results in worse performance of synchronization since it queries each row during extraction.

Stream Row

Captures only the primary key values when the trigger fires, which can reduce overhead for tables with wide data or many columns. The data will be queried using the PK values when the batch is extracted. This results in worse performance of synchronization, but it can be used when triggers for all columns won’t install or when contention from triggers is too high.

Capture Old Data

Indicates whether this trigger should capture and send the old data, which is the previous state of the row before the change. Enable this option if you need to access old data in custom trigger text, routing expression, or transform expression. Otherwise, disable this option for better performance.

Handle Key Updates

For SQL-Server and Sybase, enable this setting to capture changes to the primary key. The trigger needs to do some additional work to handle changes to the primary key, so this setting is normally disabled.

External Select

Specify a SQL select statement that returns a single row, single column result. It will be used in the generated database trigger to populate the EXTERNAL_DATA field on the data table.

Excluded Column Names

Specify a comma-delimited list of columns that should not be synchronized from this table.

Included Column Names

Specify a comma-delimited list of columns only should be synchronized from this table.

Sync Key Names

Specify a comma-delimited list of columns that should be used as the key for synchronization operations. By default, if not specified, then the primary key of the table will be used.

Channel Expression

An expression that will be used to capture the channel id in the trigger. This expression will only be used if the channel_id is set to 'dynamic'. The variable "$(schemaName)" can be used, which is replaced with the source schema of the table. See Variables.

Example 10. Sample Triggers
triggers examples
Multiple Triggers On A Table
Note that many databases allow for multiple triggers of the same type to be defined. Each database defines the order in which the triggers fire differently. If you have additional triggers beyond those SymmetricDS installs on your table, please consult your database documentation to determine if there will be issues with the ordering of the triggers.
Add Table(s) Wizard
Use the Add Table(s) Wizard to setup a complete synchronization scenario with default trigger settings.
Auto Create Triggers
By using the "Auto Create" button the on the "Table Triggers" screen you can create all your triggers through a single dialog.
Capture Changed Data

When determining whether a data change has occurred or not, by default the triggers will record a change even if the data was updated to the same value(s) they were originally. For example, a data change will be captured if an update of one column in a row updated the value to the same value it already was. There is a global property that allows you to override this behavior (defaults to false).

trigger.update.capture.changed.data.only.enabled=false

This property is currently only supported on MySQL, DB2, SQL Server, and Oracle.

4.7.1. Trigger Wildcards

The source table name may be an expression that is used to match multiple table names. Special characters include the asterisk ('*') for wildcards, the comma (',') for multiple expressions, and the exclamation ('!') for negation.

Wildcard Rules
  • Separate multiple table names or table name expressions with a comma.

  • Characters are always evaluated from left to right.

  • When a table match is made, the table is either added to or removed from the list of tables. If another trigger already exists for a table, then that table is not included in the wildcard match (the explicitly defined trigger entry take precedence).

  • System tables and any table names that start with the SymmetricDS table prefix will be excluded.

  • To negate the expression and exclude tables, start the expression with an exclamation.

  • Double up special characters to match a single literal character. (Use two asterisks to match a single asterisk.) The entire expression is processed as wildcarded when an odd number of consecutive special characters are found.

trigger wildcard
Figure 4. Sample wildcard trigger for all tables that start with "sale" or " item"

4.7.2. External Select

Occasionally, you may find that you need to capture and save away a piece of data present in another table when a trigger is firing. This data is typically needed for the purposes of determining where to 'route' the data to once routing takes place. Each trigger definition contains an optional "external select" field which can be used to specify the data to be captured. Once captured, this data is available during routing in DATA 's external_data field.

For these cases, place a SQL select statement which returns the data item you need for routing in external_select. See Trigger Variables for a list of variables available for use.

The external select SQL must return a single row, single column
Example 11. Sample External Select SQL that returns STORE_ID based on the ORDER_ID captured in the trigger.
select STORE_ID
from order
where order_id=$(curTriggerValue).$(curColumnPrefix)order_id
External select SQL statements should be used carefully as they will cause the trigger to run the additional SQL each time the trigger fires.
Using an external select on the trigger is similar to using the 'subselect' router. The advantage of this approach over the 'subselect' approach is that it guards against the (somewhat unlikely) possibility that the master Order table row might have been deleted before routing has taken place. This external select solution also is a bit more efficient than the 'subselect' approach.

4.7.3. Trigger Variables

The Sync Condition, External Select, and Custom Trigger Text configurations allow the user to provide procedure language text that is included inside the trigger. Variables can be used for configuration that works across different database platforms. When triggers are created, the variables are replaced with the syntax needed for that specific database.

Table 6. Trigger Template Variables

$(newTriggerValue)

New row alias for inserts and updates. For example, "$(newTriggerValue).MYCOLUMN" becomes ":new.MYCOLUMN" for an insert/update trigger on Oracle.

$(oldTriggerValue)

Old row alias for updates and deletes. For example, "$(oldTriggerValue).MYCOLUMN" becomes ":old.MYCOLUMN" for an update/delete trigger on Oracle.

$(curTriggerValue)

Current row alias for insert, updates, and deletes. This variable acts like $(newTriggerValue) for inserts and updates, and it acts like $(oldTriggerValue) for deletes.

$(curColumnPrefix)

Column prefix only used by H2 database. It is replaced with the NEW_ or OLD_ column prefix needed by H2. All other platforms will replace the variable with an empty string

4.7.4. Load Only Triggers

Occasionally the decision of what data to load initially results in additional triggers. These triggers, known as load only triggers, are configured such that they do not capture any data changes. In other words, the sync on insert, sync on update, and sync on delete attributes of the trigger are all set to false.

Example 12. Sample load only trigger
trigger load only
Load only triggers still participate in the following:
  • Initial Loads

  • Reverse Initial Loads

  • Table Reloads

  • Creation of tables during initial loads

Use cases for load only triggers:
  • To load a read-only lookup table, for example. It could also be used to load a table that needs populated with example or default data.

  • Recovery of data for tables that have a single direction of synchronization. For example, a retail store records sales transactions that synchronize in one direction by trickling back to the central office. If the retail store needs to recover all the sales transactions from the central office, they can be sent are part of an initial load from the central office by setting up a load only trigger that "sync" in that direction.

4.7.5. Expandable Triggers

When a set of tables is repeated for each node and the tables all contain a prefix which groups them together an expandable trigger can be used. The prefix will be resolved to the external id that the source node replicates with either through a push or pull.

Table 7. Variable

$(targetExternalId)

Resolved by looking up all the nodes the source node replicates with and will be replaced with each external id to represent a table.

Example
$(targetExternalId)-Item

If there are nodes present with external ids of client1, client2, client3 then this trigger configuration would match the following 3 tables.

client1-Item
client2-Item
client3-Item

This reduces configuration if there are for example 10 tables that are repeated for each client then only table trigger configurations would be needed and as new nodes are added (along with their tables) there is no need to add any more configuration.

4.8. Table Routing

As discussed previously, Routers define more specifically which captured data from a source node should be sent to which specific nodes in a target node group. Table Triggers define which tables should be synchronized. Table Routing defines how the Routers and Table Triggers work together, specifying what tables are associated with each router (and thus the direction the table data flows).

The simplest way to set up Table Routing is to use the Auto Create…​ button.

auto create 1 triggers

The default entry point for the auto creation is from the triggers perspective. This will allow you to select table triggers first then routers.

Clicking Next to select from available routers.

auto create 2 routers

Select the routers (direction) to associate with the previously selected triggers.

Click Save to install the triggers on the database and begin data capture.

You can also select routers on the first screen of the auto create if you would like to select routers first then triggers. Both approaches will result in the same table routing.

auto create 1 routers

Select from available routers.

Clicking Next to select from available table triggers.

auto create 2 triggers

Select the table triggers to associate with the previously selected routers.

Click Save to install the triggers on the database and begin data capture.

The table routing elements created by the wizard can be maintained from the Table Routing List Screen. New table routing entries can also be created manually from this screen if desired.

SymmetricDS will not install table triggers on the database until they are associated with a router through table routing.
Once a table routing relationship is created data capture will begin.
table routing new
Required Fields
Table Trigger

The table trigger determines the source of data.

Router

The router determines where data will be sent.

Initial Load Select

A SQL expression used in the WHERE clause of the SELECT that extracts the table during initial load. Leave blank to retrieve all rows. For correlated joins, reference the table using the "t" alias. Variables are substituted for $(groupId), $(nodeId), and $(externalId). See Variables section for formatting and manipulation of variables.

Advanced Options
Initial Load Delete

A SQL statement to run that clears the table on the target node at the beginning of the initial load before data is sent. This is usually a delete or truncate statement. Variables are substituted for $(groupId), $(nodeId), $(externalId), $(sourceGroupId), $(sourceNodeId), and $(sourceExternalId). See Variables section for formatting and manipulation of variables.

Initial Load Order

Numeric position for this table in the initial load, sent in ascending numeric order. When two numeric values are the same, the ordering is based on foreign key constraints. Use a negative number to exclude the table from initial load.

Enabled

When enabled, changes are captured and routed into batches for the table. When disabled, changes are no longer captured for the table and any outstanding changes are placed into batches with a status of "unrouted".

Ping Back Enabled

When enabled, synchronized rows with the Sync On Incoming option set on a trigger will be routed back to the source node. By default, the data is not routed back to the source node (to prevent endless loops of data changes). A reason to turn this option on is when a transform changes some of the data and the transformed data needs to be returned to the source node.

Data Refresh Type

When the data refresh jobs runs (overnight by default), this control how to refresh data in the table. Blank means exclude the table from the data refresh. Auto chooses between full reload or compare and repair. Full uses a table reload request with delete first option. Differences uses a compare and repair request.

4.9. Table Groups

Table groups are used to define collections of tables that will synchronize their data at the same time. When one row from a table in a table group is changed on the source, all corresponding rows from other tables in the same group will also be synchronized to the target.

table group
Required Fields
ID

Unique identifier for the table group.

Source Node Group ID

The node group where data changes are captured.

Target Node Group ID

The node group where data changes will be sent.

Writer Type

Type of writer: D=default, J=JSON.

Type Description

Default

The source data will remain relational on the target.

JSON

The source data will be converted to JSON documents on the target. The target database must be MongoDB, Azure Cosmos DB, or Elasticsearch.

Advanced Options
Description

Description of the table group that is available through the console.

4.10. Table Group Hierarchies

Table group hierarchies define the tables that belong to a table group and their relationships to each other, with each hierarchy representing a single table. Each table group hierarchy can have a parent, but it is not necessary for every hierarchy to have a parent. It is possible to configure multiple table group hierarchies for a single table if the table has multiple parents.

table group hier
Required Fields
ID

Unique identifier for the table group hierarchy.

Table Group ID

Unique identifier for the table group associated with this table in the hierarchy.

Source Catalog

Optional name for the catalog the configured table is in.

Source Schema

Optional name for the schema the configured table is in.

Source Table

The name of the source table that will be a part of this table group.

Parent ID

Unique identifier for the parent of this table in the hierarchy.

Relation Type

When writer type is non-default, this defines the structure of data in hierarchy when writing output: O=object, A=array.

Type Description

Object

The data will be in the form of a single object containing one key/value pair for each column in the source table.

Array

The data will be in the form of an array of objects with each object containing one key/value pair for each column in the source table.

Primary Column Names

The primary key column names on this table that uniquely identify a row, comma-separated.

Local Column Names

The local column names on this table that reference a parent table, comma-separated.

Parent Column Names

The parent column names on the parent that join with local columns, comma-separated.

4.11. File Triggers

In addition to supporting database synchronization, SymmetricDS also supports File Synchronization. Similar to database synchronization which allows configuring Table Triggers, SymmetricDS also supports setting up File Triggers. A file trigger is equivalent to specifying a directory structure or path that should be "watched" for files that need to be synchronized.

file trigger
Required Fields
File Trigger Id

Unique identifier for a trigger.

Channel

The channel_id of the channel that data changes will flow through.

Reload Channel Id

The channel_id of the channel that will be used for reloads.

Base Directory

The base directory on the source node that files will be synchronized from. To access Azure blob storage, use azure://container/optional-path.

Advanced Options
Recurse

Whether to synchronize child directories.

Include Files

Wildcard-enabled (*), comma-separated list of file to include in synchronization.

Exclude Files

Wildcard-enabled (*), comma-separated list of file to exclude from synchronization.

Sync On Create

Whether to capture and send files when they are created.

Sync On Modified

Whether to capture and send files when they are modified.

Sync On Delete

Whether to capture and send files when they are deleted.

Sync On Ctl File

Combined with sync_on_create, determines whether to capture and send files when a matching control file exists. The control file is a file of the same name with a '.ctl' extension appended to the end.

Delete After Sync

Determines whether to delete the file after it has synced successfully.

Before Copy Script

A beanshell script that is run at the target node right before the file copy to it’s destination directory.

After Copy Script

A beanshell script that is run at the target node right after the file copy to it’s destination directory.

4.11.1. BeanShell Scripts

There are two types of Bean Shell scripts that can be leveraged to customize file synchronization behavior:

Before copy script

This runs on delivery of a file before it is copied to it’s target location

After copy script

This run on delivery of a file after it is copied to it’s target location

Each of these scripts have access to local variables that can be read or set to affect the behavior of copying files.

targetBaseDir

The preset base directory as configured in file trigger or overwritten in file trigger router. This variable can be set by the before copy script to set a different target directory.

targetFileName

The name of the file that is being synchronized. This variable can be overwritten by the before copy script to rename a file at the target.

targetRelativeDir

The name of a directory relative to the target base directory to which the target file will be copied. The default value of this variable is the relative directory of the source. For example, if the source base directory is /src and the target base directory is /tgt and the file /src/subfolder/1.txt is changed, then the default targetRelativeDir will be subfolder. This variable can be overwritten by the before_copy_script to change the relative directory at the target. In the above example, if the variable is set to blank using the following script, then the target file will be copied to /tgt/1.txt.

targetRelativeDir = "";
processFile

This is a variable that is set to true by default. A custom before copy script may process the file itself and set this variable to false to indicate that the file should NOT be copied to its target location.

sourceFileName

This is the name of the file.

sourceFilePath

This is the path where the file can be found relative to the batch directory.

batchDir

This is the staging directory where the batch has been extracted. The batchDir + sourceFilePath + sourceFileName can be used to locate the extracted file.

engine

This is the bound instance of the ISymmetricEngine that is processing a file. It gives access to all of the APIs available in SymmetricDS.

sourceNodeId

This is a bound variable that represents the nodeId that is the source of the file.

log

This is the bound instance of an org.slf4j.Logger that can be used to log to the SymmetricDS log file.

Example 13. Example of a Before Copy Script
File file = new File(batchDir + "/" + sourceFilePath + "/" + sourceFileName);
if (file.exists()) {
    String path = file.getAbsolutePath();
    cp (path,"/backup/" + sourceFileName);
}

4.12. File Routing

The file routing concept is identical to the Table Routing concept except that it is for files versus database tables. The File Triggers specify which directory structures on the source node to watch, and the Routers define where the data is sent based on the configured router. Similar to Table Routing the simplest way to set up file routing is via the use of the Auto Create button.

auto create 1 triggers

The default entry point for the auto creation is from the file triggers perspective. This will allow you to select file triggers first then routers.

Click Next to select from available routers.

auto create 2 routers

Select the routers (direction) to associate with the previously selected file triggers.

Click Save to setup the file triggers and begin file sync.

You can also select routers on the first screen of the auto create if you would like to select routers first then file triggers. Both approaches will result in the same file routing.

auto create 1 routers

Select from available routers.

Click Next to select from available file triggers.

auto create 2 triggers

Select the file triggers to associate with the previously selected routers.

Click Save to setup the file routers and begin file sync.

The file routing elements created by the wizard can be maintained from the File Routing List Screen. New file routing entries can also be created manually from this screen if desired.

SymmetricDS will not create file triggers until they are associated with a router through file routing.
Once a file routing relationship is created changes on the base directory will be captured and file sync will begin.
file routing new
Required Fields
File Triggers

The file trigger to link.

Routers

The router to link.

Target Base Directory

The base directory on the target node that files will be synchronized to.

Conflict Strategy

The strategy to employ when a file has been modified at both the client and the server.

source_wins

The source file will be used when a conflict occurs.

target_wins

The target file will be used when a conflict occurs.

manual

If a conflict occurs the batch will be put in ER (error) status and require manual intervention to resolve the issue.

newer_wins

If a conflict occurs, a comparison between last modified date of the file incoming and the file currently on disk will be compared and the newest will win. This assumes the system clocks on each machine are in sync to perform an accurate comparison.

older_wins

If a conflict occurs, a comparison between last modified date of the file incoming and the file currently on disk will be compared and the oldest will win. This assumes the system clocks on each machine are in sync to perform an accurate comparison.

Advanced Options
Initial Load Enabled

Indicates whether this file trigger should be initial loaded.

Enabled

Indicates whether this file trigger router is enabled or not.

4.13. Conflicts

Conflict detection is the act of determining if an insert, update or delete is in "conflict" due to the target data row not being consistent with the data at the source prior to the insert/update/delete. Without any overriding configuration, the system uses a detection of USE_CHANGED_DATA and a resolution of NEWER_WINS by default.

conflict
Conflicts are broken into 3 key components in SymmetricDS:
  1. Detection - How to detect a conflict when loading data at the target.

  2. Resolution - When a row is in conflict, what to do with it when loading at the target.

  3. Ping Back - What data to capture at target during resolution, if any, that will be sent back to source.

Required Fields
Conflict Id

Unique identifier for a specific conflict detection setting.

Group Link

References a node group link.

Detection Type

Indicates the strategy to use for detecting conflicts during a dml action.

Conflicts are detected while data is being loaded into a target system.
Table 8. Detection Types

USE_PK_DATA

Indicates that only the primary key is used to detect a conflict. If a row exists with the same primary key, then no conflict is detected during an update or a delete. Updates and deletes rows are resolved using only the primary key columns. If a row already exists during an insert then a conflict has been detected.

USE_CHANGED_DATA

Indicates that the primary key plus any data that has changed on the source system will be used to detect a conflict. If a row exists with the same old values on the target system as they were on the source system for the columns that have changed on the source system, then no conflict is detected during an update or a delete. If a row already exists during an insert then a conflict has been detected.

USE_OLD_DATA

Indicates that all of the old data values are used to detect a conflict. Old data is the data values of the row on the source system prior to the change. If a row exists with the same old values on the target system as they were on the source system, then no conflict is detected during an update or a delete. If a row already exists during an insert then a conflict has been detected.

USE_TIMESTAMP

Indicates that the primary key plus a timestamp column (as configured in detect_expression ) will indicate whether a conflict has occurred. If the target timestamp column is not equal to the old source timestamp column, then a conflict has been detected. If a row already exists during an insert then a conflict has been detected.

USE_VERSION

Indicates that the primary key plus a version column (as configured in detect_expression ) will indicate whether a conflict has occurred. If the target version column is not equal to the old source version column, then a conflict has been detected. If a row already exists during an insert then a conflict has been detected.

Detection Expression

An expression that provides additional information about the detection mechanism. If the detection mechanism is use_timestamp or use_version then this expression will be the name of the timestamp or version column. The detect_expression is also used to exclude certain column names from being used. For example, to exclude column1 and column2, the expression is "excluded_column_names=column1,column2".

Resolution Type

The choice of how to resolve a detected conflict is configured via the resolve type.

Table 9. Resolution Types

NEWER_WINS

Indicates that when a conflict is detected that the either the source or the target will win based on which side has the newer timestamp or higher version number. With USE_TIMESTAMP detection, the column specified in detect_expression is used, otherwise the time of capture is used.

FALLBACK

Indicates that when a conflict is detected the system should automatically apply the changes anyway. If the source operation was an insert, then an update will be attempted. If the source operation was an update and the row does not exist, then an insert will be attempted. If the source operation was a delete and the row does not exist, then the delete will be ignored.

IGNORE

Indicates that when a conflict is detected the system should automatically ignore the incoming change. Use IGNORE between two node groups in one direction, and FALLBACK in the other direction to establish which group wins a conflict.

MANUAL

Indicates that when a conflict is detected, the batch will remain in error until manual intervention occurs. A row in error is inserted into the INCOMING_ERROR table, which includes the conflict ID, old data, new data, and current data at the target. The user can specify the resolve data to use on the next load attempt. The resolve_ignore flag can also be used to indicate that the row should be ignored.

To make a primary node group always win a conflict, use a "fallback" resolution on group links where primary is the source and an "ignore" resolution on group links where primary is the target.
Ping Back

For each configured conflict, you also have the ability to control if and how much "resolved" data is sent back to the node whose data change is in conflict.

Table 10. Ping Backs

SINGLE_ROW

The resolved data of the single row in the batch that caused the conflict is sent back to the originating node. Recommended use with MANUAL resolution, so the resolved data is sent back to the originating node.

REMAINING_ROWS

The resolved data of the single row in the batch in conflict, along with the entire remainder of the batch, is sent back to the originating node.

OFF

No data is sent back to the originating node, even if the resolved data doesn’t match the data the node sent. Recommended use with resolution types that choose a winning row, including NEWER_WINS and when IGNORE and FALLBACK are used on opposing group links.

Advanced Options
Channel

Optional channel that this setting will be applied to.

Target Catalog

Optional database catalog that the target table belongs to. Only use this if the target table is not in the default catalog.

Target Schema

Optional database schema that the target table belongs to. Only use this if the target table is not in the default schema.

Target Table

Optional database table that this setting will apply to. If left blank, the setting will be for any table in the channel (if set) and in the specified node group link.

Be aware that conflict detection will not detect changes to binary columns in the case where use_stream_lobs is true in the trigger for the table. In addition, some databases do not allow comparisons of binary columns whether use_stream_lobs is true or not.
Some platforms do not support comparisons of binary columns. Conflicts in binary column values will not be detected on the following platforms: DB2, DERBY, ORACLE, and SQLSERVER.

4.14. Transforms

Transforms allow you to manipulate data on a source node or target node, as the data is being loaded or extracted.

The source table must first be configured for synchronization (see Table Triggers) before it can be transformed.

The source trigger captures data changes that are passed through transformations during extraction from the source and during loading at the target. If any enhancement with additional data is needed, an extract transform can access the source database, while a load transform can access the target database. A single row change may be processed by multiple transforms on the same group link, which transforms it into multiple target rows. Rows for the same target table do not merge, which means multiple transforms with the same target table results in multiple rows for that target table.

transform
Required Fields
Transform Id

Unique identifier of a specific transform.

Group Link

The group link defining which direction the transform will process.

Transform Point

Where this transform will occur. The options include:

Table 11. Transform Points

EXTRACT

The transform will execute while data is being extracted from the source. This means the transform will have access to the source’s database.

LOAD

The transform will execute while data is being loaded into the target. This means the transform will have access to the target’s database.

Column Policy

Indicates whether unspecified columns are passed thru or if all columns must be explicitly defined. The options include:

SPECIFIED

Indicates that only the transform columns that are defined will be the ones that end up as part of the transformation.

IMPLIED

Indicates that if not specified, then columns from the source are passed through to the target. This is useful if you just want to map a table from one name to anther or from one schema to another. It is also useful if you want to transform a table, but also want to pass it through. You would define an implied transform from the source to the target and would not have to configure each column.

Source Catalog

Name of the catalog of the configured source table. This should only be set if Use Source Catalog/Schema or Target Catalog are set on the Router. Parameter values can be substituted using $(name) syntax.

Source Schema

Name of the schema for the configured source table. This should only be set if Use Source Catalog/Schema or Target Schema are set on the Router. Parameter values can be substituted using $(name) syntax.

Source Table

The name of the source table that will be transformed. Parameter values can be substituted using $(name) syntax.

Target Catalog

Optional name for the catalog a target target table is in. Only use this if the target table is not in the default catalog. Parameter values can be substituted using $(name) syntax.

Target Schema

Optional name of the schema a target target table is in. Only use this if the target table is not in the default schema. Parameter values can be substituted using $(name) syntax.

Target Table

The name of the target table. Parameter values can be substituted using $(name) syntax.

Advanced Options
Update First

This option overrides the default behavior for an Insert operation. Instead of attempting the Insert first, SymmetricDS will always perform an Update first and then fall back to an Insert if that fails. Note that, by default, fall back logic always applies for Insert and Updates. Here, all you a specifying is whether to always do an Update first, which can have performance benefits under certain situations you may run into.

Delete Action

An action to take upon delete of a row.

Table 12. Transform Points

DEL_ROW

The delete results in a delete of the row as specified by the pk columns defined in the transformation configuration.

UPDATE_COL

The delete results in an update operation on the target which updates the specific rows and columns based on the defined transformation.

NONE

The delete results in no target changes.

Update Action

An action to take upon update of a row.

UPD_ROW

The update performs normally.

INS_ROW

The update is transformed into an insert instead.

DEL_ROW

The update is transformed into a delete instead.

NONE

The update is ignored and no changes are made.

Transform Order

For a single source operation that is mapped to a transformation, there could be more than one target operation that takes place. You may control the order in which the target operations are applied through a configuration parameter defined for each source-target table combination. This might be important, for example, if the foreign key relationships on the target tables require you to execute the transformations in a particular order.

4.14.1. Columns

Transforms are not complete until the columns involved in the transformation have been defined. Typically there will be several columns defined for each transform, each of which will define a source column and a target column.

To see a transform’s columns, select the transform and click "Edit Columns".

transform edit columns
PK

Indicates that this mapping is used to define the "primary key" for identifying the target row(s) (which may or may not be the true primary key of the target table). This is used to define the "where" clause when an Update or Delete on the target is occurring.

Unless the column policy is "IMPLIED" at least one row marked as a pk should be present for each transform_id.
Source

The source column name to be transformed.

Target

The target column name to be transformed.

Transform On

Defines whether this entry applies to source operations of Insert, Update, Delete, or All.

Type

The name of a specific type of transform, default type is "copy". See Transform Types for more information.

Expression

An expression that is specific to the type of transform that is configured in transform_type. See Transform Types for more information.

Order

In the event there are more than one columns to transform, this defines the relative order in which the transformations are applied.

4.14.2. Data Format

Transforms have access to the change data in character form using a standard set of data formats. Modification to change data should also be character data following the format for the target field.

Table 13. Data Formats
Type Format

binary

Hex or Base64 encoding, depending on the source database dialect. Use context.getBatch().encodeBinary() and context.getBatch().decodeBinary().

boolean

Either 1 or 0. (Some databases require a small integer to represent boolean.)

date

Formatted as yyyy-MM-dd.

float

Decimal digits with a decimal point expressed as one period or comma. A negative (-) sign and an exponent (e or E) are also allowed.

geometry

Well known text (WKT) format

integer

Decimal digits. A negative (-) sign is also allowed.

numeric

Decimal digits with a decimal point expressed as one period or comma. A negative (-) sign is also allowed.

time

Formatted as time HH:mm:ss.S or a timestamp yyyy-MM-dd HH:mm:ss.S.

timestamp

Formatted as yyyy-MM-dd HH:mm:ss.S.

4.14.3. Transform Types

There are several pre-defined transform types. Additional ones can be defined by creating and configuring an extension point which implements the IColumnTransform interface.

Table 14. Transform Types
Type Description

additive

For numeric fields, apply the difference of new minus old to the target. Calculate the difference between the old and new source number and apply the difference to the target.

bsh

Run a BeanShell script to return a value.

bleft

Copy the leftmost bytes of the value.

clarionDateTime

Convert a source column with a Clarion integer into a timestamp.

const

Use a provided constant value.

copy

Copy source column value to the target column.

copyIfChanged

Copy the source column value only if its old and new value are different.

columnsToRowsKey

Convert columns to rows, specifying which columns are the key for all rows.

columnsToRowsValue

Convert column to rows, specifying which columns are the changing values.

deletedColumns

Comma-separated list of column names that were changed to null.

identity

Set INSERT_IDENTITY to ON for generating the next identity value.

isBlank

If value is empty after trimming spaces, then replace its value.

isEmpty

If value is empty, then replace its value.

java

Run Java code to return a value.

isNull

If value is null, then replace its value.

left

Copy the leftmost characters of the value.

lookup

Run a SQL query to return a value.

math

Perform a math expression.

multiply

Run a SQL query with the primary key to create multiple rows.

remove

Remove the column from the row.

substring

Copy a substring of the value.

valueMap

Map from one literal value to another.

variable

Use a built-in variable value.

Copy Transform

This transformation type copies the source column value to the target column. This is the default behavior.

transform type copy
Remove Transform

This transformation type excludes the source column. This transform type is only valid for a table transformation type of 'IMPLIED' where all the columns from the source are automatically copied to the target.

transform type remove
Constant Transform

This transformation type allows you to map a constant value to the given target column. The constant itself is placed in transform expression.

transform type const
Variable Transform

This transformation type allows you to map a built-in dynamic variable to the given target column. The variable name is placed in transform expression. The following variables are available:

Table 15. Variables

batch_id

Batch number

batch_start_time

Timestamp of when the batch started to process

delete_indicator_flag

Whether or not this is a delete event, in format of Y or N.

null

null value

old_column_value

column’s old value prior to the DML operation.

source_catalog_name

Catalog name of source table

source_dml_type

Event type, usually one of INSERT, UPDATE, or DELETE, but possibly one of RELOAD, SQL, CREATE, or BSH.

source_node_id

node id of the source (from the batch)

source_node_id_from_data

source_node_id value from sym_data (source of a captured synchronization data change)

source_schema_name

Schema name of the source table

source_table_name

Name of the source table

system_date

current system date

system_timestamp

current system date and time using default timezone

system_timestamp_utc

current system date and time using UTC timezone

target_node_id

node id of the target (from the batch)

transform type variable
Additive Transform

This transformation type is used for numeric data. It computes the change between the old and new values on the source and then adds the change to the existing value in the target column. That is, target = target + multiplier (source_new - source_old), where multiplier is a constant found in the transform expression (default is 1 if not specified).

Example 14. Additive Transform Example

If the source column changed from a 2 to a 4, the target column is currently 10, and the multiplier is 3, the effect of the transform will be to change the target column to a value of 16 ( 10+3*(4-2) ⇒ 16 ).

In the case of deletes, the new column value is considered 0 for the purposes of the calculation.
transform type additive
Substring Transform

This transformation computes a substring of the source column data and uses the substring as the target column value. The transform expression can be a single integer ( n , the beginning index), or a pair of comma-separated integers ( n,m - the beginning and ending index). The transform behaves as the Java substring function would using the specified values in transform expression.

transform type substring
Left Transform

This transform copies the left most number of characters specified.

BLeft Transform

This transform copies the left most number of bytes specified.

Lookup Transform

This transformation determines the target column value by using a query, contained in transform expression to lookup the value in another table. The query must return a single row, and the first column of the query is used as the value. Your query references source column values by prefixing with a colon (e.g., :MY_COLUMN). Also, you can reference old values with :OLD_COLUMN and previously transformed columns (see transform order) with :TRM_COLUMN.

transform type lookup
Multiply Transform

This transformation allows for the creation of multiple rows in the target table based on the transform expression. This transform type can only be used on a primary key column. The transform expression is a SQL statement, similar to the lookup transform, except it can return multiple rows that result in multiple rows for the target table. The first column of the query is used as the value for the target column. The query can reference source column values by prefixing them with a colon (e.g., :MY_COLUMN).

BeanShell Script Transform

This transformation allows you to provide a BeanShell script in the transform expression and executes the script at the time of transformation. Beanshell transforms can return either a String value or an instance of NewAndOldValue. Some variables are provided to the script:

Table 16. Variables

<COLUMN_NAME>

The variable name is the source column name in uppercase of the row being changed (replace <COLUMN_NAME> with your column)

currentValue

The value of the current source column

oldValue

The old value of the source column for an updated row

sqlTemplate

org.jumpmind.db.sql.ISqlTemplate object for querying or updating the database

channelId

name of the channel on which the transformation is happening

sourceNode

org.jumpmind.symmetric.model.Node object that represents the node from where the data came

sourceNodeId

same as sourceNode.getNodeId()

sourceNodeGroupId

same as sourceNode.getNodeGroupId()

sourceNodeExternalId

same as sourceNode.getNodeExternalId()

targetNode

org.jumpmind.symmetric.model.Node object that represents the node where the data is being loaded.

targetNodeId

same as targetNode.getNodeId()

targetNodeGroupId

same as targetNode.getNodeGroupId()

targetNodeExternalId

same as targetNode.getNodeExternalId()

transformColumn

org.jumpmind.symmetric.io.data.transform.TransformColumn that is the transform configuration

includeOn

org.jumpmind.symmetric.io.data.transform.TransformColumn.IncludeOnType, same as transformColumn.getIncludeOn(), tells whether column transform is configured for all, insert, update, or delete

sourceSchemaName

source schema name that the transform matched

sourceCatalogName

source catalog name that the transform matched

sourceTableName

source table name that the transform matched

transformedData

org.jumpmind.symmetric.io.data.transform.TransformedData, the model object representing the outputted transformed data

sourceDmlType

org.jumpmind.symmetric.io.data.DataEventType that is the source row change type, either insert, update, or delete

sourceDmlTypeString

same as sourceDmlType.toString(), returning insert, update, or delete

log

org.slf4j.Logger, write to the log file

context

org.jumpmind.symmetric.io.data.DataContext containing internal variables and also acts like a Map for sharing variables between transforms for the current sync session

bshContext

java.util.Map, static map of variables to share between transforms

engine

org.jumpmind.symmetric.ISymmetricEngine, access to engine functions and services

Example 15. Transform Expression Example Returning a String
if (currentValue > oldValue) {
	return currentValue * .9;
} else {
	return PRICE;
}
Example 16. Transform Expression Example Returning a NewAndOldValue object
if (currentValue != null && currentValue.length() == 0) {
	return new org.jumpmind.symmetric.io.data.transform.NewAndOldValue(null, oldValue);
} else {
	return currentValue;
}
Example 17. Transform Expression Example Accessing Old/New Values for the Additional Column 'path'
String newFilePath = PATH;
String oldFilePath = null;
if (transformedData.getOldSourceValues() != null) {
    oldFilePath = transformedData.getOldSourceValues().get("path");
}
if (oldFilePath == null) {
    return newFilePath;
} else {
    return oldFilePath;
}
transform type beanshell
Identity Transform

This transformation allows you to insert into an identity column by letting the database compute a new identity, instead of loading an explicit value from a source database. This transform is needed on databases like SQL-Server and Sybase, which have an INSERT_IDENTITY option that is normally ON for normal data sync. By using the identity transform, the INSERT_IDENTITY is set to OFF, so the next value is generated by the database.

transform type identity
Mathematical Transform

This transformation allows you to perform mathematical equations in the transform expression. Some variables are provided to the script:

#{COLUMN_NAME}

A variable for a source column in the row, where the variable name is the column name in uppercase (replace COLUMN_NAME with your column name).

#{currentValue}

The value of the current source column

#{oldValue}

The old value of the source column for an updated row.

transform type math
Figure 5. Transform Expression Example
Copy If Changed

This transformation will copy the value to the target column if the source value has changed. More specifically, the copy will occur if the the old value of the source does not equal the new value.

Table 17. Target Expression Options

IgnoreColumn

If old and new values are equal, the COLUMN will be ignored

{empty string}

If old and new values are equal, the ROW will be ignored

transform type copy if change
Figure 6. If values are equal, ignore only the COLUMN
transform type copy if change blank
Figure 7. If values are equal, ignore the entire ROW
Value Map Transform

This transformation allows for simple value substitutions through use of the transform expression. The transform expression should consist of a space separated list of value pairs of the format sourceValue=TargetValue. The column value is used to locate the correct sourceValue, and the transform will change the value into the corresponding targetValue. A sourceValue of * can be used to represent a default target value in the event that the sourceValue is not found. Otherwise, if no default value is found, the result will be null.

Example 18. Value Map Examples
transform expression source value target value (result)

s1=t1 s2=t2 s3=t3 *=t4

s1

t1

s1=t1 s2=t2 s3=t3 *=t4

s2

t2

s1=t1 s2=t2 s3=t3 *=t4

s3

t3

s1=t1 s2=t2 s3=t3 *=t4

s4

t4

s1=t1 s2=t2 s3=t3 *=t4

s5

t4

s1=t1 s2=t2 s3=t3 *=t4

null

t4

transform type valueMap
Clarion Date Time

Convert a Clarion date column with optional time column into a timestamp. Clarion dates are stored as the number of days since December 28, 1800, while Clarion times are stored as hundredths of a second since midnight, plus one. Use a source column of the Clarion date and a target column of the timestamp. If the Clarion time exists in a separate column it can optionally be provided through the transform expression to be included in the target timestamp column.

Columns To Rows

Convert column values from a single source row into a row per column value at the target. Two column mappings are needed to complete the work:

columnsToRowsKey

Maps which source column is used

column1=key1,column2=key2
columnsToRowsValue

Maps the value

changesOnly=true

Convert only rows when the old and new values have changed

ignoreNulls=true

Convert only rows that are not null

TODO add image

Example 19. Example

"fieldid" mapped as "columnsToRowsKey" with expression of "user1=1,user2=2" and column "color" mapped as "columnsToRowsValue" would convert a row with columns named "user1" and "user2" containing values "red" and "blue" into two rows with columns "fieldid" and "color" containing a row of "1" and "red" and a row of "2" and "blue".

isEmpty Transform

This transformation checks to see if a string is null or zero length. If it is empty the replacement value will be used. If no value is provided null will be used as a default replacement for empty values.

isBlank Transform

This transformation checks to see if a string is null or zero length after trimming white spaces. If it is blank the replacement value will be used. If no value is provided null will be used as a default replacement for blank values.

Null Value Transform

This transformation checks to see if the source value is null and if so replaces it with the provided value.

Deleted Columns Transform

For an update, this transform returns a comma-separated list of columns names that were set to null and previously not null.

Java Transform

Java Transform ('java'): Use Java code in the transform expression that is included in the transform method of a class that extends JavaColumnTransform. The class is compiled whenever the transform expression changes and kept in memory for runtime. The code must return a String for the new value of the column being mapped.

Some variables are provided to the code:

Table 18. Variables
Variable Name Java Type Description

platform

org.jumpmind.db.platform.IDatabasePlatform

The platform for the database that this node is connected to

context

org.jumpmind.symmetric.io.data.DataContext

The data cotext for the synchronization of the current batch

column

org.jumpmind.symmetric.io.data.transform.TransformColumn

The transform column

data

org.jumpmind.symmetric.io.data.transform.TransformedData

The transformed data

sourceValues

java.util.Map<java.lang.String, java.lang.String>

The map of source values

newValue

java.lang.String

The captured new value

oldValue

java.lang.String

The captured old value

Example 20. Transform Expression Example Returning a String
if (sourceValues.containsKey("OLDKEY")) {
    return sourceValues.get("OLDKEY");
} else {
    return sourceValues.get("NEWKEY");
}

4.14.4. Virtual Columns

Transforms provide the ability to create "virtual columns" which can pass data between nodes for use by other SymmetricDS processes.

Use cases for virtual columns
  1. Extract transform adds virtual column to be processed by a target load transform.

  2. Extract transform adds virtual column to be processed by a target load filter.

  3. Extract transform adds virtual column to be processed by a source router.

Example 21. Example of an extract transform passing a virtual column to a target load transform
transform virtual columns
Figure 8. Create two transforms, one for extract and one for target using different group links
transform virtual columns extract
Figure 9. Create lookup transform for the extract transform to create a new virtual column to be sent to target.
transform virtual columns load
Figure 10. Create copy transform for the load transform to populate the cost column from the virtual column that was sent over.

4.15. Load Filters

Load Filters are a way to take a specific action when a row of data is loaded by SymmetricDS at a destination database node.

Load filters run for each row of data being loaded.
load filter
Required Fields
Filter Id

The unique identifier for the load filter

Group Link

The group link for with the load filter will be applied.

Type

The type of load filter. Today only Bean Shell, Java, and SQL are supported ('BSH', 'Java', 'SQL').

Target Table

The table on the target which the load filter will execute when changes occur on it.

Use the wildcard * to specify all tables configured through the group link. Partial table names in conjunction with a wildcard are NOT supported. If the wildcard is used it should be the only value.
Filter Order

The order in which load filters should execute if there are multiple scripts pertaining to the same source and target data.

Advanced Options
Filter On Update

Determines whether the load filter takes action (executes) on a database update statement.

Filter On Insert

Determines whether the load filter takes action (executes) on a database insert statement.

Filter On Delete

Determines whether the load filter takes action (executes) on a database delete statement.

Fail On Error

Whether we should fail the batch if the filter fails.

Target Catalog

The name of the target catalog for which you would like to watch for changes.

Target Schema

The name of the target schema for which you would like to watch for changes.

4.15.1. Load Filter Scripts

Load filters are based on the execution of a script. You have the ability to set the execution point of the script at 6 different points. A script can be provided for one or more of these execution points.

Return Values
  • Return true to load the row of data.

  • Return false to not load the row of data.

load filter edit scripts
Figure 11. Edit load filter scripts by selecting the load filter and clicking the Edit Scripts button.
Available Load Filter Scripts
Before Write Script

The script to execute before the database write occurs.

After Write Script

The script to execute after the database write occurs.

Batch Complete Script

The script to execute after the entire batch completes.

Batch Commit Script

The script to execute after the entire batch is committed.

Batch Rollback Script

The script to execute if the batch rolls back.

Handle Error Script

A script to execute if data cannot be processed.

Table 19. Variables available within scripts
Variable BSH SQL JAVA Description

engine

X

The Symmetric engine object.

COLUMN_NAME

X

X

The source values for the row being inserted, updated or deleted.

OLD_COLUMN_NAME

X

X

The old values for the row being inserted, updated or deleted.

context

X

X

The data context object for the data being inserted, updated or deleted. .

table

X

X

The table object for the table being inserted, updated or deleted.

data

X

X

The CsvData object for the data change.

error

X

X

java.lang.Exception

Example 22. Example of simple load filter
load filter example 1
Figure 12. Create new load filter
load filter example 1 bsh
Figure 13. Provide a "Before Write" bsh script.
Example 23. Example load filter to send email on error
load filter example
Figure 14. Create new load filter
load filter example bsh
Figure 15. Select filter and hit "Edit Scripts" button, then select "Handle Error Script" and enter the bsh script below.
Beanshell Expression
authListener = new javax.mail.Authenticator() {
  protected javax.mail.PasswordAuthentication getPasswordAuthentication() {
    return new javax.mail.PasswordAuthentication(engine.getParameterService().getString("mail.smtp.username"),
       engine.getParameterService().getString("mail.smtp.password"));
  }
};

if (bsh.shared.mailMap == void) {
  bsh.shared.mailMap = new HashMap();
}

String batchId = context.getBatch().getNodeBatchId();
String targetNodeId = context.getBatch().getTargetNodeId();
if (!bsh.shared.mailMap.containsKey(batchId)) {
  bsh.shared.mailMap.put(batchId, Boolean.TRUE);
  javax.mail.Session session = javax.mail.Session.getInstance
    (engine.getParameterService().getAllParameters(), authListener);
  javax.mail.internet.MimeMessage msg = new
    javax.mail.internet.MimeMessage(session);
  msg.setFrom(new javax.mail.internet.InternetAddress
    (engine.getParameterService().getString("mail.smtp.from")));
  msg.setRecipients(javax.mail.Message.RecipientType.TO,
    engine.getParameterService().getString("mail.smtp.to"));
  msg.setSubject("SymmetricDS - batch " + batchId + " is in error at node " + targetNodeId);
  msg.setSentDate(new java.util.Date());
  msg.setText(org.apache.commons.lang.exception.ExceptionUtils.
    getFullStackTrace(error));
  javax.mail.Transport.send(msg);

}

4.15.2. Custom Load Filters

Custom load filters can be created by implementing the IDatabaseWriterFilter, see IDatabaseWriterFilter for more information.

4.16. Extensions

Extensions are custom code written to a plug-in interface, which allows them to run inside the engine and change its default behavior. Saving extension code in the configuration has the advantage of dynamically running without deployment or restarting. Configured extensions are available to other nodes and move between environments when configuration is exported and imported.

Extension Id

Identifier for a unique extension entry.

Extension Type

Type of extension, either written in Java or BeanShell. Java extensions are compiled to bytecode on first use and may be compiled to native code by the Just-In-Time (JIT) compiler, giving them the best performance. BeanShell extensions are parsed on first use and interpreted at runtime, but they are easier to write because of loose typing and short-cuts with syntax.

Table 20. Options for Extension Type

Java

Indicates that Java code is provided in the extension text.

BSH

Indicates that BeanShell code is provided in the extension text. Built-in variables are available for engine, sqlTemplate, and log.

Interface Name

The full class name for the interface implemented by the extension, including the package name. Only needed for extension type of BSH.

Node Group Id

The node group where this extension will be active and run.

Enabled

Whether or not the extension should be run.

Extension Order

The order to register extensions when multiple extensions for the same interface exist.

Extension Text

The code for the extension that will be compiled or interpreted at runtime.

Example 24. BSH extension that adds a new transform for masking characters

Add a new transform type called "mask" that replaces all characters in a string with an asterisk except the last number of characters specified by the user in the expression. This BeanShell extension uses the ISingleValueColumnTransform interface and applies only to the "corp" node group.

Click the New button to create a new extension.

extension new

Click the Edit Script button to edit the script for the extension.

import org.apache.commons.lang.StringUtils;

isExtractColumnTransform() {
    return true;
}

isLoadColumnTransform() {
    return true;
}

transform(platform, context, column, data, sourceValues, newValue, oldValue) {
    if (StringUtils.isNotBlank(newValue)) {
        String expression = column.getTransformExpression();
        if (StringUtils.isNotBlank(expression)) {
            count = newValue.length() - Integer.parseInt(expression.trim());
            return StringUtils.repeat("*", count) + newValue.substring(count);
        }
    }
    return newValue;
}
Extensions Tips and Tricks
For BeanShell, implement only the methods needed from an interface, then write a special method of "invoke(method, args) {}" that will be called for any unimplemented methods.

4.17. Jobs

In SymmetricDS, jobs are tasks that are scheduled to run by a job manager. These jobs do the majority of synchronization work for SymmetricDS. In addition to the built in jobs, you can create custom jobs in Configuration to run for specified node groups.

customJob
Required Fields
Job Name

The Job Name is a unique identifier to refer to the job

Job Type

The type of job. Possible types are listed below.

Type Provided Description

BSH

x

A job that will compile and run a beanshell script.

JAVA

x

A job that will compile and run a java script.

SQL

x

A job that will compile and run a sql script.

Node Group

Targets the job at a specific node group id. To target all groups, use the value of 'ALL'.

Schedule

Specifies how often this job should run. If the schedule is just a number, it is interpreted as milliseconds. So 60000 means to run the job every minute. Otherwise, a cron expression can be specified. Some sample cron expressions include:

  • Every fifth minute: 0 0/5 * * * *

  • Every night at midnight: 0 0 0 * * *

  • Every ten minutes at the 1 o’clock hour: 0 0/10 1 * * *

Note that parameter values will be created for the schedule, in the form of job.<name>.cron OR job.<name>.period.time.ms, depending on if a cron of period of time was entered. If both the cron and period parameters are present, the cron will take precedence.

Advanced Topics
Job Expression

The payload of the job. For BSH jobs, this should be a beanshell script. For Java jobs, this should be Java code of a class which implements the IJob interface. For SQL jobs, this should be a sql script. BSH jobs are written in beanshell script (http://www.beanshell.org/). The beanshell has the following variables available:

  • engine - An instances of the current ISymmetricEngine

  • sqlTemplate - An instance of ISqlTemplate which can be used to run SQL statements.

  • log - An logger which can be used to write messages to the log.

This is an example BSH job:

    // Configuration for this extension...
    final String TABLE_NAME = "item";
    final String CHANNEL_ID = "item";
    final String ROUTER_ID = "corp_2_store";
    // End Configuration.
    String INSERT_RELOAD_DATA_TEMPLATE =
            "insert into sym_data ( table_name, event_type, row_data, trigger_hist_id, channel_id, transaction_id, create_time)" +
            "    select  t.source_table_name, 'R', 'item_id > ''1''', " +
            "            h.trigger_hist_id, t.channel_id, '1', current_timestamp" +
            "        from sym_trigger t inner join sym_trigger_router tr on" +
            "            t.trigger_id=tr.trigger_id inner join sym_trigger_hist h on" +
            "            h.trigger_hist_id=(select max(trigger_hist_id) from sym_trigger_hist" +
            "                where trigger_id=t.trigger_id)" +
            "    where channel_id=? and" +
            "        tr.router_id like ? and" +
            "        (t.source_table_name = ?)" +
            "    order by tr.initial_load_order asc;";

        int updatedCount = sqlTemplate.update(INSERT_RELOAD_DATA_TEMPLATE, new Object[]{CHANNEL_ID, ROUTER_ID, TABLE_NAME});
        if (updatedCount == 1) {
            log.info("Issued reload for table '" + TABLE_NAME + "'");
        } else {
            throw new SymmetricException("SyncOnHeartbeatExtension is designed to insert exactly 1 sym_data row.  Instead inserted " +
                    updatedCount + ". Check TABLE_NAME, CHANNEL_ID, ROUTER_ID parameters in the extension itself.",  null);
        }

This is the same job as a Java job. This might perform a little better, but you’ll need to have a JDK (not just JRE) available on your nodes for this to compile.

    // Configuration for this extension...
    final String TABLE_NAME = "item";
    final String CHANNEL_ID = "item";
    final String ROUTER_ID = "corp_2_store";
    // End Configuration.
    String INSERT_RELOAD_DATA_TEMPLATE =
            "insert into sym_data ( table_name, event_type, row_data, trigger_hist_id, channel_id, transaction_id, create_time)" +
            "    select  t.source_table_name, 'R', 'item_id > ''1''', " +
            "            h.trigger_hist_id, t.channel_id, '1', current_timestamp" +
            "        from sym_trigger t inner join sym_trigger_router tr on" +
            "            t.trigger_id=tr.trigger_id inner join sym_trigger_hist h on" +
            "            h.trigger_hist_id=(select max(trigger_hist_id) from sym_trigger_hist" +
            "                where trigger_id=t.trigger_id)" +
            "    where channel_id=? and" +
            "        tr.router_id like ? and" +
            "        (t.source_table_name = ?)" +
            "    order by tr.initial_load_order asc;";

        log.info("TESTING\n\n\n");
        int updatedCount = sqlTemplate.update(INSERT_RELOAD_DATA_TEMPLATE, CHANNEL_ID, ROUTER_ID, TABLE_NAME);
        if (updatedCount == 1) {
            log.info("Issued reload for table '" + TABLE_NAME + "'");
        } else {
            throw new org.jumpmind.symmetric.SymmetricException("SyncOnHeartbeatExtension is designed to insert exactly 1 sym_data row.  Instead inserted " +
                    updatedCount + ". Check TABLE_NAME, CHANNEL_ID, ROUTER_ID parameters in the extension itself.");
        }

This is an example of a SQL job that updates the heartbeat_time of the current node. Available tokens are: :NODE_ID and :NODE_GROUP_ID.

update sym_node_host set heartbeat_time = now() where node_id = :NODE_ID;
delete from item where item_id = 0;
insert into ITEM (ITEM_ID, NAME) values (0,'tesing');

To add your job expression/script navigate to the Configure Jobs page, select your job and click the button "Edit Script". The following editor will appear.

editScript

4.18. Parameters

Parameters can be used to help tune and configure your SymmetricDS configuration. Parameters can be set for an individual node or for all nodes in your network.

See Parameter List for a complete list of parameters.

4.18.1. Targeting

Parameters can be set for a given node or node group to allow more flexibility in configuration.

parameters target
Figure 16. Setting parameters by target
Only certain parameters can be edited on this screen. To edit a SymmetricDS startup parameter, you must go to the Manage→Startup Parameters screen or locate and edit the engine.properties file.

4.18.2. Filtering

Parameters can be filtered in order to help find specific parameters or parameters named a certain way.

parameters filter
Figure 17. Filter parameters by typing in a portion of the parameter name

4.19. Users

Additional users may be set up to access the SymmetricDS web console.

users
Required Fields
Authentication Method

Method to authenticate the user.

Internal

Internal refers to users being stored in the SYM_CONSOLE_USER table.

LDAP

LDAP refers to users being authenticated through an LDAP server. See the LDAP section for more information.

SAML

SAML refers to users being authenticated through a SAML identity provider. See the SAML section for more information.

User Id

Unique Identifier for a User.

Password

Password for a User.

First Name

First Name for a User.

Last Name

Last Name for a User.

Role

Role for the user.

Admin

Admin users have access to all functionalities of the SymmetricDS web console, including the Manage, Configure, and Explore tabs. Admin users also have full access to the connected databases.

Read

Read users can view the Manage and Configure tabs, but have no control over editing. Read users don’t have any access to the Explore tab, SSO settings, Mail Server Settings, Monitor settings, or Notification settings.

Write

Write users have access to most functionalities of the SymmetricDS web console, including the Manage, Configure, and Explore tabs. Write users also have full access to the connected databases, but do not have any access to SSO settings, Mail Server Settings, Monitor settings, or Notification settings.

Email

Email address for a user which can be used to recover a forgotten password.

Password Reset Required

If enabled, requires the user to change their password the next time they login.

Glass Break Required

If enabled, another user must break the glass for this user to login. This option is only compatible with internal authentication.

Enabled

If disabled, prevents the user from logging in.

users add

4.20. Roles

Additional user roles may be set up to facilitate fine-grained control of permissions and user creation.

Role Id

Unique identifier for a role.

Role Description

Description for the role.

Parent Role Id

Role Id of this role’s parent role. Leaving this blank makes this role non-hierarchical.

Non-Hierarchical

Roles that have neither a parent role nor any children roles, setting them apart from any hierarchy. Users with these roles have control over the creation and manipulation of any other users (except for admins) with non-hierarchical roles. The built-in Read and Write roles are both non-hierarchical by default.

Hierarchical

Roles that have a parent role, children roles, or both, which situates them in a hierarchy. Users with these roles have control over the creation and manipulation of other users with hierarchical roles that are descendent relative to their own.

Admin

The built in admin role. Users with this role have control over the creation and manipulation of all users, as well as access to any connected databases and all functionalities of the SymmetricDS web console.

4.21. Single Sign-On

Users can utilize single sign-on using LDAP or SAML authentication.

4.21.1. LDAP

Configuring a Users authentication on SymmetricDS using LDAP is as simple as setting up the following four Parameters.

console.auth.ldap.baseDN

The LDAP base DN to search for a user or group (e.g. dc=example,dc=com). [ Default: ]

A list of base DNs can be specified by separating each entry by a pipe ("|") symbol.
For example:
ou=Users,o=IT,c=US,dc=corp,dc=local|ou=Users,o=Tech,c=US,dc=corp,dc=local

console.auth.ldap.host

The LDAP server host name (e.g. ldap://hostname:389/). [ Default: ]

console.auth.ldap.searchAttribute

LDAP attribute to search for (e.g. uid). [ Default: ]

console.auth.ldap.securityPrincipal

Format to use for the LDAP security principal (just ${username} may work for Active Directory). [ Default: ${searchAttribute}=${username},${baseDN} ]

It is also possible for LDAP users to log in without having a corresponding console user. This feature relies on the following five Parameters.

console.auth.ldap.role.group.map

Map of LDAP groups to console role IDs. Allows LDAP users to log in without having a corresponding console user if they are a member of one of the specified groups. [ Default: ]

If a user is a member of multiple groups, the group that is specified first will take priority. This parameter does not account for group inheritance, meaning that each individual group must be specified. The entire DN must be specified for each group. The map must use the following format:
"group0==roleId0","group1==roleId1","group2==roleId2"

console.auth.ldap.adminDN

The DN of the user to log in as while determining whether a user belongs to one of the groups specified by the console.auth.ldap.role.group.map parameter. If blank, the credentials from the login screen will be used. [ Default: ]

console.auth.ldap.admin.password

The password for the user specified by the console.auth.ldap.adminDN parameter. [ Default: ]

console.auth.ldap.admin.use.memberOf

Whether or not the LDAP server maintains a "memberOf" attribute. This only has an effect if the console.auth.ldap.role.group.map parameter is set. [ Default: true ]

console.auth.ldap.admin.memberAttribute

The attribute of an LDAP group to check to determine whether or not a user is a member of the group. This only has an effect if the console.auth.ldap.role.group.map parameter is set and the console.auth.ldap.use.memberOf parameter is set to false. [ Default: member ]

These parameters can be set within the LDAP tab on the Configure SSO page by filling out the form and clicking Save.

ldap save

These parameters can also be set by manually by entering the desired values in Configure→Parameters.

ldap parameters

You may test the values that have been input into the form to test the connection to your LDAP directory. To do this, click the Test button, enter the user credentials and click Test to test the authentication. If the authentication succeeded, save the form.

ldap authentication
You must have values for the LDAP URL, Base DN, and Search Attributes fields or the test will fail.

4.21.2. SAML

To configure a Users authentication on SymmetricDS using SAML, you have to set four Parameters.

console.auth.saml.sp.nameid.format

Specifies constraints on the name identifier to be used to represent the requested subject during SAML authentication.

console.auth.saml.idp.entityid

Identifier of the SAML identity provider entity (must be a URI).

console.auth.saml.idp.sso.service.url

SSO endpoint info of the SAML identity provider. URL target of the identity provider where the service provider will send the authentication request message.

console.auth.saml.idp.cert

Path to the public x509 certificate of the SAML identity provider.

It is also possible for SAML users to log in without having a corresponding console user. This feature relies on the following two Parameters.

console.auth.saml.role.group.map

Map of SAML groups to console role IDs. Allows SAML users to log in without having a corresponding console user if they are a member of one of the specified groups. [ Default: ]

If a user is a member of multiple groups, the group that is specified first will take priority. This parameter does not account for group inheritance, meaning that each individual group must be specified. The map must use the following format:
"group0==roleId0","group1==roleId1","group2==roleId2"

console.auth.saml.admin.groupAttribute

The attribute of an SAML response to check to determine whether or not a user is a member of a SAML group. This only has an effect if the console.auth.saml.role.group.map parameter is set. [ Default: group ]

These parameters can be set within the SAML tab on the Configure SSO page by filling out the form and clicking Save.

saml save

These parameters can also be set by manually by entering the desired values in Configure→Parameters.

saml parameters

You may test the values that have been input into the form to test the connection to your SAML identity provider. You must save your changes in the form if you want them to apply when you run the SAML test. To do this, make sure that the current node is one that you set the parameters for, and then click the Test button. Then click Test to test the authentication.

saml authentication
You must have values for all four of the required parameter fields or the test will fail.

If you have only specified the first four parameters, the console user’s ID must match the name ID returned by the identity provider for authentication to succeed. The console user must also be configured to use SAML authentication.

When configuring SymmetricDS as a service provider, specify the below endpoints in your identity provider’s settings. You can obtain the service provider metadata in XML format by visiting the Metadata URL.

Service Provider Entity ID / Metadata URL

http://{hostname}:{port}/saml/{engine.name}/metadata

Service Provider ACS URL

http://{hostname}:{port}/saml/{engine.name}/acs

SymmetricDS is compatible with many different SAML identity providers. The main requirements for compatibility are that the identity provider must accept a SAML authentication request using the HTTP-Redirect binding and return a SAML response using the HTTP-POST binding. While compatibility is not limited to only these identity providers, the following IdPs have been tested and are confirmed to be compatible with SymmetricDS:

  • Auth0

  • CyberArk

  • JumpCloud

  • miniOrange

  • Okta

  • OneLogin

  • Ping Identity

4.22. License Key

The web console requires a valid license key before it will allow access to screens. The user is prompted automatically to enter a valid license key if one is missing. The license key can also be updated from the Configure License screen.

License Info

A description of the license type.

Holder

The company name or email address of the license holder.

Effective Date

Date and time when the license will become valid.

Expiration Date

Date and time when the license will no longer be valid.

Max Maintenance

The end date and time for maintenance and support. Defaults to expiration date if not present.

Max Nodes

The maximum number of nodes that can be created for synchronization.

Max Tables

The maximum number of table triggers that can be created for synchronization.

Max Rows

The maximum number of rows that can be synchronized in a 24-hour period.

Instance ID

If present in the license key, this unique identifier must match the installation. If not present, the instance ID for the current installation is shown in parentheses.

The license key is a block of encoded text that is sent to the user over email, or it can be retrieved from the Support Center. When an installed license key expires, the web console will require a new valid key before allowing access, but data synchronization and command line access will continue to function normally.

license key

4.23. Mail Server

A mail server can be configured for sending email notifications.

Target Nodes

The node group ID that will use this configuration.

Hostname

The hostname or IP address of the mail server to contact for sending mail.

Transport

The transport mechanism is either SMTP (Simple Mail Transfer Protocol) or SMTPS (encrypted with SSL).

Port

The default port for SMTP is 25, while the default port for SMTPS is 465.

Use StartTLS

After connecting over SMTP, the TLS protocol is used to encrypt content.

Use Authentication

The mail server requires a login and password before email can be sent.

Use SSL Authentication

The mail server requires SSL authentication (Gmail, AWS).

User

The user login to use for authentication.

Password

The login password to use for authentication.

The "Test" button will use the current settings on the screen to connect to the mail server with the configured transport and report any errors.

4.24. Monitors

A monitor watches some part of the system for a problem, checking to see if the monitored value exceeds a threshold. (To be notified immediately of new monitor events, configure a notification.)

Monitor ID

The monitor ID is a unique name to refer to the monitor.

Target Nodes

The group of nodes that will run this monitor.

Monitor Type

The monitor type is one of several built-in or custom types that run a specific check and return a numeric value that can be compared to a threshold value.

Type Description Insight

cpu

Percentage from 0 to 100 of CPU usage for the server process.

disk

Percentage from 0 to 100 of disk usage (tmp folder staging area) available to the server process.

memory

Percentage from 0 to 100 of memory usage (tenured heap pool) available to the server process.

batchError

Number of incoming and outgoing batches in error.

batchUnsent

Number of outgoing batches waiting to be sent to online nodes.

batchUnsentOffline

Number of outgoing batches waiting to be sent to offline nodes.

dataUnrouted

Number of change capture rows that are waiting to be batched and sent.

dataGaps

Number of active data gaps that are being checked during routing for data to commit.

offlineNodes

The number of nodes that are offline based on the last heartbeat time. The console.report.as.offline.minutes parameter controls how many minutes before a node is considered offline.

log

Number of entries found in the log for the specified severity level.

loadAverage

Sum of the number of runnable entities queued to the available processors and the number of runnable entities running on the available processors averaged over a period of time. Not implemented for Windows.

fileHandles

Percentage from 0 to 100 of Operating System’s open file handles. Not implemented for Windows.

jobError

Number of jobs that are in error. This only applies to jobs that record statistics in the NODE_HOST_JOB_STATS table. The built-in jobs that write to this table are Routing, Purge Outgoing, Purge Incoming, Push, Pull, and SyncTriggers.

licenseExpire

Percentage from 0 to 100 of the license usage, with expiration occurring at 100%.

certExpire

Percentage from 0 to 100 of the TLS/SSL certificate usage, with expiration occurring at 100%.

licenseRows

Percentage from 0 to 100 of rows used out of the maximum number of rows allowed by the license.

jvm64Bit

Value of 0 or 1 indicating whether or not the operating system is 64-bit and the JVM is 32-bit.

jvmCrash

Number of Java crash files found that were created or modified in the last 24 hours.

jvmOutOfMemory

Number of times a java.lang.OutOfMemoryError appears in the wrapper.log file.

jvmThreads

Number of threads that are blocked or calling the same method.

block

Number of seconds that a transaction has been blocked for.

mySqlMode

Value of 0 or 1 indicating whether or not a MySQL node is incompatible with one or more other nodes.

nextDataInGap

Value of 0 or 1 indicating whether the next data ID is within a data gap.

channelsDisabled

Number of channels that are disabled.

maxBatchSize

Largest Max Batch Size for a channel.

maxDataToRoute

Largest Max Data to Route for a channel.

maxBatchToSend

Smallest Max Batch to Send for a channel.

maxChannels

Number of channels.

channelSuspend

Number of channels that are suspended or ignored.

missingPrimaryKey

Number of tables that are configured for replication and missing a primary key.

channelsForeignKey

Number of tables that are configured to use a different channel than other tables that they have a foreign key relationship with.

jobTrending

Largest percentage by which a job’s most recent runtime has exceeded its previous peak runtime. This only applies to jobs that record statistics in the NODE_HOST_JOB_STATS table (see jobError monitor above). This will only take effect for a particular job when the NODE_HOST_JOB_STATS table contains a sufficient amount of statistics for that job (10 or 100 rows, depending on the job).

connectionPool

Value of 0 or 1 indicating whether all connections in the database connection pool are in use.

connectionReset

Value of 0 or 1 indicating whether or not a "connection reset" error has occurred recently. The duration in the past during which this insight checks for "connection reset" errors is determined by the purge.log.summary.retention.minutes parameter, which defaults to 60 minutes.

lob

Number of tables that could store rows that are wider than the limit for data capture and have not been configured to account for these wide rows.

strandedOrExpiredData

Number of stranded or expired data rows detected by the Purge Outgoing Job and recorded in the NODE_HOST_STATS table within the past 4 hours.

unknownCa

Number of nodes that have recently experienced a "PKIX path building failed" error due to an unknown certificate authority. The duration in the past during which this insight checks for errors is determined by the purge.log.summary.retention.minutes parameter, which defaults to 60 minutes.

Expression

An expression used by the monitor to set options specific to the monitor type. For batchError monitors, setting the expression to "notifyOnIncrease=true" will trigger an additional notification if the number of batches in error has increased since the last time the monitor job ran.

Threshold

When this threshold value is reached or exceeded, an event is recorded.

Run Period

The time in seconds of how often to run this monitor. The monitor job runs on a period also, so the monitor can only run as often as the monitor job.

Run Count

The number of times to run the monitor before calculating an average value to compare against the threshold.

Severity Level

The importance of this monitor event when it exceeds the threshold.

Display Order

The order in which this monitor will be displayed in the web console.

Is Insight

Whether or not this monitor is an insight. Insights are optional recommendations for changing the system settings, while other monitors are for errors that must be resolved. Insights do not trigger notifications and they are displayed via the Insight Manager and the Insights Dialog rather than the Manage Monitors screen. This option is only available for some monitor types.

Is Pinned

Whether or not this monitor is pinned in the web console. A pinned monitor will be displayed to the user even if there are no unresolved events for it.

Enabled

Whether or not this monitor is enabled to run.

4.25. Notifications

A notification sends a message to the user when a monitor event records a system problem. First configure a monitor to watch the system and record events with a specific severity level. Then, configure a notification to match the severity level and write to the log or send an email.

Notification ID

The notification ID is a unique name to refer to the notification.

Target Nodes

The group of nodes that will run this monitor.

Notification Type

The notification type is either a built-in or custom type that is given the list of monitor events to send.

Type Description

log

The monitor events are written to the log using the same severity level. The web console will indicate WARN and ERROR level notifications in the top-right corner, which are also displayed on the main Dashboard screen.

email

The monitor events are sent in an email to a list of recipients. Use the expression for the comma-separated list of email addresses. Use the Configure→Mail Server screen to configure a mail server to use for sending emails.

Expression

Additional information to configure the notification type.

Severity Level

Find monitor events that occur at this severity level or above.

Enabled

Whether or not this notification is enabled to run.

5. Manage

This section will talk about how to manage and monitor SymmetricDS.

5.1. Nodes

Nodes can be added and managed in the Nodes screen under the Manage tab.

manage nodes

5.1.1. Add Node

Multiple nodes can be hosted in a single SymmetricDS instance. SymmetricDS will start a node for each properties file it finds in the engines directory.

Add nodes to your synchronization scenario by using the Connect Database Wizard. This wizard is automatically shown when less the 2 nodes are configured. It can also be accessed from the Manage Nodes screen by clicking the Add button.

wizard connect menu
Choosing node location
wizard connect second

SymmetricDS uses HTTP protocol to send data between nodes and uses JDBC to connect to databases. HTTP is designed to for sending across networks and will perform better than JDBC connections. As a result the maximum performance will be achieved the closer each SymmetricDS node is to its data source. This will provide the best JDBC performance to read and write data from the database and utilizes the HTTP protocol to maximize the transfer of data across a network.

HTTP can also be converted to use HTTPS so that the information is encrypted. JDBC drivers usually do not supply the ability to encrypt the data. This is a big plus for setting up your nodes remotely and taking full advantage of HTTPS to be sure your data is safe

Remote Nodes
wizard connect remote

Remote nodes require that another installation of SymmetricDS be installed on a remote machine. The registration url shown on this screen will be used by the remote node during its installation process in order to properly connect to the SymmetricDS network.

Auto registration by default is turned off. This ensures a person manually allows remote nodes to connect. By clicking this button registration will occur automatically and not require manual intervention.

Local Nodes
Database Setup
wizard connect database 2

Provide database credentials to connect the node to a database.

After selecting a database type the jdbc url will be populated for you based on the driver. Be sure to replace host name, port, and database name if necessary (varies by type).

Registration URL
wizard connect reg url

Provide the registration url in order to register the node on the SymmetricDS network.

This url is the sync url of the Master Node.

Node Group
wizard connect group

The list of groups will be provided from the Master Node.

A new group can be typed in though if the desired group is not present. Note, that the new group will need to be linked and associated to triggers and routers before synchronization will occur.

External ID
wizard connect external

This is a unique identifier for the node.

For example, in retail examples this might be a register number when the node is the store.

Communication Settings
wizard connect host 2

If your network will be accessing the web console through a proxy or a different host name that the one shown in the first bullet you can override the hostname to be used.

You can also select HTTPS or HTTP to be used for transferring data.

Finish
wizard connect install

The node is now ready to be installed.

5.1.2. Load Data

A load is the process of seeding tables at a target node with data from a source node. Instead of capturing data, data is selected from the source table using a SQL statement and then it is streamed to the client.

As of 3.8 loading data has been consolidated into a single wizard. Request a load by clicking the Load Data button on the manage nodes screen.

manage load data

This will open a new window that will walk through the steps screen by screen to build the load needed. If there were no selected rows on the manage nodes screen when this button was clicked the first screen will prompt for a source and target set of nodes. Multiple selections can be made here as long as all source nodes or all target nodes belong to the same group.

manage load data source target

The next screen allows for a selection of all tables configured (full load) or a subset of tables configured (partial load). Note, this will be the first screen provided if rows were already selected on the manage nodes screen when the load data button was selected.

The second option on this screen is to determine what "before action" should occur before the load is run.

Table 21. Before Load Options

Nothing just send the data

This assumes the tables are present and not need cleared out before the load can be run. If data already exists it will fall back to an update if the insert fails. This allows the load to be run repeatedly as needed.

Create tables

This will send create table statements to the target to match the source before loading any data. Use this option when the target database does not already contain the tables used on the source.

Delete all data

This will delete all data from the target table prior to loading data. This can be useful to accommodate data that is in the target but no longer in the source. By default just sending the data will only insert/update with the rows from the source so any old data will remain in place on the target unless this delete action is taken prior to loading.

Truncate tables

This will perform a truncate on all target tables prior to loading.

Run custom SQL

Allows a custom SQL statement to be executed on each table prior to loading. The use of %s will server as a variable to be replaced at runtime with the table name. Example: truncate table %s

manage load data full partial before

Partial loads will then see a table selection screen. Full loads will immediately be taken to the summary screen. The table selection screen allows individual tables to be selected as part of the load.

manage load data tables

There is also the ability to provide a where clause to be used to extract the data for a load. This allows control over what data should be loaded for a given table.

Where clause variable replacements
  • $(groupId)

  • $(nodeId)

  • $(externalId)

See Variables.

manage load data where clauses

Finally a summary screen is presented to review all the settings for the load prior to adding it to the request queue. Loads are checked by the routing process so once the load has been saved it will be picked up and begin processing on the next run of the routing job.

manage load data summary

A load’s progress can be monitored on the Outgoing Loads screen or the Outgoing Batches screen.

5.1.3. Control

Stopping a Node

Installed nodes are started automatically when the SymmetricDS server is started. An individual node instance can be stopped while other nodes continue to run.

To stop a node, select the node you want to stop and click on the Control button and choose Stop. The node’s status will indicate that is has been stopped.

From the command line, you can use JMX to stop a node. The following is an example. You would replace <engine name> with the name of the engine as found in the Node Properties File

bin/jmx --bean org.jumpmind.symmetric.<engine name>:name=Node --method stop

Uninstalling a Node

Uninstalling a node will remove all SymmetricDS database artifacts and delete the engine’s property file.

This can not be undone so be sure to uninstall with caution.

To uninstall a node, select the node you want to uninstall and click on the Control button and choose Uninstall.

uninstall
Figure 18. Control→Uninstall

If the node has no children, you will be prompted by a confirm dialog to make sure you want to uninstall the node.

uninstall single
Figure 19. Uninstall single node

If the node has child nodes you will be told that uninstalling the parent node will uninstall all of the child nodes.

uninstall with children
Figure 20. Uninstall current node and all children

From the command line you can use the symadmin utility to uninstall a node.

bin/symadmin --engine <engine name> uninstall

Changing a Node’s Priority

Setting a node’s priority to high means that the node will always be able to reserve a connection with the current node regardless of how busy it is. The priority settings will last until SymmetricDS restarts.

Setting too many nodes to high priority can starve the node of database connections.

To change a node’s priority from normal to high or vice-versa, select the node and click on the Control button and choose Toggle Priority.

From the command line, you can use JMX to change a node’s priority. Below are examples of how to increase priority, decrease priority, and get the list of prioritized node IDs.

bin/jmx --engine <engine name> --bean org.jumpmind.symmetric.<engine name>:name=Node --method addNodeToWhiteList --args "<node ID to prioritize>"

bin/jmx --engine <engine name> --bean org.jumpmind.symmetric.<engine name>:name=Node --method removeNodeFromWhiteList --args "<node ID to deprioritize>"

bin/jmx --engine <engine name> --bean org.jumpmind.symmetric.<engine name>:name=Node --method getNodesInWhiteList

5.1.4. Registration

In order for a node to synchronize with other nodes it must be registered. When a node is registered, it downloads its configuration as well as references to the nodes that it should sync with.

To allow a node to register, navigate to Manage Nodes on the server, click the Registration button and select Allow. Registration requests appear at the top of the list with a question mark (?) for the Node ID.

manage nodes allow registration

The registration menu has the following actions:

Allow

Accept the request and register the node. The node is assigned a node ID and password, which are sent along with the configuration. If the auto.reload or auto.reload.reverse parameters are true, an initial load is started.

Reject

Reject the request by removing it. The node can still make additional registration requests.

Reopen

Open registration for a node that is already registered, which forces the node to use the registration.url and pull configuration again. The auto.reload and auto.reload.reverse parameters do not apply to reopen. If the reopen.use.same.password parameter is false, a new node password is generated.

Unregister

Remove the node from any further data synchronization by removing registration, deleting any incoming batches, and marking any outgoing batches with status OK.

You can remove a node from the command line with the following command:

bin/symadmin remove-node --engine <engine name> -n <node id>

A node is considered unregistered if it does not have an NODE_IDENTITY row. When a node is unregistered, it will use the registration.url defined in the Node Properties File to request registration. The registration.url of the new node is the sync.url of the node that is being registered with.

Before a node is allowed to register, it must have an open registration. If there is no open registration, then a REGISTRATION_REQUEST is recorded.

You can open registration from the command line with the following command:

bin/symadmin open-registration --engine <engine name> <node group> <external id>

The <node group> and <external id> should match the group.id and external.id in the registering node’s Node Properties File.

Node registration is stored in the NODE and NODE_SECURITY tables. Nodes are only allowed to register if rows exist for the registering node and the registration_enabled flag is set to 1.

If the auto.registration SymmetricDS parameter is set to true, then when a node attempts to register, the node will automatically be accepted.

SymmetricDS allows you to have multiple nodes with the same external_id. In order to enable this you must set external.id.is.unique.enabled to false.

5.1.5. Push Registration

If the client is unable to pull from the server, it can register by having the server push to it instead. The group link for server to client should be set to push.

From the Manage Nodes screen on the server, click on the Add button and then select Remote.

manage nodes push registration

You can open registration from the command line with the following command:

bin/symadmin open-registration --engine <engine name> <node group> <external id> <sync url>

The <node group>, <external id>, and <sync url> should match the group.id, external.id, and sync.url in the registering node’s Node Properties File.

5.1.6. Initial Loads

Loading data for 3.8 and above has been modified, see Load Data.

Initial Load Extract In Background

By default, initial loads for a table are broken into multiple batches, with the size of batches based on the max_batch_size of the Channels for the reload channel being used. Batches are pre-extracted to staging in the background, instead of waiting for a push or pull to extract them. An estimated count of rows for the table are queried from the database statistics, or it will execute a count(*) query from the table if statistics are not available. The extract process creates batches based on the number of rows in the table divided by the max_batch_size.

If the background job is disabled by setting initial.load.use.extract.job.enabled to false, then all data for a given table will be extracted into a single batch during a push or pull, regardless of channel settings. For large tables, this can result in a batch that can take a long time to extract and load.

Initial Load Parameters

There are several parameters that can be used to modify the behavior of an initial load.

auto.reload

A load is queued up to a node automatically when it registers.

auto.reload.reverse

A reverse initial load is queued up for a node automatically when it registers.

initial.load.delete.first / initial.load.delete.first.sql

By default, an initial load will not delete existing rows from a target table before loading the data. If a delete is desired, the parameter initial.load.delete.first can be set to true. If true, the command found in initial.load.delete.first.sql will be run on each table prior to loading the data. The default value for initial.load.delete.first.sql is

delete from %s

Note that individual reload batches are created that represent the deletes in the reverse order that load batches are created. All delete batches are inserted first. The initial.load.delete.first.sql can be overwritten at the TRIGGER_ROUTER level by entering a initial_load_delete_stmt.

initial.load.create.first

By default, an initial load will not create the table on the target if it doesn’t already exist. If the desired behavior is to create the table on the target if it is not present, set the parameter intial.load.create.first to true. SymmetricDS will attempt to create the table and indexes on the target database before doing the initial load. Note that individual create batches are created to represent each of the table creates.

Sometimes when creating tables across different database platforms default values do not translate. You can turn off the use of default values during the table create by setting create.table.without.defaults.

5.1.7. Send

Events other than data changes can be sent to nodes, which include the following:

SQL Scripts

Sql can be sent to be executed on a target node

BSH Scripts

Beanshell scripts can be sent to be executed on a target node

Table Schema

The schema the source node can be replicated to the target node individually

Configuration

The configuration is sent, similar to exporting the configuration and importing it, which excludes node, user, role, and privilege tables.

5.2. Jobs

Most work done by SymmetricDS is initiated by jobs. Jobs are tasks that are started and scheduled by a job manager. Jobs are enabled by the start.<name>.job parameter.

Most jobs are enabled by default. The frequency at which a job runs is controlled by one of two parameters:

  • job.<name>.period.time.ms

  • job.<name>.cron

If a valid cron property exists in the configuration, then it will be used to schedule the job. Otherwise, the job manager will attempt to use the period.time.ms property.

From the manage jobs screen the frequency that a job will run for the current node in the cluster can be managed. If you want to change the frequency a job runs across a group or for everyone, you modify the parameter in the Parameters screen.

A job can also be stopped and restarted. If a job is scheduled to run infrequently and the job is needed to run immediately you can select job and click Run Now.

manage jobs

SymmetricDS utilizes Spring’s CRON support, which includes seconds as the first parameter. This differs from the typical Unix-based implementation, where the first parameter is usually minutes. For example, */15 * * * * * means every 15 seconds, not every 15 minutes. See Spring’s documentation for more details.

5.2.1. Route Job

The Route Job is responsible for creating outgoing batches of captured data that are targeted at specific nodes.

The job processes Channels, one at a time, reading up to Max Data To Route data rows which have not been routed.

The data is assigned to outgoing batches based on the Batch Algorithm defined for the channel. Note that, for the default and transactional algorithm Max Data To Route rows may be exceeded depending on the transaction boundaries.

An outgoing batch is initially created with a status of "RT". Data is assigned to the batch by inserting into data event. When a batch is complete, the batch is committed and the status is changed to "NE".

The route job will respect the Max Batch Size as configured in Channels. If the max batch size is reached before the end of a captured database transaction and the batch algorithm is set to something other than nontransactional the batch may exceed the specified max size.

The route job delegates to a router to decide which nodes need to receive the data. The correct router is looked up by referencing the captured trigger_hist_id in the DATA table and using Table Routing configuration.

After outgoing batches have been created by the Route Job, they need to be transported to the target node.

Data Gaps

The DATA to route is selected based on the values in the DATA_GAP table. For efficiency, DATA_GAP tracks gaps in the data ids in DATA table that have not yet been processed.

A gap while routing in DATA can occur because concurrently running transactions have not yet committed. They can also be caused by rolled back transactions.

Most of gaps are only temporarily and fill in at some point after routing and need to be picked up with the next routing run.

This table completely defines the entire range of data that can be routed at any point in time. For a brand new instance of SymmetricDS, this table is empty and SymmetricDS creates a gap starting from data id of zero and ending with a very large number (defined by routing.largest.gap.size ).

At the start of a route job, the list of valid gaps (gaps with status of 'GP') is collected, and each gap is evaluated in turn. If a gap is sufficiently old (as defined by routing.stale.dataid.gap.time.ms, SymmetricDS assumes that a transaction has been rolled back and deletes the gap.

If the gap is not stale, then DATA_EVENT is searched for data ids present in the gap. If one or more data ids is found in DATA_EVENT, then the current gap is deleted, and new gap(s) are created to represent the data ids still missing in the gap’s range. This process is done for all gaps. If the very last gap contained data, a new gap starting from the highest data id and ending at (highest data id + routing.largest.gap.size) is then created.

This results in an updated gap list that can be used to select DATA for routing.

5.2.2. Push Job

The Push Job is responsible for assigning nodes that need to be pushed to individual threads. See Push Threads for more details.

The job sends Outgoing Batches to the target node using an HTTP PUT. By default an HTTP PUT buffers data at the client. If large batches are going to be sent using the push job, then consider turning on http.push.stream.output.enabled.

The push job is considered to be slightly more efficient than the Pull Job because it only needs to make a network connection if there are batches available to send.

In order to be more efficient, the push job sends an HTTP HEAD to request a reservation at the target node. If the target nodes responds and accepts the request, then the job issues the HTTP PUT with the data pay load in Data Format

5.2.3. Pull Job

The Pull Job is responsible for assigning nodes that need to be pulled to individual threads. See Pull Threads for more details.

The job expects to receive Incoming Batches from a source node using an HTTP GET.

5.2.4. Purge Outgoing Job

The Purge Outgoing Job is responsible to purging outgoing data that has successfully been loaded at the target and is older than purge.retention.minutes.

This job purges the following tables:

5.2.5. Purge Incoming Job

The Purge Incoming Job is responsible for purging the INCOMING_BATCH table.

5.2.6. Statistics Job

The Statistics Job flushes captured statistics to following tables:

It also purges the same tables based on the purge.stats.retention.minutes parameter.

5.2.7. Sync Triggers Job

The Sync Triggers Job runs when a node is started and on the prescribed job schedule. The job checks for missing SymmetricDS database triggers and creates them. It also updates the SymmetricDS database triggers that have had a change to its configuration or the database table has had a change to its structure.

5.2.8. Heartbeat Job

The Heartbeat Job updates its own NODE_HOST row with a new heartbeat_time so that it is synchronized to it’s created_at_node_id node to indicate that the node is online and healthy.

5.2.9. Watchdog Job

The Watchdog Job looks for nodes that have been offline for offline.node.detection.period.minutes and disables them.

5.2.10. Stage Management Job

The Stage Management Job purges the staging area according to the stream.to.file.ttl.ms parameter.

5.2.11. Refresh Cache Job

The Refresh Cache Job checks the last_update_time on each cached configuration resource and determines if it needs to refresh the cached items. This job is mostly relevant when using Clustering.

5.2.12. File Sync Tracker Job

The File System Tracker job is responsible for monitoring and recording the events of files being created, modified, or deleted. It records the current state of files to the FILE_SNAPSHOT table.

See File Synchronization for more details.

5.2.13. File Sync Pull Job

The File Sync Pull Job is responsible for assigning nodes that need to be pulled to individual threads.

See File Synchronization and Pull Threads for more details.

5.2.14. File Sync Push Job

The File Sync Push Job is responsible for assigning nodes that need to be pushed to individual threads.

See File Synchronization and Pull Threads for more details.

5.2.15. Initial Load Extract Job

The Initial Load Extract Job processes EXTRACT_REQUESTs. See Initial Load Extract In Background for more details.

5.3. Installed Triggers

SymmetricDS installs database triggers to capture changes in the DATA table. A record of the triggers that were installed and what columns are being captured is stored in the TRIGGER_HIST table. When data is captured in DATA it references the TRIGGER_HIST record that represented the trigger at the time data was captured.

This is necessary because if a trigger is rebuilt after columns are added or removed and data that was captured by the old trigger has not yet been delivered, we need a record of what columns were in play at the time the data had been captured.

The Installed Triggers Screen is used to view tables that have successfully been installed with triggers.

This screen allows you to Rebuild an individual database trigger or Rebuild All database triggers. It also allows you to Drop triggers. Only reason you might need to drop triggers to alter a table or perform a batch job where you don’t want data to be captured.

manage installed triggers

The TRIGGER_HIST table records the reason a trigger was rebuilt. The following reasons are possible:

N

New trigger that has not been created before

S

Schema changes in the table were detected

C

Configuration changes in Trigger

T

Trigger was missing

A configuration entry in Trigger without any history in Trigger Hist results in a new trigger being created (N). The Trigger Hist stores a hash of the underlying table, so any alteration to the table causes the trigger to be rebuilt (S). When the last_update_time is changed on the Trigger entry, the configuration change causes the trigger to be rebuilt ©. If an entry in Trigger Hist is missing the corresponding database trigger, the trigger is created (T).

5.4. Outgoing Loads

The Outgoing Loads screen shows the number of outgoing loads that have been queued. It also lists loads that have had Outgoing Batches created. The screen shows loads that are sourced from the current node.

You can cancel a load that is in progress by selecting the load and pressing the Cancel button.

manage outgoing loads

5.5. Incoming Loads

The Incoming Loads screen shows the number of incoming loads that have been queued. It also lists loads that have had Incoming Batches created. The screen shows loads that are targeted at the current node.

You can cancel a load that is in progress by selecting the load and pressing the Cancel button.

manage incoming loads

5.6. Outgoing Batches

The Outgoing Batches screens allows you to monitor data that has been routed.

manage outgoing batches

Outgoing batches are delivered to the target node when the source node pushes or when the target node pulls.

A single push or pull connection is called a synchronization.

For one synchronization, each enabled channel will be processed. Channels are processed in the order defined by the Processing Order setting on the channel with two exceptions:

  • If there are reload channels available to be sent and the reload channels are not in error, then only reload channels will be sent

  • If a channel is in error it will be moved to the bottom of the list

When outgoing batches are selected for a node and a channel, the maximum number of batches that are extracted per synchronization is controlled by the Max Batch To Send setting on the channel.

There is also a setting that controls the max number of bytes to send in one synchronization. If SymmetricDS has extracted more than the number of bytes configured by the transport.max.bytes.to.sync parameter, then it will finish extracting the current batch and then finish synchronization so the client has a chance to process and acknowledge the "big" batch. This may happen before the configured Max Batch To Send has been reached.

When extracting a batch, data is first extracted to the Staging Area and then sent across the network from the Staging Area. The staging area is used to minimize the amount of time a database connection is being used when streaming over slower networks. The use of the staging area can be turned off by setting the stream.to.file.enabled parameter.

5.6.1. Extract Frequency By Channel

The pull and push frequency cannot be adjusted by channel. If you want to adjust the frequency that outgoing batches for a specific channel are sent, you have two options:

  1. Batches are extracted by channel at an interval controlled by the extract_period_millis in the Channels settings. The last_extract_time is always recorded, by channel, on the NODE_CHANNEL_CTL table for the host node’s id. When the Pull and Push Job run, if the extract period has not passed according to the last extract time, then the channel will be skipped for this run. If the extract_period_millis is set to zero, data extraction will happen every time the jobs run.

  2. SymmetricDS provides the ability to configure windows of time when synchronization is allowed. This is done using the NODE_GROUP_CHANNEL_WND table. A list of allowed time windows can be specified for a node group and a channel. If one or more windows exist, then data will only be extracted and transported if the time of day falls within the window of time specified. The configured times are always for the target node’s local time. If the start_time is greater than the end_time, then the window crosses over to the next day.

5.6.2. Outgoing Sync Status

The status of outgoing synchronization can be queried at the source database.

The following query will show outgoing synchronization failures by node:

select count(*), node_id from sym_outgoing_batch
  where error_flag=1 group by node_id;

The following query will show the number of data rows that have not been delivered to target nodes:

select sum(data_row_count), node_id from sym_outgoing_batch
  where status != 'OK' group by node_id;

The following queries summed together give an approximation of the number of rows that have not been routed:

select sum(end_id-start_id) from sym_data_gap
  where start_id < (select max(start_id) from sym_data_gap);

select count(*) from sym_data
  where data_id >= (select max(start_id) from sym_data_gap);

5.6.3. Outgoing Batch Errors

By design, whenever SymmetricDS encounters an issue with synchronization, the batch containing the error is marked as being in an error state, and all subsequent batches on the same channel to the same node are not synchronized until the batch error is resolved.

SymmetricDS will retry the batch in error until the situation creating the error is resolved (or the data for the batch itself is changed). If the error is caused by network or database failures, then the error might eventually resolve itself when the network or database failures are resolved.

Analyzing and resolving issues can take place on the outgoing or incoming side. The techniques for analysis are slightly different in the two cases, however, due to the fact that the node with outgoing batch data also has the data and data events associated with the batch in the database. On the incoming node, however, all that is available is the incoming batch header and data present in an incoming error table.

Analyzing the Issue

The first step in analyzing the cause of a failed batch is to locate information about the data in the batch.

To show only batches in error, select the status of Error:

manage outgoing batches error

In order to view the error message, click the bar chart link in the details column of the outgoing batch row of interest.

manage outgoing batches error show

In order to view the data row in error, click the Sent Data tab on the captured batch data screen. The row that is in error will be red. You can view the data that changed at the bottom of the dialog.

manage outgoing batches error data

The error message and the data should give the user clues as to why the batch failed.

Common reasons a batch might fail include:

  • The schema at the destination has a column that is not nullable yet the source has the column defined as nullable and a data change was sent with the column as null.

  • A foreign key constraint at the destination is preventing an insertion or update, which could be caused from data being deleted at the destination or the foreign key constraint is not in place at the source.

  • The data size of a column on the destination is smaller than the data size in the source, and data that is too large for the destination has been synced.

Resolving the Issue

Once you have decided upon the cause of the issue, you’ll have to decide the best course of action to fix the issue. If, for example, the problem is due to a database schema mismatch, one possible solution would be to alter the destination database in such a way that the SQL error no longer occurs. Whatever approach you take to remedy the issue, once you have made the change, on the next push or pull SymmetricDS will retry the batch and the channel’s data will start flowing again.

If you have instead decided that the batch itself is wrong, or does not need synchronized, or you wish to remove a particular data change from a batch, you do have the option of changing the data associated with the batch directly.

Be cautious when using the following two approaches to resolve synchronization issues. By far, the best approach to solving a synchronization error is to resolve what is truly causing the error at the destination database. Skipping a batch or removing a data id as discussed below should be your solution of last resort, since doing so results in differences between the source and destination databases.

Now that you’ve read the warning, if you still want to change the batch data itself, you do have several options, including:

  • Causing SymmetricDS to skip the batch completely can be accomplished from the user interface. Select the batch that should be ignored. Click Change and select Ignore Batch.

    manage outgoing batches error ignore
  • Removing the failing data id from the batch by deleting the corresponding row in DATA_EVENT. Eliminating the data id from the list of data ids in the batch will cause future synchronization attempts of the batch to no longer include that particular data change as part of the batch. For example:

    delete from sym_data_event where batch_id='XXXXXX' and data_id='YYYYYY'
    where XXXXXX is the failing batch and YYYYYY is the data id to longer be included in the batch.

    After modifying the batch you will have to clear the Staging Area manually or wait for the staged version of the batch to timeout and clear itself.

5.7. Incoming Batches

The Incoming Batches screens allows you to monitor batches that have been delivered to the target node.

manage incoming batches

Incoming batches are delivered to the target node when the source node pushes or when the target node pulls.

Incoming batches are written to the Staging Area first and then loaded. The use of the staging area can be turned off by setting the stream.to.file.enabled parameter.

5.7.1. Incoming Sync Status

The status of incoming synchronization can be queried at the target database.

The following query will show incoming synchronization failures by node:

select count(*), node_id from sym_incoming_batch
  where error_flag=1 group by node_id;

Client nodes update their heartbeat_time in the NODE_HOST table. If a client node is online and actively syncing you can monitor the NODE_HOST table at the server to find client nodes that are offline. Note that at times there could be more than one NODE_HOST row per node_id. This could be the case if the nodes are clustered or the host_name changes.

The following query will give you nodes that have not synchronized in the last 24 hours. Note that the SQL might vary slightly for some databases as some of the supported databases do not support current_timestamp.

select node_id, host_name from sym_node_host
  where heartbeat_time < current_timestamp-1;

5.7.2. Incoming Batch Errors

When a batch fails to load it is marked with an Error status.

Analyzing the Issue

Analysis using an incoming batch is different than that of outgoing batches.

Batches in Error can be viewed when the Status is set to <Not Ok> or Error.

manage incoming batches error

The error message can be viewed by clicking the Bar Chart link in the details column.

manage incoming batches error show
Resolving the Issue

For batches in error, from the incoming side you’ll also have to decide the best course of action to fix the issue.

Incoming batch errors that are in conflict can by fixed by taking advantage of two columns in INCOMING_ERROR which are examined each time batches are processed. The first column, resolve_data if filled in will be used in place of row_data. The second column, resolve_ignore if set will cause this particular data item to be ignored and batch processing to continue. This is the same two columns used when a manual conflict resolution strategy is chosen, as discussed in Conflicts.

5.8. Staging Area

SymmetricDS creates temporary extraction and data load files with the CSV payload of a synchronization when the value of the stream.to.file.threshold.bytes SymmetricDS property has been reached. Before reaching the threshold, files are streamed to/from memory. The default threshold value is 0 bytes. This feature may be turned off by setting the stream.to.file.enabled property to false.

SymmetricDS creates these temporary files in the directory specified by the java.io.tmpdir Java System property.

The location of the temporary directory may be changed by setting the Java System property passed into the Java program at startup. For example,

-Djava.io.tmpdir=/home/.symmetricds/tmp

The Staging Area screen shows files that are currently in the staging area. Staging area can be cleared at any time by clicking the Clear Staging button.

Clear Staging Performance Impacts

The staging area is designed to hold extracted data while it is waiting to be processed through a push or pull request. This will contain a mix of staged batch files that have been completed, files that are in process, and files that are waiting to be processed. So clearing this will remove all types of files both for both incoming and outgoing batches. If a file is not found in staging as a result of clearing it, the batch will need to be re-extracted. If an incoming batch is not found when its ready to be loaded a resent request will be made to the source node that it can be sent across the network again. So use caution when clearing staging if there are a lot of files present otherwise there could be an impact on performance while these batches are re-extracted or resent.

staging

5.9. Processes

The Processes screen shows what is actively going on in SymmetricDS for the current node.

processes

5.10. Pull Threads

Both the Pull Job and the File Sync Pull Job can be configured to pull multiple nodes in parallel. In order to take advantage of this the pull.thread.per.server.count or file.pull.thread.per.server.count should be adjusted (from the default value of 1) to the number to the number of concurrent pulls you want to occur per period on each SymmetricDS instance.

This screen shows pull activity for both table and file synchronization.

pull threads

Pull activity is recorded in the NODE_COMMUNICATION table. This table is also used as a semaphore to lock pull activity across multiple servers in a cluster.

5.11. Push Threads

Both the Push Job and the File Sync Push Job can be configured to push multiple nodes in parallel. In order to take advantage of this the push.thread.per.server.count or file.push.thread.per.server.count should be adjusted (from the default value of 1) to the number to the number of concurrent pushes you want to occur per period on each SymmetricDS instance.

This screen shows pull activity for both table and file synchronization.

push threads

Push activity is recorded in the NODE_COMMUNICATION table. This table is also used as a semaphore to lock push activity across multiple servers in a cluster.

5.12. JVM Properties

The JVM Properties screen is a convenience screen that shows JVM System Properties.

jvm properties

5.13. JVM Threads

The JVM Threads Screen shows active threads in the JVM.

jvm threads

5.14. Monitors

The Monitors screen allows you to view events of system problems recorded by both local and remote nodes. The list of events can be filtered to limit the number of events displayed, event type, severity level, and node ID. Filtering by severity level will match the level you choose and any level above it. Events are listed in descending order by event time, but the order can be changed by clicking column headings. The remove button will clear the event from the table on the current node. Events are purged automatically each night using the purge.retention.minutes parameter to remove by event time.

manage monitors

When a Monitor is configured, it is run periodically to check the current value of a system metric and compare it to a threshold value. Different monitor types can check the CPU usage, disk usage, memory usage, batch errors, outstanding batches, unrouted data, number of data gaps, and job errors. Custom monitor types can be created using Extensions that use the IMonitorType interface. When the value returned from the check meets or exceeds the threshold value, a MONITOR_EVENT is recorded. The MONITOR_EVENT table is synchronized on the "monitor" channel, which allows a central server to see events from remote nodes, but this behavior can be disabled by setting the monitor.events.capture.enabled parameter to false.

To be immediately notified of a monitor event, use Notifications to match on the severity level. Different notification type can send a message by writing to the log or sending an email. Custom notification types can be created using Extensions that use the INotificationType interface. In order to send email, the Mail Server should be configured.

5.15. Security

The Security panel manages the SSL/TLS certificates and keys. It also manages break glass users.

5.15.1. Server Certificate

The server certificate and private key are used for encrypting the HTTPS transport of data. The certificate can also be used to verify a node when client authentication is enabled. The server certificate is stored in a protected keystore named security/keystore.

Import certificate and key from file

Upload a file containing the server certificate and private key, which can then be imported. If the file is a keystore with multiple entries, the user is prompted to select one from the list. If the file is a protected keystore, the user is prompted for the password. File types supported are PEM, CRT, CER, PFX, and P12, with support for PKCS 1, PKCS 8, and PKCS 12 standards.

Backup certificate key to file

Backup the current server certificate and private key to a PEM-encoded text file.

Export certificate to file

Export the current certificate to a PEM-encoded text file.

Generate self-signed certificate

Generate a new key and self-signed certificate.

5.15.2. Certificate Authorities

The certificates for authorities are public keys used for verifying authorized certificates used for encryption and authentication. Certificates are stored in a trusted keystore named security/cacerts.

View

View one certicate to display its subject, issuer, effective date, expiration date, and key type.

Delete

Delete one or more certificates from the keystore.

Export

Export one public certificate to a file.

Import

Import one public certificate from a file.

5.15.3. Client Authentication

Authentication of clients using certificates is enabled in the conf/symmetric-server.properties file.

https.need.client.auth=true

The client will be required to send a valid certificate before HTTPS requests can be made. During TLS negotiation, the server will send a list of certificate authorities to the client (See Certificate Authorities). If the client has a certificate signed by one of the authorities, it will send it to the server as authentication (See Server Certificate).

5.15.4. Console Client Authentication

Authentication of console users using client certificates can be enabled with parameter console.auth.client.cert. The console username is parsed from the distinguished name of the certificate’s subject using the first matching group of a regular expression specified by parameter console.auth.client.cert.subject.regex. By default, it will match on the common name of the subject.

5.15.5. Break Glass

If the user has the necessary permissions, they can break the glass for another user that has the is_glass_break_required flag set. This will generate a new password for the user and allow them to login for the selected duration. It is also possible to expire the glass break before the time period is over.

5.16. Rest Keys

Rest API keys must be set up in order to access the REST API. Attempting to access the REST API without a key will result in HTTP error code 659 (forbidden). When a new key is created, it will only be displayed once, so be sure to copy it and save it for future use.

The following is an example of how to include the key in a curl command:

curl -X GET "http://localhost:31415/api/engine/corp-000/channelstatus" -H "accept: */*" -H "X-REST-API-KEY: gSrLaAEuKa3FLdUbuUZ1cRv7Mi2nVS"

A user interface for the REST API is available on the Help → REST API screen. When the screen is first visited, the padlock icons are open to indicate that a REST API key is not being used. To enter the key, click the Authorize button, which is highlighted in red in the below image. After entering the key, the padlocks will close.

rest api
Required Fields
Key Name

Name of application that will use this key.

Key Description

Additional description of the use.

Access Level

Level of access, either read-only or read-write.

Readonly

Read-only users can only access the get methods.

Readwrite

Read-write users can access all methods, including post, put, and get.

5.17. Logging

The standalone SymmetricDS installation uses Log4J for logging. The configuration file is conf/log4j.xml. The log4j.xml file has hints as to what logging can be enabled for useful, finer-grained logging.

There is a command line option to turn on preconfigured debugging levels. When the --debug option is used the conf/debug-log4j.xml is used instead of log4j.xml.

SymmetricDS proxies all of its logging through SLF4J. When deploying to an application server or if Log4J is not being leveraged, then the general rules for SLF4J logging apply.

You can use the logging screen to dynamically adjust the log level in order to help triage issues. When developing transforms it is useful to turn on Log Sync Details.

logging

6. Advanced Topics

This chapter focuses on a variety of topics, including deployment options, jobs, clustering, encryptions, synchronization control, and configuration of SymmetricDS.

6.1. Advanced Synchronization

6.1.1. Disabling Synchronization

All data loading may be disabled by setting the dataloader.enable property to false. This has the effect of not allowing incoming synchronizations, while allowing outgoing synchronizations. All data extractions may be disabled by setting the dataextractor.enable property to false. These properties can be controlled by inserting into the master node’s PARAMETER table. These properties affect every channel with the exception of the 'config' channel.

6.1.2. Bi-Directional Synchronization

SymmetricDS allows tables to be synchronized bi-directionally. Note that an outgoing synchronization does not process changes during an incoming synchronization on the same node unless the trigger was created with the sync_on_incoming_batch flag set. If the sync_on_incoming_batch flag is set, then update loops are prevented by a feature that is available in most database dialects. More specifically, during an incoming synchronization the source node_id is put into a database session variable that is available to the database trigger. Data events are not generated if the target node_id on an outgoing synchronization is equal to the source node_id.

By default, only the columns that changed will be updated in the target system.

Conflict resolution strategies can be configured for specific links and/or sets of tables.

6.1.3. Multi-Tiered Synchronization

There may be scenarios where data needs to flow through multiple tiers of nodes that are organized in a tree-like network with each tier requiring a different subset of data. For example, you may have a system where the lowest tier may be a computer or device located in a store. Those devices may connect to a server located physically at that store. Then the store server may communicate with a corporate server for example. In this case, the three tiers would be device, store, and corporate. Each tier is typically represented by a node group. Each node in the tier would belong to the node group representing that tier.

A node can only pull and push data to other nodes that are represented in the node’s NODE table and in cases where that node’s sync_enabled column is set to 1. Because of this, a tree-like hierarchy of nodes can be created by having only a subset of nodes belonging to the same node group represented at the different branches of the tree.

If auto registration is turned off, then this setup must occur manually by opening registration for the desired nodes at the desired parent node and by configuring each node’s registration.url to be the parent node’s URL. The parent node is always tracked by the setting of the parent’s node_id in the created_at_node_id column of the new node. When a node registers and downloads its configuration it is always provided the configuration for nodes that might register with the node itself based on the Node Group Links defined in the parent node.

Registration Redirect

When deploying a multi-tiered system it may be advantageous to have only one registration server, even though the parent node of a registering node could be any of a number of nodes in the system. In SymmetricDS the parent node is always the node that a child registers with. The REGISTRATION_REDIRECT table allows a single node, usually the root server in the network, to redirect registering nodes to their true parents. It does so based on a mapping found in the table of the external id (registrant_external_id) to the parent’s node id (registration_node_id).

For example, if it is desired to have a series of regional servers that workstations at retail stores get assigned to based on their external_id, the store number, then you might insert into REGISTRATION_REDIRECT the store number as the registrant_external_id and the node_id of the assigned region as the registration_node_id. When a workstation at the store registers, the root server sends an HTTP redirect to the sync_url of the node that matches the registration_node_id.

Please see Initial Loads for important details around initial loads and registration when using registration redirect.

6.2. Offline Synchronization

6.2.1. Setup an Offline Node

Configuring a node as offline will still allow changes to be captured and batched for replication. However the push and/or pull jobs that are used to interact with this node will not use the standard http or https protocols to communicate with other nodes for changes. Instead the local file system will be used for replication. It is up to the user transport batch (*.csv) files to and from the node based on incoming or outgoing changes.

  • Turn on the offline push and pull jobs.

offline start jobs
Figure 21. Configure the offline jobs to run
  • Turn on the offline.node parameter for the node that should be offline.

offline node parameter
Figure 22. Configure the offline parameter for the node that will be offline
  • Setting these parameters immediately affects the behavior of the push and pull jobs, so outgoing batches intended for the offline node are now written as files. Other nodes are unaffected and will continue to synchronize normally. All outstanding batches for this node are immediately written to files. As new changes occur going forward, any batches for this will also be written as files. In this example, two batches of data were waiting to sync, so they are written to files.

  • At the offline store node, the parameter immediately affects the behavior of the push and pull jobs, so outgoing batches intended for CORP are now written as files. Other nodes are unaffected. All outstanding batches for CORP are immediately written to files. As new changes occur going forward, any batches for CORP will also be written as files. In this example, two batches of data are written.

offline sync node1 before
  • Finally, move the batch files to their respective incoming folder. After moving the files, the folders will contain the files depicted below.

offline sync

6.2.2. Turn offline node online again

To configure the node online again simply remove the parameter entries from step 2 above. As an additional step to save resources the offline jobs can be stopped as well from step 1 above if there are not any nodes operating in an offline mode.

6.3. Encrypted Passwords

This section describes command line usage. The Manage → Startup Parameters screen in the web console automatically encrypts database properties.

The db.user and db.password properties can be protected by encrypting them. This encryption can be done two different ways.

The first option is to pass the text as an argument:

symadmin encrypt-text "text-to-encrypt"
The text-to-encrypt on the command line should be surrounded with double quotes if spaces or symbols are used.

The second option is without an argument, and it will prompt you to enter text:

symadmin encrypt-text
Enter Text:

Encrypted text starts with "enc:" to differentiate it from plain text. See the Keystores section for an explanation of the encryption key.

6.4. Secure Transport

By specifying the "https" protocol for a URL, SymmetricDS will communicate over Secure Sockets Layer (SSL) for an encrypted transport. The following properties need to be set with "https" in the URL:

sync.url

This is the URL of the current node, so if you want to force other nodes to communicate over SSL with this node, you specify "https" in the URL.

registration.url

This is the URL where the node will connect for registration when it first starts up. To protect the registration with SSL, you specify "https" in the URL.

6.4.1. Standalone

The SymmetricDS service and the "sym" launch command use Jetty as an embedded web server. Edit the conf/symmetric-server.properties file to change port numbers to listen on and whether or not to use HTTP and/or HTTPS.

http.enable=true
http.port=31415
https.enable=true
https.port=31417

6.4.2. Tomcat

If you deploy SymmetricDS to Apache Tomcat, it can be secured by editing the TOMCAT_HOME/conf/server.xml configuration file. There is already a line that can be uncommented and changed to the following:

<Connector port="8443" protocol="HTTP/1.1" SSLEnabled="true"
  maxThreads="150" scheme="https" secure="true"
  clientAuth="false" sslProtocol="TLS"
  keystoreFile="/symmetric-ds-1.x.x/security/keystore" />

6.4.3. Keystores

The security subdirectory contains a private keystore and a trusted keystore. System properties are used to specify the location of each keystore file and a password to protect it. The system properties are set in the bin/setenv (or bin\setenv.bat on Windows) and conf/sym_service.conf files.

Filename Store Type System Property for File System Property for Password Description

keystore

PKCS12

sym.keystore.file

javax.net.ssl.keyStorePassword

Contains private encryption key and TLS certificate.

cacerts

JKS

javax.net.ssl.trustStore

javax.net.ssl.trustStorePassword

Contains public certificates for trusted authorities who sign keys.

The following entries in the keystore file are used.

Table 22. Alias entries in keystore file

sym

The TLS certificate used for handling incoming HTTPS communication.

sym.secret

The encryption key used for protecting secrets like database password.

If an entry is missing when encryption is requested, SymmetricDS will automatically generate a random key for use. It tries to use the strongest encryption algorithm and the largest key size available on the system. If the keystore file is missing, it will be created. Starting in SymmetricDS 3.14, it uses PKCS12 as the store type for new keystores, but it is backwards compatible with the older JCEKS store type.

6.4.4. Finding Keystore Password

The keystores and each key entry is protected with a password. The default password is changeit. During installation, the setup program chooses a random password and obfuscates it for the private keystore file.

To obtain the current password, use the following steps:

  • Look in the bin/setenv (or bin\setenv.bat on Windows) or the conf/sym_service.conf files. (The password should be the same in both files.)

  • For the password to keystore, look for the javax.net.ssl.keyStorePassword system property.

  • For the password to cacerts, look for the javax.net.ssl.trustStorePassword system property.

  • If the password starts with "obf:" then it is obfuscated. To obtain the cleartext password, use the following command from the bin subdirectory:

symadmin unobfuscate-text obf:cHVuYXRydmc=
  • If the password does not start with "obf:" then it is the cleartext password.

6.4.5. Changing Keystore Password

To change the keystore password, use the following steps:

  • Open a command prompt and navigate to the SymmetricDS installation.

  • In the security subdirectory, use the following commands to enter the old and new password for the keystore and each key entry.

keytool -keystore keystore -storepasswd
keytool -keystore keystore -alias sym -keypasswd
keytool -keystore keystore -alias sym.secret -keypasswd
  • Edit bin/setenv (or bin\setenv.bat on Windows) and conf/sym_service.conf files to update the new password.

-Djavax.net.ssl.keyStorePassword=changeit
  • Optionally, obfuscate the new password in the previous step to prevent casual observation.

syadmin obfuscate-text changeit

6.4.6. Generating Certificates

This section describes command line usage. See Security for using the web console instead.

To generate new keys and install a server certificate, use the following steps:

  • Open a command prompt and navigate to the security subdirectory of SymmetricDS.

  • Delete the old key pair and certificate.

keytool -keystore keystore -delete -alias sym
keytool -keystore cacerts -delete -alias sym
If you receive a message like, "Alias <sym> does not exist" - then the key entry does not exist and you can skip this step.
  • Generate a new key pair. Note that the first name/last name (the "CN") must match the fully qualified hostname the client will be using to communcate to the server.

keytool -keystore keystore -alias sym -genkey -keyalg RSA -validity 10950
  • Export the certificate from the private keystore.

keytool -keystore keystore -export -alias sym -rfc -file sym.cer
  • Install the certificate into the trusted keystore.

keytool -keystore cacerts -import -alias sym -file sym.cer
  • Copy the cacerts file that is generated by this process to the security directory of each client’s SymmetricDS installation.

6.5. Java Management Extensions

Monitoring and administrative operations can be performed using Java Management Extensions (JMX). SymmetricDS exposes JMX attributes and operations that can be accessed from the jmx command, Java’s jconsole, or a third party tool.

In order to use jconsole, you must enable JMX remote management in the JVM. You can edit the startup scripts to set the following system parameters.

-Dcom.sun.management.jmxremote.port=31417
-Dcom.sun.management.jmxremote.authenticate=false
-Dcom.sun.management.jmxremote.ssl=false

More details about enabling JMX for JConsole can be found here.

Using the Java jconsole command, SymmetricDS is listed as a local process named SymmetricLauncher. In jconsole, SymmetricDS appears under the MBeans tab under the name defined by the engine.name property. The default value is SymmetricDS.

The management interfaces under SymmetricDS are organized as follows:

Node

administrative operations

Parameters

access to properties set through the parameter service

6.6. JMS Publishing

With the proper configuration SymmetricDS can publish XML messages of captured data changes to JMS during routing or transactionally while data loading synchronized data into a target database. The following explains how to publish to JMS during synchronization to the target database.

The XmlPublisherDatabaseWriterFilter is a IDatabaseWriterFilter that may be configured to publish specific tables as an XML message to a JMS provider. See Extensions for information on how to configure an extension point. If the publish to JMS fails, the batch will be marked in error, the loaded data for the batch will be rolled back and the batch will be retried during the next synchronization run.

The following is an example extension point configuration that will publish four tables in XML with a root tag of 'sale'. Each XML message will be grouped by the batch and the column names identified by the groupByColumnNames property which have the same values.

<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xmlns:context="http://www.springframework.org/schema/context"
    xsi:schemaLocation="http://www.springframework.org/schema/beans
           http://www.springframework.org/schema/beans/spring-beans-3.0.xsd
           http://www.springframework.org/schema/context
           http://www.springframework.org/schema/context/spring-context-3.0.xsd">

    <bean id="configuration-publishingFilter"
      class="org.jumpmind.symmetric.integrate.XmlPublisherDatabaseWriterFilter">
        <property name="xmlTagNameToUseForGroup" value="sale"/>
        <property name="tableNamesToPublishAsGroup">
            <list>
               <value>SALE_TX</value>
               <value>SALE_LINE_ITEM</value>
               <value>SALE_TAX</value>
               <value>SALE_TOTAL</value>
            </list>
        </property>
        <property name="groupByColumnNames">
            <list>
               <value>STORE_ID</value>
               <value>BUSINESS_DAY</value>
               <value>WORKSTATION_ID</value>
               <value>TRANSACTION_ID</value>
            </list>
        </property>
        <property name="publisher">
           <bean class="org.jumpmind.symmetric.integrate.SimpleJmsPublisher">
               <property name="jmsTemplate" ref="definedSpringJmsTemplate"/>
           </bean>
        </property>
    </bean>
</beans>

The publisher property on the XmlPublisherDatabaseWriterFilter takes an interface of type IPublisher. The implementation demonstrated here is an implementation that publishes to JMS using Spring’s JMS template. Other implementations of IPublisher could easily publish the XML to other targets like an HTTP server, the file system or secure copy it to another server.

The above configuration will publish XML similar to the following:

<?xml version="1.0" encoding="UTF-8"?>
<sale xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
  id="0012010-01-220031234" nodeid="00001" time="1264187704155">
  <row entity="SALE_TX" dml="I">
    <data key="STORE_ID">001</data>
    <data key="BUSINESS_DAY">2010-01-22</data>
    <data key="WORKSTATION_ID">003</data>
    <data key="TRANSACTION_ID">1234</data>
    <data key="CASHIER_ID">010110</data>
  </row>
  <row entity="SALE_LINE_ITEM" dml="I">
    <data key="STORE_ID">001</data>
    <data key="BUSINESS_DAY">2010-01-22</data>
    <data key="WORKSTATION_ID">003</data>
    <data key="TRANSACTION_ID">1234</data>
    <data key="SKU">9999999</data>
    <data key="PRICE">10.00</data>
    <data key="DESC" xsi:nil="true"/>
  </row>
  <row entity="SALE_LINE_ITEM" dml="I">
    <data key="STORE_ID">001</data>
    <data key="BUSINESS_DAY">2010-01-22</data>
    <data key="WORKSTATION_ID">003</data>
    <data key="TRANSACTION_ID">1234</data>
    <data key="SKU">9999999</data>
    <data key="PRICE">10.00</data>
    <data key="DESC" xsi:nil="true"/>
  </row>
  <row entity="SALE_TAX" dml="I">
    <data key="STORE_ID">001</data>
    <data key="BUSINESS_DAY">2010-01-22</data>
    <data key="WORKSTATION_ID">003</data>
    <data key="TRANSACTION_ID">1234</data>
    <data key="AMOUNT">1.33</data>
  </row>
  <row entity="SALE_TOTAL" dml="I">
    <data key="STORE_ID">001</data>
    <data key="BUSINESS_DAY">2010-01-22</data>
    <data key="WORKSTATION_ID">003</data>
    <data key="TRANSACTION_ID">1234</data>
    <data key="AMOUNT">21.33</data>
  </row>
</sale>

To publish JMS messages during routing the same pattern is valid, with the exception that the extension point would be the XmlPublisherDataRouter and the router would be configured by setting the router_type of a ROUTER to the Spring bean name of the registered extension point. Of course, the router would need to be linked through TRIGGER_ROUTERs to each TRIGGER table that needs published.

6.7. File Synchronization

SymmetricDS not only supports the synchronization of database tables, but it also supports the synchronization of files and folders from one node to another.

6.7.1. File Synchronization Overview

File synchronization features include:
  • Monitoring one or more file system directory locations for file and folder changes

  • Support synchronizing a different target directory than the source directory

  • Use of wild card expressions to “include” or “exclude” files

  • Choice of whether to recurse into subfolders of monitored directories

  • Use of existing SymmetricDS routers to subset target nodes based on file and directory metadata

  • Ability to specify if files will be synchronized on creation, or deletion, and/or modification

  • Ability to specify the frequency with which file systems are monitored for changes

  • Ability to extend file synchronization through scripts that run before or after a file is copied to its source location

  • Support for bidirectional file synchronization

  • Like database synchronization, file synchronization is configured in a series of database tables. The configuration was designed to be similar to database synchronization in order to maintain consistency and to give database synchronization users a sense of familiarity.

For database synchronization, SymmetricDS uses Table Triggers to configure which tables will capture data for synchronization and Routers to designate which nodes will be the source of data changes and which nodes will receive the data changes. Table Routing links triggers to routers.

Likewise, for file synchronization, SymmetricDS uses File Triggers to designate which base directories will be monitored. Each entry in File Triggers designates one base directory to monitor for changes on the source system. The columns on File Triggers provide additional settings for choosing specific files in the base directory that will be monitored, and whether to recurse into subdirectories, etc. File triggers are linked to routers using File Routing. The file trigger router not only links the source and the target node groups, but it also optionally provides the ability to override the base directory name at the target. File Routing also provides a flag that indicates if the target node should be seeded with the files from the source node during SymmetricDS’s initial load process.

File synchronization does require a database for runtime information about the synchronization scenario. File Triggers will also need to be linked to an appropriate router like table triggers in order to complete the setup.
H2 database works great as a small lightweight database to support file synchronization runtime information if you do not have a relational database readily available to support file sync.

6.7.2. How File Synchronization Works

Not only is file synchronization configured similar to database synchronization, but it also operates in a very similar way. The file system is monitored for changes via a background job that tracks the file system changes (this parallels the use of triggers to monitor for changes when synchronizing database changes). When a change is detected it is written to the FILE_SNAPSHOT table. The file snapshot table represents the most recent known state of the monitored files. The file snapshot table has a SymmetricDS database trigger automatically installed on it so that when it is updated the changes are captured by SymmetricDS on an internal channel named filesync.

The changes to FILE_SNAPSHOT are then routed and batched by a file-synchronization-specific router that delegates to the configured router based on the File Routing configuration. The file sync router can make routing decisions based on the column data of the snapshot table, columns which contain attributes of the file like the name, path, size, and last modified time. Both old and new file snapshot data are also available. The router can, for example, parse the path or name of the file and use it as the node id to route to.

Batches of file snapshot changes are stored on the filesync channel in OUTGOING_BATCH. The existing SymmetricDS pull and push jobs ignore the filesync channel. Instead, they are processed by file-synchronization-specific push and pull jobs. These jobs, file sync tracker, file sync pull, and file sync push are turned off by default. They need to be started before any changes are processed.

When transferring data, the file sync push and pull jobs build a zip file dynamically based on the batched snapshot data. The zip file contains a directory per batch. The directory name is the batch_id. A sync.bsh Bean Shell script is generated and placed in the root of each batch directory. The Bean Shell script contains the commands to copy or delete files at their file destination from an extracted zip in the staging directory on the target node. The zip file is downloaded in the case of a pull, or, in the case of a push, is uploaded as an HTTP multi-part attachment. Outgoing zip files are written and transferred from the outgoing staging directory. Incoming zip files are staged in the filesync_incoming staging directory by source node id. The filesync_incoming/{node_id} staging directory is cleared out before each subsequent delivery of files.

The acknowledgement of a batch happens the same way it is acknowledged in database synchronization. The client responds with an acknowledgement as part of the response during a file push or pull.

6.8. Variables

Variables can be used throughout configuration with the $(variableName) syntax. Check the documentation for each configuration item to see which variables it supports. A substring of the variable value can be specified with a starting index and an optional ending index. The first character is at index 0, and the end index is not included in the substring. Therefore, the length of the substring will be end index minus start index.

$(variableName:start)
$(variableName:start:end)

Examples when $(externalId) is set to 00001-002:

$(externalId:0:5) - returns 00001
$(externalId:6) - returns 002

The variable value can be formatted using a format string supported by java.lang.String.format().

$(variableName|format_string)

Examples when $(externalId) is set to 1:

$(variableName|%05d) - returns 00001

6.9. Display Settings

There are a number of display settings available for the web console. They are accessible by clicking the cog icon in the top right corner of the screen and navigating to the "Display" tab of the settings dialog. These settings apply to the console user that is currently logged in.

display settings
Appearance

Determines whether to display the web console in light mode or dark mode.

Date Format

Determines what format dates are displayed in. Options include YYYY-MM-DD, DD/MM/YYYY, and MM/DD/YYYY.

Time Format

Determines whether to display times in a 12-hour or 24-hour clock format.

Time Zone

Determines what time zone dates and times are displayed in. Options include the system’s time zone, the browser’s time zone, and specific time zone offsets. In order to access this setting, the JVM’s time zone must be set to GMT or UTC. The JVM’s time zone is determined by the user.timezone property, which can be set in conf/sym_service.conf.

6.10. Preventing SymmetricDS from Running DDL Statements

It is possible to prevent SymmetricDS from running DDL statements and instead use the symadmin command to export the DDL statements and run them separately. To create a new node that does not run its own DDL statements, set up its properties file and add the following parameters:

auto.config.database=false
auto.sync.triggers=false

To export the DDL statements used to create the database objects that SymmetricDS requires (not including triggers), run the symadmin export-sym-objects subcommand and optionally specify an output file:

symadmin export-sym-objects <output file> --engine <engine name>

Run the exported DDL statements and then insert or import your configuration. If your configuration is saved to a .csv or .sql file, then you can import it using the symadmin import-config subcommand:

symadmin import-config <configuration file> --engine <engine name>

To export the trigger creation statements, run the symadmin sync-triggers subcommand with the -o option and specify an output file:

symadmin sync-triggers -o <output file> --engine <engine name>

Run the exported trigger creation statements and then start SymmetricDS. When making future configuration changes, it will be necessary to export any trigger changes using the above method and run them yourself.

7. Developer

This chapter focuses on a variety of ways for developers to build upon and extend some of the existing features found within SymmetricDS.

7.1. Extension Points

SymmetricDS has a pluggable architecture that can be extended. A Java class that implements the appropriate extension point interface, can implement custom logic and change the behavior of SymmetricDS to suit special needs. All supported extension points extend the IExtensionPoint interface. The available extension points are documented in the following sections.

When SymmetricDS starts up, the ExtensionPointManager searches a Spring Framework context for classes that implement the IExtensionPoint interface, then creates and registers the class with the appropriate SymmetricDS component.

Extensions should be configured in the conf/symmetric-extensions.xml file as Spring beans. The jar file that contains the extension should be placed in the web/WEB-INF/lib directory.

If an extension point needs access to SymmetricDS services or needs to connect to the database it may implement the ISymmetricEngineAware interface in order to get a handle to the ISymmetricEngine.

The INodeGroupExtensionPoint interface may be optionally implemented to indicate that a registered extension point should only be registered with specific node groups.

/**
 * Only apply this extension point to the 'root' node group.
 */
 public String[] getNodeGroupIdsToApplyTo() {
     return new String[] { "root" };
 }

7.1.1. IParameterFilter

Parameter values can be specified in code using a parameter filter. Note that there can be only one parameter filter per engine instance. The IParameterFilter replaces the deprecated IRuntimeConfig from prior releases.

public class MyParameterFilter
    implements IParameterFilter, INodeGroupExtensionPoint {

    /**
     * Only apply this filter to stores
     */
    public String[] getNodeGroupIdsToApplyTo() {
        return new String[] { "store" };
    }

    public String filterParameter(String key, String value) {
        // look up a store number from an already existing properties file.
        if (key.equals(ParameterConstants.EXTERNAL_ID)) {
            return StoreProperties.getStoreProperties().
              getProperty(StoreProperties.STORE_NUMBER);
        }
        return value;
    }

    public boolean isAutoRegister() {
        return true;
    }

}

7.1.2. IDatabaseWriterFilter

Data can be filtered or manipulated before it is loaded into the target database. A filter can change the data in a column, save it somewhere else or do something else with the data entirely. It can also specify by the return value of the function call that the data loader should continue on and load the data (by returning true) or ignore it (by returning false). One possible use of the filter, for example, might be to route credit card data to a secure database and blank it out as it loads into a less-restricted reporting database.

A DataContext is passed to each of the callback methods. A new context is created for each synchronization. The context provides a mechanism to share data during the load of a batch between different rows of data that are committed in a single database transaction.

The filter also provides callback methods for the batch lifecycle. The DatabaseWriterFilterAdapter may be used if not all methods are required.

A class implementing the IDatabaseWriterFilter interface is injected onto the DataLoaderService in order to receive callbacks when data is inserted, updated, or deleted.

public class MyFilter extends DatabaseWriterFilterAdapter {

    @Override
    public boolean beforeWrite(DataContext context, Table table, CsvData data) {
        if (table.getName().equalsIgnoreCase("CREDIT_CARD_TENDER")
                && data.getDataEventType().equals(DataEventType.INSERT)) {
            String[] parsedData = data.getParsedData(CsvData.ROW_DATA);
            // blank out credit card number
            parsedData[table.getColumnIndex("CREDIT_CARD_NUMBER")] = null;
        }
        return true;
    }
}

The filter class should be specified in conf/symmetric-extensions.xml as follows.

<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xmlns:context="http://www.springframework.org/schema/context"
    xsi:schemaLocation="http://www.springframework.org/schema/beans
           http://www.springframework.org/schema/beans/spring-beans-3.0.xsd
           http://www.springframework.org/schema/context
           http://www.springframework.org/schema/context/spring-context-3.0.xsd">

    <bean id="myFilter" class="com.mydomain.MyFilter"/>

</beans>

7.1.3. IDatabaseWriterErrorHandler

Implement this extension point to override how errors are handled. You can use this extension point to ignore rows that produce foreign key errors.

7.1.4. IDataLoaderFactory

Implement this extension point to provide a different implementation of the org.jumpmind.symmetric.io.data.IDataWriter that is used by the SymmetricDS data loader. Data loaders are configured for a channel. After this extension point is registered it can be activated for a CHANNEL by indicating the data loader name in the data_loader_type column.

SymmetricDS has two out of the box extensions of IDataLoaderFactory already implemented in its PostgresBulkDataLoaderFactory and OracleBulkDataLoaderFactory classes. These extension points implement bulk data loading capabilities for Oracle, Postgres and Greenplum dialects. See Appendix C. Database Notes for details.

Another possible use of this extension point is to route data to a NOSQL data sink.

7.1.5. IAcknowledgeEventListener

Implement this extension point to receive callback events when a batch is acknowledged. The callback for this listener happens at the point of extraction.

7.1.6. IReloadListener

Implement this extension point to listen in and take action before or after a reload is requested for a Node. The callback for this listener happens at the point of extraction.

7.1.7. IReloadVariableFilter

Implement this extension point to filter the SQL used by the initial load to query source tables and purge target tables. The extension receives the SQL and can replace variable names with values, which allows for adding new variables. The org.jumpmind.util.FormatUtils.replace() method can be used to find and replace variables.

7.1.8. ISyncUrlExtension

This extension point is used to select an appropriate URL based on the URI provided in the sync_url column of sym_node.

To use this extension point configure the sync_url for a node with the protocol of ext://beanName. The beanName is the name you give the extension point in the extension xml file.

7.1.9. IColumnTransform

This extension point allows custom column transformations to be created. There are a handful of out-of-the-box implementations. If any of these do not meet the column transformation needs of the application, then a custom transform can be created and registered. It can be activated by referencing the column transform’s name transform_type column of TRANSFORM_COLUMN

7.1.10. INodeIdCreator

This extension point allows SymmetricDS users to implement their own algorithms for how node ids and passwords are generated or selected during the registration process. There may be only one node creator per SymmetricDS instance (Please note that the node creator extension has replaced the node generator extension).

7.1.11. ITriggerCreationListener

Implement this extension point to get status callbacks during trigger creation.

7.1.12. IBatchAlgorithm

Implement this extension point and set the name of the Spring bean on the batch_algorithm column of the Channel table to use. This extension point gives fine grained control over how a channel is batched.

7.1.13. IDataRouter

Implement this extension point and set the name of the Spring bean on the router_type column of the Router table to use. This extension point gives the ability to programmatically decide which nodes data should be routed to.

7.1.14. IHeartbeatListener

Implement this extension point to get callbacks during the heartbeat job.

7.1.15. IOfflineClientListener

Implement this extension point to get callbacks for offline events on client nodes.

7.1.16. IOfflineServerListener

Implement this extension point to get callbacks for offline events detected on a server node during monitoring of client nodes.

7.1.17. INodePasswordFilter

Implement this extension point to intercept the saving and rendering of the node password.

7.1.18. ISmtpPasswordFilter

Implement this extension point to intercept the saving and rendering of the SMTP password.

7.1.19. IPurgeListener

Implement this extension point to execute some functionality before and after the execution of the Outgoing Purge or the Incoming Purge.

7.2. Embedding in Android

SymmetricDS has its web-enabled, fault-tolerant, database synchronization software available on the Android mobile computing platform. The Android client follows all of the same concepts and brings to Android all of the same core SymmetricDS features as the full-featured, Java-based SymmetricDS client. The Android client is a little bit different in that it is not a stand-alone application, but is designed to be referenced as a library to run in-process with an Android application requiring synchronization for its SQLite database.

By using SymmetricDS, mobile application development is simplified, in that the mobile application developer can now focus solely on interacting with their local SQLite database. SymmetricDS takes care of capturing and moving data changes to and from a centralized database when the network is available

The same core libraries that are used for the SymmetricDS server are also used for Android. SymmetricDS’s overall footprint is reduced by eliminating a number of external dependencies in order to fit better on an Android device. The database access layer is abstracted so that the Android specific database access layer could be used. This allows SymmetricDS to be efficient in accessing the SQLite database on the Android device.

In order to convey how to use the SymmetricDS Android libraries, the example below will show how to integrate an Android application with the Quick-Start demo configuration. This Android application will create an embedded SQLite database, all SymmetricDS run-time tables, and configure the Quick-Start demo tables (ITEM, ITEM_SELLING_PRICE, SALE_TRANSACTION, SALE_RETURN_LINE_ITEM) for synchronization. The application has a basic user interface that can run queries on the SQLite database to demonstrate synchronization with a Quick-Start demo corp server.

Android Studio 3.0.1 and Android SDK 26 were used for this example. The example Android application can be cloned into Android Studio using the GitHub project located here: https://github.com/JumpMind/symmetric-android-client-demo

SymmetricDS for Android comes as a zip file of Java archives (jar files) that are required by the SymmetricDS client at runtime. The libs directory containing the necessary .jar files can be downloaded as a zip archive (symmetric-android-VERSION.zip) from the SymmetricDS downloads page. Alternatively, the libs directory can be generated by cloning the symmetric-ds GitHub repository and running the following command in the symmetric-assemble directory:

./gradlew androidDistZip

The first step to using SymmetricDS in an Android application is to unzip the jar files into a location where the project will recognize them. The latest Android SDK requires that these jar files be put into a libs directory under the app directory of the Android application project.

In order to sync properly, the Sync URL of the corp-000 node must be updated to use the IP address of host rather than localhost. Then, update the String REGISTRATION_URL in the DbProvider class of the Android project to the new Sync URL of the corp-000 node.

Next, set up an Android Emulator or connect and Android device. This can be done by opening the Android Virtual Device Manager. Click New and follow the steps. The higher the Emulator’s API, the better.

Run your Android Application by pressing the Run button in Android Studio. When prompted, select the emulator you just created. Monitor the Console in Android Studio. Let the apk install on the emulator. Now watch the LogCat and wait as it attempts to register with your SymmetricDS Master Node.

The core functionality of SymmetricDS on Android is implemented by starting the SymmetricService class as an Android service. This requires building the SymmetricDS Android libraries using the steps mentioned above and adding them to your Android project’s dependencies.

The SymmetricService Intent is defined in the AndroidManifest.xml using the following XML snippet:

<service android:name="org.jumpmind.symmetric.android.SymmetricService" android:enabled="true" >
        <intent-filter>
                <action android:name="org.jumpmind.symmetric.android.SymmetricService" />
        </intent-filter>
</service>

The SymmetricService Intent is started using the following java code:

Intent intent = new Intent(getContext(), SymmetricService.class);

// Replace extras with desired node configuration
intent.putExtra(SymmetricService.INTENTKEY_SQLITEOPENHELPER_REGISTRY_KEY, DATABASE_NAME);
intent.putExtra(SymmetricService.INTENTKEY_REGISTRATION_URL, REGISTRATION_URL);
intent.putExtra(SymmetricService.INTENTKEY_EXTERNAL_ID, NODE_ID);
intent.putExtra(SymmetricService.INTENTKEY_NODE_GROUP_ID, NODE_GROUP);
intent.putExtra(SymmetricService.INTENTKEY_START_IN_BACKGROUND, true);

Properties properties = new Properties();
// Put any additional SymmetricDS parameters into properties
intent.putExtra(SymmetricService.INTENTKEY_PROPERTIES, properties);
getContext().startService(intent);

7.3. Embedding in C/C++

A minimal implementation of the SymmetricDS client is written in C, which includes a shared library named "libsymclient" and a command line executable named "sym" for synchronizing a database. It currently only supports the SQLite database. The SymmetricDS C library and client are built from the following projects:

symmetric-client-clib

This project contains most of the code and builds the libsymclient C library. It depends on libcurl, libsqlite3, and libcsv.

symmetric-client-clib-test

This project links against the C library to runs unit tests. It also depends on the CUnit library.

symmetric-client-native

This project links against the C library to build the sym executable.

The binaries are built using Eclipse CDT (C/C++ Development Tooling), which is an Integrated Developer Environment based on the Eclipse platform. A distribution of Eclipse CDT can be downloaded or an existing Eclipse installation can be updated to include the CDT. (See https://eclipse.org/cdt/ for information and downloads.) In the future, the projects above will switch to a general build system like Autotools for automating builds, but for now Eclipse is required.

The "sym" executable can be run from the command line and expects the "libsymclient.so" library to be installed on the system. If running from the project directories during development, the path to the library can be specified with the LD_LIBRARY_PATH environment variable on Linux, the DYLD_LIBRARY_PATH on Mac OS X, or PATH on Windows. The executable will look for a "symmetric.properties" file containing startup parameters in the user’s home directory or in the current directory:

LD_LIBRARY_PATH=../../symmetric-client-clib/Debug ./sym

It will also accept an argument of the path and filename of the properties file to use:

LD_LIBRARY_PATH=../../symmetric-client-clib/Debug ./sym /path/to/client.properties

The client uses Startup Parameters to connect to a database, identify itself, and register with a server to request synchronization. Here is an example client.properties file:

db.url=sqlite:file:test.db
group.id=store
external.id=003
registration.url=http://localhost:31415/sync/corp-000

The symmetric-client-native project is an example of how to use the SymEngine API provided by the C library. The C library uses an object-oriented pattern and follows the same naming conventions as the Java project. All symbol names in the C library are prefixed with "Sym". Each Java class is represented in C with a struct that contains member data and pointers to member functions. Here is an example C program that runs the SymmetricDS engine:

#include "libsymclient.h"

int main(int argCount, char **argValues) {

    // Startup and runtime parameters
    SymProperties *prop = SymProperties_new(NULL);
    prop->put(prop, SYM_PARAMETER_DB_URL, "sqlite:file:data.db");
    prop->put(prop, SYM_PARAMETER_GROUP_ID, "store");
    prop->put(prop, SYM_PARAMETER_EXTERNAL_ID, "003");
    prop->put(prop, SYM_PARAMETER_REGISTRATION_URL, "http://localhost:31415/sync/corp-000");

    // Uncomment to read parameters from a file instead
    //SymProperties *prop = SymProperties_newWithFile(NULL, fileName);

    SymEngine *engine = SymEngine_new(NULL, prop);
    // Connects to database, creates config/runtime tables and triggers
    engine->start(engine);

    // Pull changes from remote nodes
    engine->pull(engine);

    // Create batches of captured changes
    engine->route(engine);

    // Push changes to remote nodes
    engine->push(engine);

    // Create a heartbeat batch with current host information
    engine->heartbeat(engine, 0);

    // Purge old batch data that has successfully synced
    engine->purge(engine);

    // Clean up
    engine->stop(engine);
    engine->destroy(engine);
    prop->destroy(prop);

    return 0;
}

8. By Example

This chapter focuses on using examples for a variety of use cases with SymmetricDS.

8.1. Replicating CSV File Into a Database

8.1.1. SymmetricDS Setup

  • Create or choose a target table for your target node group. For this example, we will create a table, person, for the "corp" node group.

createTable
  • Next, a file trigger should be created to watch for csv files at the specified base directory. You can also specify files that should be included to set off the trigger.

FileSync needs to be enabled for each node group to create file triggers.
createFileTrigger
  • After the file trigger is created, a router should be created so that the changes can be routed to the correct location. The communication link needs to be specified as well as the router type and target table. In this example, "store" is the source node group which communicates to the target node group,"corp". The router type is csv and person is the target table.

createRouter
  • Finally, a file routing link should be created between the file trigger and router.

createFileRouting
  • Testing. You are now ready to test your csv to database routing. Add a csv file that matches your table specifications to your base directory or make a change to a pre-existing csv file in the directory. File sync tracker checks for changes in already sync’d files and runs every 5 minutes. File sync pull checks for new files to pull down and runs every 1 minute. Depending on the change you choose (changing a file or adding a file), wait the appropriate amount of time and then verify that the changes are shown in the target table.

The following SQL statement will verify changes to the person table.
select * from person;

Appendix A: Data Model

What follows is the complete SymmetricDS data model. Note that all tables are prepended with a configurable prefix so that multiple instances of SymmetricDS may coexist in the same database. The default prefix is sym_.

SymmetricDS configuration is entered by the user into the data model to control the behavior of what data is synchronized to which nodes.

data model config
Figure 23. Configuration Data Model

At runtime, the configuration is used to capture data changes and route them to nodes. The data changes are placed together in a single unit called a batch that can be loaded by another node. Outgoing batches are delivered to nodes and acknowledged. Incoming batches are received and loaded. History is recorded for batch status changes and statistics.

data model runtime
Figure 24. Runtime Data Model

A.1. CHANNEL

This table represents a category of data that can be synchronized independently of other channels. Channels allow control over the type of data flowing and prevents one type of synchronization from contending with another.

Table 23. CHANNEL

Name

Type

Size

Default

Keys

Not Null

Description

CHANNEL_ID

VARCHAR

128

PK

X

A unique identifer, usually named something meaningful, like 'sales' or 'inventory'.

PROCESSING_ORDER

INTEGER

1

X

Order of sequence to process channel data.

MAX_BATCH_SIZE

INTEGER

1000

X

The maximum number of Data Events to process within a batch for this channel.

MAX_BATCH_TO_SEND

INTEGER

60

X

The maximum number of batches to send during a 'synchronization' between two nodes. A 'synchronization' is equivalent to a push or a pull. If there are 12 batches ready to be sent for a channel and max_batch_to_send is equal to 10, then only the first 10 batches will be sent.

MAX_DATA_TO_ROUTE

INTEGER

100000

X

The maximum number of data rows to route for a channel at a time.

EXTRACT_PERIOD_MILLIS

INTEGER

0

X

The minimum number of milliseconds allowed between attempts to extract data for targeted at a node_id.

ENABLED

TINYINT

1

1

X

Indicates whether channel is enabled or not.

USE_OLD_DATA_TO_ROUTE

TINYINT

1

1

X

Indicates whether to read the old data during routing.

USE_ROW_DATA_TO_ROUTE

TINYINT

1

1

X

Indicates whether to read the row data during routing.

USE_PK_DATA_TO_ROUTE

TINYINT

1

1

X

Indicates whether to read the pk data during routing.

RELOAD_FLAG

TINYINT

1

0

X

Indicates that this channel is used for reloads.

FILE_SYNC_FLAG

TINYINT

1

0

X

Indicates that this channel is used for file sync.

CONTAINS_BIG_LOB

TINYINT

1

0

X

Provides SymmetricDS a hint on how to treat captured data. Currently only supported by Oracle, Interbase and Firebird. If set to '0', then selects for routing and data extraction will be more efficient and lobs will be truncated at 4k in the trigger text. When it is set to '0' there is a 4k limit on the total size of a row and on the size of a LOB column.

BATCH_ALGORITHM

VARCHAR

50

default

X

The algorithm to use when batching data on this channel. Possible values are: 'default', 'transactional', and 'nontransactional'

DATA_LOADER_TYPE

VARCHAR

50

default

X

Identify the type of data loader this channel should use. Allows for the default dataloader to be swapped out via configuration for more efficient platform specific data loaders.

DESCRIPTION

VARCHAR

255

Description on the type of data carried in this channel.

QUEUE

VARCHAR

25

default

X

User provided queue name for channel to operate on. Creates multi-threaded channels. Defaults to 'default' thread

MAX_NETWORK_KBPS

DECIMAL

10,3

0.000

X

The maximum network transfer rate in kilobytes per second. Zero or negative means unlimited. Channels running in serial or parallel can have an effect on how much bandwidth can be used and when a channel will be processed. This is currently only implemented when staging is enabled.

DATA_EVENT_ACTION

CHAR

1

For a node group link with a data event action of B (both), select how to send changes to the target node group. (P = Push, W = Wait for Pull)

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.2. COMPARE_REQUEST

Compare one or more tables between source and target nodes with option to repair data

Table 24. COMPARE_REQUEST

Name

Type

Size

Default

Keys

Not Null

Description

SOURCE_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the source node.

TARGET_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the target node.

TRIGGER_ID

VARCHAR

128

PK

X

Unique identifier for a trigger associated with the request, or ALL for all tables.

ROUTER_ID

VARCHAR

50

PK

X

Unique description of the router associated with the request, or ALL for all tables.

CREATE_TIME

TIMESTAMP

2

PK

X

Timestamp when this entry was created.

USE_TABLE_CHECKSUM

TINYINT

1

0

X

Use a checksum on the table to detect changes, otherwise a simple row count is used for comparison.

USE_INFLIGHT

TINYINT

1

0

X

Use inflight changes , which are changes in the process of syncing. Currently, it assumes inflight changes will sync successfully, so they are ignored during comparisons. Requires use_table_checksum = 1.

USE_ROW_LEVEL

TINYINT

1

0

X

When a table count or checksum doesn’t match, find the rows that are different, missing, and extra.

USE_CHUNKING

TINYINT

1

0

X

When finding row differences, use a chunking technique that performs a checksum on groups of rows. Instead of sending all rows with their checksum for comparison, it will send chunks for efficient network communication. The algorithm will limit the number of chunks to 100.

USE_SCRIPT

TINYINT

1

0

X

Generates a script of the differences as inserts, updates, and deletes, but does not apply the script. The use_repair option will apply the script.

USE_REPAIR

TINYINT

1

0

X

Runs the repair script of inserts, updates, and deletes on the target so it matches the source.

CHECKSUM_ALGORITHM

VARCHAR

15

crc32

X

One of CRC32, PureJavaCrc32, Adler32, or XXHash32

DECIMAL_SCALE

SMALLINT

-1

Number of digits to the right of the decimal point to use when comparing numbers.

TIMESTAMP_SCALE

SMALLINT

-1

Number of digits to the right of the decimal point for fractional seconds to use when comparing timestamps.

REAL_SCALE

SMALLINT

-1

Number of digits to use when comparing floating point numbers.

NOT_BEFORE

TIMESTAMP

2

User requested start time.

PROCESSED

TINYINT

1

0

X

System field indicates compare_table_status was created.

COMPLETED

TINYINT

1

0

X

System field to indicate request is complete.

CANCELLED

TINYINT

1

0

X

User can set this flag to interrupt and cancel.

COMPARE_ID

BIGINT

System field with compare ID sequence.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

2

X

Timestamp when a user or the system last updated this entry.

A.3. COMPARE_STATUS

Compare one or more tables between source and target nodes with option to repair data

Table 25. COMPARE_STATUS

Name

Type

Size

Default

Keys

Not Null

Description

COMPARE_ID

BIGINT

PK

X

System field with compare ID sequence.

SOURCE_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the source node.

TARGET_NODE_ID

VARCHAR

50

X

Unique identifier for the target node.

START_TIME

TIMESTAMP

2

Timestamp when this entry was created.

END_TIME

TIMESTAMP

2

Timestamp when this entry was created.

TABLES_REQUESTED

INTEGER

0

X

The number of tables associated with this compare, which is generated by the system.

TABLES_COMPARED

INTEGER

0

X

The number of tables completed for this compare, which is generated by the system.

COMPLETED

TINYINT

1

0

X

System field to indicate request is complete.

CANCELLED

TINYINT

1

0

X

System field to indicate request was cancelled.

ERROR_FLAG

TINYINT

1

0

X

A flag that indicates that this request is in error during the last attempt to process.

CREATE_TIME

TIMESTAMP

2

PK

X

Timestamp when this entry was created.

LAST_UPDATE_TIME

TIMESTAMP

2

X

Timestamp when a user or the system last updated this entry.

A.4. COMPARE_TABLE_STATUS

Status by table of a compare_request to compare one or more tables.

Table 26. COMPARE_TABLE_STATUS

Name

Type

Size

Default

Keys

Not Null

Description

COMPARE_ID

BIGINT

PK

X

Compare ID sequence.

SOURCE_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the source node.

STEP_NUM

INTEGER

PK

X

Step number for each table to compare.

TARGET_NODE_ID

VARCHAR

50

X

Unique identifier for the target node.

CURRENT_LEVEL

CHAR

1

T

X

Level of current comparision. T=table, C=chunk, R=row

STATUS

CHAR

2

RQ

X

Status of this table comparisons. RQ=requested and queued, NE=new and processing, SE=sent payload for level, OK=complete

SOURCE_CATALOG_NAME

VARCHAR

255

The catalog name where the source table resides.

SOURCE_SCHEMA_NAME

VARCHAR

255

The schema name where the source table resides.

SOURCE_TABLE_NAME

VARCHAR

255

X

The name of the source table.

SOURCE_ROW_COUNT

BIGINT

0

X

The number of rows in the source table.

SOURCE_ROW_COUNT_MILLIS

BIGINT

0

X

Time it took to count rows at source when use_table_checksum = 0.

SOURCE_ROW_EXTRACT_MILLIS

BIGINT

0

X

Time it took to extract all rows at source, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different.

SOURCE_ROW_EXTRACT_SIZE

BIGINT

0

X

The number of bytes extracted to staging from the source table.

SOURCE_ROW_SORT_MILLIS

BIGINT

0

X

Time it took to sort all rows at source, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different

SOURCE_ROW_CHECKSUM_MILLIS

BIGINT

0

X

Time it took to checksum all rows at source, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different

SOURCE_ROW_FILTER_MILLIS

BIGINT

0

X

Time it took to filter rows for chunks that didn’t match at the source.

SOURCE_TABLE_CHECKSUM_MILLIS

BIGINT

0

X

Time it took to checksum the table at source when use_table_checksum = 1

SOURCE_INFLIGHT_COUNT

BIGINT

0

X

The number of inflight rows waiting to sync for the source table.

SOURCE_INFLIGHT_MILLIS

BIGINT

0

X

The number of millis to process inflight rows.

SOURCE_CHUNK_MILLIS

BIGINT

0

X

Time it took to process rows into chunks at source.

TARGET_CATALOG_NAME

VARCHAR

255

The catalog name where the target table resides.

TARGET_SCHEMA_NAME

VARCHAR

255

The schema name where the target table resides.

TARGET_TABLE_NAME

VARCHAR

255

X

The name of the target table.

TARGET_ROW_COUNT

BIGINT

0

X

The number of rows in the target table.

TARGET_ROW_COUNT_MILLIS

BIGINT

0

X

Time it took to count rows at target when use_table_checksum = 0.

TARGET_ROW_EXTRACT_MILLIS

BIGINT

0

X

Time it took to extract all rows at target, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different.

TARGET_ROW_EXTRACT_SIZE

BIGINT

0

X

The number of bytes extracted to staging from the target table.

TARGET_ROW_SORT_MILLIS

BIGINT

0

X

Time it took to sort all rows at target, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different

TARGET_ROW_CHECKSUM_MILLIS

BIGINT

0

X

Time it took to checksum all rows at target, either because use_table_checksum = 1 or because use_repair = 1 and the row count is different

TARGET_ROW_FILTER_MILLIS

BIGINT

0

X

Time it took to filter rows for chunks that didn’t match at the target.

TARGET_TABLE_CHECKSUM_MILLIS

BIGINT

0

X

Time it took to checksum the table at target when use_table_checksum = 1

TARGET_INFLIGHT_COUNT

BIGINT

0

X

The number of inflight rows waiting to sync for the target table.

TARGET_INFLIGHT_MILLIS

BIGINT

0

X

The number of millis to process inflight rows.

TARGET_CHUNK_MILLIS

BIGINT

0

X

Time it took to process rows into chunks at target.

CHUNK_MATCH_COUNT

BIGINT

0

X

Number of chunks that matched, if chunk level is used.

CHUNK_DIFF_COUNT

BIGINT

0

X

Number of chunks that did not match, if chunk level is used.

CHUNK_DIFF_LIST

VARCHAR

800

List of chunk numbers (up to 200) that are different and need row-level comparison.

CHUNK_COMPARE_MILLIS

BIGINT

0

X

Time it took to compare chunks.

ROW_MATCH_COUNT

BIGINT

0

X

Number of rows that matched.

ROW_DIFF_COUNT

BIGINT

0

X

Number of rows that did not match.

ROW_MISSING_COUNT

BIGINT

0

X

Number of rows missing on the target.

ROW_EXTRA_COUNT

BIGINT

0

X

Number of rows that were extra on the target.

ROW_COMPARE_MILLIS

BIGINT

0

X

Time it took to compare rows.

REPAIR_QUERY_MILLIS

BIGINT

0

X

Time it took to query and write repair script.

REPAIR_LOAD_MILLIS

BIGINT

0

X

Time it took to load repair script.

REPAIR_REJECT_COUNT

INTEGER

0

X

The number of SQL statements that failed to apply from the repair SQL script.

MATCH_FLAG

TINYINT

1

0

X

After status is OK, whether or not source and target tables match.

IGNORE_FLAG

TINYINT

1

0

X

Indicates this table was ignored by user cancelling request.

ERROR_FLAG

TINYINT

1

0

X

Indicates an error occurred.

ERROR_MESSAGE

LONGVARCHAR

When error_flag is set, this contains details on the error.

START_TIME

TIMESTAMP

2

When comparison for this table started.

END_TIME

TIMESTAMP

2

When comparison for this table ended.

LAST_UPDATE_TIME

TIMESTAMP

2

X

Timestamp when the system last updated this entry.

A.5. CONFLICT

Defines how conflicts in row data should be handled during the load process.

Table 27. CONFLICT

Name

Type

Size

Default

Keys

Not Null

Description

CONFLICT_ID

VARCHAR

50

PK

X

Unique identifier for a specific conflict detection setting.

SOURCE_NODE_GROUP_ID

VARCHAR

50

FK

X

The source node group for which this setting will be applied to. References a node group link.

TARGET_NODE_GROUP_ID

VARCHAR

50

FK

X

The target node group for which this setting will be applied to. References a node group link.

TARGET_CHANNEL_ID

VARCHAR

128

Optional channel that this setting will be applied to.

TARGET_CATALOG_NAME

VARCHAR

255

Optional database catalog that the target table belongs to. Only use this if the target table is not in the default catalog.

TARGET_SCHEMA_NAME

VARCHAR

255

Optional database schema that the target table belongs to. Only use this if the target table is not in the default schema.

TARGET_TABLE_NAME

VARCHAR

255

Optional database table that this setting will apply to. If left blank, the setting will be for any table in the channel (if set) and in the specified node group link.

DETECT_TYPE

VARCHAR

128

X

Indicates the strategy to use for detecting conflicts during a dml action. The possible values are: use_pk_data (manual, fallback, ignore), use_changed_data (manual, fallback, ignore), use_old_data (manual, fallback, ignore), use_timestamp (newer_wins), use_version (newer_wins)

DETECT_EXPRESSION

LONGVARCHAR

An expression that provides additional information about the detection mechanism. If the detection mechanism is use_timestamp or use_version then this expression will be the name of the timestamp or version column.

RESOLVE_TYPE

VARCHAR

128

X

Indicates the strategy for resolving update conflicts. The possible values differ based on the detect_type that is specified.

PING_BACK

VARCHAR

128

X

Indicates the strategy for sending resolved conflicts back to the source system. Possible values are: OFF, SINGLE_ROW, and REMAINING_ROWS.

RESOLVE_CHANGES_ONLY

TINYINT

1

0

Indicates that when applying changes during an update that only data that has changed should be applied. Otherwise, all the columns will be updated. This really only applies to updates.

RESOLVE_ROW_ONLY

TINYINT

1

0

Indicates that an action should take place for the entire batch if possible. This applies to a resolve type of 'ignore'. If a row is in conflict and the resolve type is 'ignore', then the entire batch will be ignored.

CREATE_TIME

TIMESTAMP

X

The date and time when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

The date and time when a user last updated this entry.

A.6. CONSOLE_EVENT

Events triggered by console users

Table 28. CONSOLE_EVENT

Name

Type

Size

Default

Keys

Not Null

Description

USER_ID

VARCHAR

50

PK

X

The user who last updated this entry.

EVENT_CODE

VARCHAR

50

PK

X

CREATE_TIME

TIMESTAMP

PK

X

Timestamp when this entry was created.

EVENT_NODE_ID

VARCHAR

50

Node the event occurred on

SOURCE_NODE_ID

VARCHAR

50

Source node of the event

TARGET_NODE_ID

VARCHAR

50

Target node of the event

EVENT_VALUE

LONGVARCHAR

Placeholder for additional information about the event

A.7. CONSOLE_ROLE

Role for users to assign privileges to use screens

Table 29. CONSOLE_ROLE

Name

Type

Size

Default

Keys

Not Null

Description

ROLE_ID

VARCHAR

50

PK

X

PARENT_ROLE_ID

VARCHAR

50

FK

ROLE_DESCRIPTION

VARCHAR

255

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.8. CONSOLE_ROLE_PRIVILEGE

List of privileges for a role

Table 30. CONSOLE_ROLE_PRIVILEGE

Name

Type

Size

Default

Keys

Not Null

Description

ROLE_ID

VARCHAR

50

PK FK

X

PRIVILEGE_NAME

VARCHAR

80

PK

X

IS_READ

TINYINT

1

0

X

Indicates that this privilege is read only.

IS_WRITE

TINYINT

1

0

X

Indicates that this privilege is read/write.

A.9. CONSOLE_TABLE_STATS

Statistics for all tables replicating

Table 31. CONSOLE_TABLE_STATS

Name

Type

Size

Default

Keys

Not Null

Description

TABLE_NAME

VARCHAR

255

PK

X

The table name representing the statistics

EVENT_TYPE

CHAR

1

PK

X

The type of event representing this entry. Valid values are 'I' for insert, 'U' for update, or 'D' for delete.

START_TIME

TIMESTAMP

PK

X

The start time for the period which this row represents.

END_TIME

TIMESTAMP

PK

X

The start time for the period which this row represents.

LOADED_ROWS

BIGINT

0

X

The rows loaded for this table during this period.

EXTRACTED_ROWS

BIGINT

0

The rows extracted for this table during this period.

A.10. CONSOLE_USER

A console specific table that is used to store console users

Table 32. CONSOLE_USER

Name

Type

Size

Default

Keys

Not Null

Description

USER_ID

VARCHAR

50

PK

X

FIRST_NAME

VARCHAR

50

X

LAST_NAME

VARCHAR

50

X

EMAIL

VARCHAR

50

HASHED_PASSWORD

VARCHAR

255

SALT

VARCHAR

50

Salt used in encryption, null if no salt

ENCRYPT_METH

VARCHAR

50

Encryption method for password, i.e. SHA-1 or SHA-512

FAILED_LOGINS

TINYINT

0

Number of failed login attempts

AUTH_METH

VARCHAR

50

INTERNAL

USER_ROLE

VARCHAR

50

ADMIN

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_LOGIN_TIME

TIMESTAMP

Timestamp when user last logged in.

LAST_LOGIN_IP_ADDR

VARCHAR

39

The IP from which the user last logged in.

LAST_LOGIN_NODE_ID

VARCHAR

50

The node where the user last logged in.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

LAST_PASSWORD_TIME

TIMESTAMP

Timestamp when a user last updated their password.

OTP_KEY

VARCHAR

32

Two Factor Authentication Key.

IS_RESET_REQUIRED

TINYINT

1

0

X

Indicates if a password reset is required.

CONFIRM_CODE

VARCHAR

32

Confirmation code for email verification and forgot password

CONFIRM_CREATE_TIME

TIMESTAMP

Create time for a confirmation code

IS_GLASS_BREAK_REQUIRED

TINYINT

1

0

X

Indicates if a glass break is required to login as the user.

BREAK_GLASS_START_TIME

TIMESTAMP

Timestamp when the glass was last broken.

BREAK_GLASS_END_TIME

TIMESTAMP

Timestamp when the glass break expires.

BREAK_GLASS_USER_ID

VARCHAR

50

ID of the user who last broke the glass for this user.

ENABLED

TINYINT

1

1

X

A.11. CONSOLE_USER_HIST

A console specific table that is used to store console users login history

Table 33. CONSOLE_USER_HIST

Name

Type

Size

Default

Keys

Not Null

Description

HIST_ID

INTEGER

PK

X

ID for a users history, lower ID’s represent earlier passwords

USER_ID

VARCHAR

50

PK

X

HASHED_PASSWORD

VARCHAR

255

SALT

VARCHAR

50

Salt used in encryption, null if no salt

ENCRYPT_METH

VARCHAR

50

Encryption method for password, i.e. SHA-1 or SHA-512

AUTH_METH

VARCHAR

50

INTERNAL

A.12. CONTEXT

Context variables used by runtime services on a single node

Table 34. CONTEXT

Name

Type

Size

Default

Keys

Not Null

Description

NAME

VARCHAR

80

PK

X

The name of the context variable.

CONTEXT_VALUE

LONGVARCHAR

The value of the context variable.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when emtry was last updated.

A.13. DATA

The captured data change that occurred to a row in the database. Entries in data are created by database triggers.

Table 35. DATA

Name

Type

Size

Default

Keys

Not Null

Description

DATA_ID

BIGINT

PK

X

Unique identifier for a data.

TABLE_NAME

VARCHAR

255

X

The name of the table in which a change occurred that this entry records.

EVENT_TYPE

CHAR

1

X

The type of event captured by this entry. For triggers, this is the change that occurred, which is 'I' for insert, 'U' for update, or 'D' for delete. Other events include: 'R' for reloading the entire table (or subset of the table) to the node; 'S' for running dynamic SQL at the node, which is used for adhoc administration.

ROW_DATA

LONGVARCHAR

The captured data change from the synchronized table. The column values are stored in comma-separated values (CSV) format.

PK_DATA

LONGVARCHAR

The primary key values of the captured data change from the synchronized table. This data is captured for updates and deletes. The primary key values are stored in comma-separated values (CSV) format.

OLD_DATA

LONGVARCHAR

The captured data values prior to the update. The column values are stored in CSV format.

TRIGGER_HIST_ID

INTEGER

X

The foreign key to the trigger_hist entry that contains the primary key and column names for the table being synchronized.

CHANNEL_ID

VARCHAR

128

The channel that this data belongs to, such as 'prices'

TRANSACTION_ID

VARCHAR

255

An optional transaction identifier that links multiple data changes together as the same transaction.

SOURCE_NODE_ID

VARCHAR

50

If the data was inserted by a SymmetricDS data loader, then the id of the source node is record so that data is not re-routed back to it.

EXTERNAL_DATA

VARCHAR

50

A field that can be populated by a trigger that uses the EXTERNAL_SELECT

NODE_LIST

VARCHAR

255

A field that can be populated with a comma separated subset of node ids which will be the only nodes available to the router

IS_PREROUTED

TINYINT

1

0

X

Set to true when routing should ignore this row because data_event and outgoing_batch rows are manually entered.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

A.14. DATA_EVENT

Each row represents the mapping between a data change that was captured and the batch that contains it. Entries in data_event are created as part of the routing process.

Table 36. DATA_EVENT

Name

Type

Size

Default

Keys

Not Null

Description

DATA_ID

BIGINT

PK

X

Id of the data to be routed.

BATCH_ID

BIGINT

PK

X

Id of the batch containing the data.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

A.15. DATA_GAP

Used only when routing.data.reader.type is set to 'gap.' Table that tracks gaps in the data table so that they may be processed efficiently, if data shows up. Gaps can show up in the data table if a database transaction is rolled back.

Table 37. DATA_GAP

Name

Type

Size

Default

Keys

Not Null

Description

START_ID

BIGINT

PK

X

The first missing data_id from the data table where a gap is detected. This could be the last data_id inserted plus one.

END_ID

BIGINT

PK

X

The last missing data_id from the data table where a gap is detected. If the start_id is the last data_id inserted plus one, then this field is filled in with a -1.

IS_EXPIRED

TINYINT

0

X

Routing will watch a gap for data to commit up until the time of routing.stale.dataid.gap.time.ms parameter is reached, then this flag is set to expire it. The purge job will see any data that commits after expiration as stranded data.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_HOSTNAME

VARCHAR

255

The host who last updated this entry.

A.16. DESIGN_DIAGRAM

Table 38. DESIGN_DIAGRAM

Name

Type

Size

Default

Keys

Not Null

Description

ID

VARCHAR

36

PK

X

A unique identifier

ROW_ID

VARCHAR

36

X

A unique identifier for this row across different versions

NAME

VARCHAR

255

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

CREATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_BY

VARCHAR

50

The user who created this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.17. DIAGRAM_GROUP

Table 39. DIAGRAM_GROUP

Name

Type

Size

Default

Keys

Not Null

Description

ID

VARCHAR

36

PK

X

A unique identifier

DIAGRAM_ID

VARCHAR

36

PK FK

X

A unique identifier

NAME

VARCHAR

50

The name to be displayed

X

INTEGER

10

X

null

Y

INTEGER

10

X

null

DATABASE_TYPE

VARCHAR

50

The database type this node group represents.

APPROXIMATE_ORDER

INTEGER

0

X

null

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

CREATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_BY

VARCHAR

50

The user who created this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.18. EXTENSION

Dynamic extensions stored in the database that plug-in to the running engine and receive callbacks according to their interface.

Table 40. EXTENSION

Name

Type

Size

Default

Keys

Not Null

Description

EXTENSION_ID

VARCHAR

50

PK

X

The unique id of the extension.

EXTENSION_TYPE

VARCHAR

10

X

The type of the extension. Types are 'java' and 'bsh'

INTERFACE_NAME

VARCHAR

255

Name of interface, required for 'bsh' only.

NODE_GROUP_ID

VARCHAR

50

X

Target the extension at a specific node group id. To target all groups, use the value of 'ALL'.

ENABLED

TINYINT

1

1

X

Whether or not the extension is enabled.

EXTENSION_ORDER

INTEGER

1

X

Specifies the order in which to install extensions when multiple extensions implement the same interface.

EXTENSION_TEXT

LONGVARCHAR

The script or code of the extension.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.19. EXTRACT_REQUEST

This table is used internally to request the extract of initial loads asynchronously when the initial load extract job is enabled.

Table 41. EXTRACT_REQUEST

Name

Type

Size

Default

Keys

Not Null

Description

REQUEST_ID

BIGINT

PK

X

Unique identifier for a request.

SOURCE_NODE_ID

VARCHAR

50

default

PK

X

Unique identifier for the node that will be the source of the extract.

NODE_ID

VARCHAR

50

X

The node_id of the batch being loaded.

QUEUE

VARCHAR

128

The channel queue name of the batch being loaded.

STATUS

CHAR

2

NE, OK

START_BATCH_ID

BIGINT

X

A load can be split across multiple batches. This is the first of N batches the load will be split across.

END_BATCH_ID

BIGINT

X

This is the last of N batches the load will be split across.

TRIGGER_ID

VARCHAR

128

X

Unique identifier for a trigger associated with the extract request.

ROUTER_ID

VARCHAR

50

X

Unique description of the router associated with the extract request.

LOAD_ID

BIGINT

The load id associated with the extract request.

TABLE_NAME

VARCHAR

255

The table name for this extract request

EXTRACTED_ROWS

BIGINT

0

X

The rows in this table that have been extracted to target

EXTRACTED_MILLIS

BIGINT

0

X

The time spent extracting this table

TRANSFERRED_ROWS

BIGINT

0

X

The rows in this table that have been transferred to target

TRANSFERRED_MILLIS

BIGINT

0

X

The time spent transfering this table

LAST_TRANSFERRED_BATCH_ID

BIGINT

The last batch id that was successfully transferred to the target

LOADED_ROWS

BIGINT

0

X

The rows in this table that have been loaded to the target

LOADED_MILLIS

BIGINT

0

X

The time spent loading this table

LAST_LOADED_BATCH_ID

BIGINT

The last batch id that was successfully loaded on the target

TOTAL_ROWS

BIGINT

The rows in this table to be extracted

LOADED_TIME

TIMESTAMP

Timestamp when this table was loaded.

PARENT_REQUEST_ID

BIGINT

0

X

Parent request_id that will actually handle the extract for this request.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a process last updated this entry.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

A.20. FILE_INCOMING

As files are loaded from another node the file and source node are captured here for file sync to use to prevent file ping backs in bidirectional file synchronization.

Table 42. FILE_INCOMING

Name

Type

Size

Default

Keys

Not Null

Description

RELATIVE_DIR

VARCHAR

255

PK

X

The path to the file starting at the base_dir and excluding the file name itself.

FILE_NAME

VARCHAR

260

PK

X

The name of the file that has been loaded.

LAST_EVENT_TYPE

CHAR

1

X

The type of event that caused the file to be loaded from another node. 'C' is for create, 'M' is for modified, and 'D' is for deleted.

NODE_ID

VARCHAR

50

X

The node_id of the source of the batch being loaded.

FILE_MODIFIED_TIME

BIGINT

The last modified time of the file at the time the file was loaded.

A.21. FILE_SNAPSHOT

Table used to capture file changes. Updates to the table are captured and routed according to the configured file trigger routers.

Table 43. FILE_SNAPSHOT

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_ID

VARCHAR

128

PK

X

The id of the trigger that caused this snapshot to be taken.

ROUTER_ID

VARCHAR

50

PK

X

The id of the router that caused this snapshot to be taken.

RELATIVE_DIR

VARCHAR

255

PK

X

The path to the file starting at the base_dir

FILE_NAME

VARCHAR

260

PK

X

The name of the file that changed.

CHANNEL_ID

VARCHAR

128

filesync

X

The channel_id of the channel that data changes will flow through.

RELOAD_CHANNEL_ID

VARCHAR

128

filesync_reload

X

The channel_id of the channel that data changes will flow through.

LAST_EVENT_TYPE

CHAR

1

X

The type of event captured by this entry. 'C' is for create, 'M' is for modified, and 'D' is for deleted.

CRC32_CHECKSUM

BIGINT

File checksum. Can be used to determine if file content has changed.

FILE_SIZE

BIGINT

The size in bytes of the file at the time this change was detected.

FILE_MODIFIED_TIME

BIGINT

The last modified time of the file at the time this change was detected.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

A.22. FILE_TRIGGER

This table defines files or sets of files for which changes will be captured for file synchronization

Table 44. FILE_TRIGGER

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_ID

VARCHAR

128

PK

X

Unique identifier for a trigger.

CHANNEL_ID

VARCHAR

128

filesync

X

The channel_id of the channel that data changes will flow through.

RELOAD_CHANNEL_ID

VARCHAR

128

filesync_reload

X

The channel_id of the channel that will be used for reloads.

BASE_DIR

VARCHAR

255

X

The base directory on the client that will be synchronized.

RECURSE

TINYINT

1

1

X

Whether to synchronize child directories.

INCLUDES_FILES

VARCHAR

255

Wildcard-enabled, comma-separated list of file to include in synchronization.

EXCLUDES_FILES

VARCHAR

255

Wildcard-enabled, comma-separated list of file to exclude from synchronization.

SYNC_ON_CREATE

TINYINT

1

1

X

Whether to capture and send files when they are created.

SYNC_ON_MODIFIED

TINYINT

1

1

X

Whether to capture and send files when they are modified.

SYNC_ON_DELETE

TINYINT

1

1

X

Whether to capture and remove files when they are deleted.

SYNC_ON_CTL_FILE

TINYINT

1

0

X

Combined with sync_on_create, determines whether to capture and send files when a matching control file exists. The control file is a file of the same name with a '.ctl' extension appended to the end.

DELETE_AFTER_SYNC

TINYINT

1

0

X

Determines whether to delete the file after it has synced successfully.

BEFORE_COPY_SCRIPT

LONGVARCHAR

A bsh script that is run right before the file copy.

AFTER_COPY_SCRIPT

LONGVARCHAR

A bsh script that is run right after the file copy.

CREATE_TIME

TIMESTAMP

X

Timestamp of when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp of when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for file_trigger

A.23. FILE_TRIGGER_ROUTER

Maps a file trigger to a router.

Table 45. FILE_TRIGGER_ROUTER

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_ID

VARCHAR

128

PK FK

X

The id of a file trigger.

ROUTER_ID

VARCHAR

50

PK FK

X

The id of a router.

ENABLED

TINYINT

1

1

X

Indicates whether this file trigger router is enabled or not.

INITIAL_LOAD_ENABLED

TINYINT

1

1

X

Indicates whether this file trigger should be initial loaded.

TARGET_BASE_DIR

VARCHAR

255

The base directory on the destination that files will be synchronized to.

CONFLICT_STRATEGY

VARCHAR

128

source_wins

X

The strategy to employ when a file has been modified at both the client and the server. Possible values are: source_wins, target_wins, manual

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for file_trigger_router

A.24. GROUPLET

Deprecated in 3.9. This tables defines named groups to which nodes can belong to based on their external id. Grouplets are used to designate that synchronization should only affect an explicit subset of nodes in a node group.

Table 46. GROUPLET

Name

Type

Size

Default

Keys

Not Null

Description

GROUPLET_ID

VARCHAR

50

PK

X

Unique identifier for the grouplet.

GROUPLET_LINK_POLICY

CHAR

1

I

X

Specified whether the external ids in the grouplet_link are included in the group or excluded from the grouplet. In the case of excluded, the grouplet starts with all external ids and removes the excluded ones listed. Use 'I' for inclusive and 'E' for exclusive.

DESCRIPTION

VARCHAR

255

A description of this grouplet.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

Deprecated in 3.9. This tables defines nodes belong to a grouplet based on their external.id

Table 47. GROUPLET_LINK

Name

Type

Size

Default

Keys

Not Null

Description

GROUPLET_ID

VARCHAR

50

PK FK

X

Unique identifier for the grouplet.

EXTERNAL_ID

VARCHAR

255

PK

X

Provides a means to select the nodes that belong to a grouplet.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

A.26. INCOMING_BATCH

The incoming_batch is used for tracking the status of loading an outgoing_batch from another node. Data is loaded and commited at the batch level. The status of the incoming_batch is either successful (OK) or error (ER).

Table 48. INCOMING_BATCH

Name

Type

Size

Default

Keys

Not Null

Description

BATCH_ID

BIGINT

50

PK

X

The id of the outgoing_batch that is being loaded.

NODE_ID

VARCHAR

50

PK

X

The node_id of the source of the batch being loaded.

CHANNEL_ID

VARCHAR

128

The channel_id of the batch being loaded.

STATUS

CHAR

2

The current status of the batch can be loading (LD), successfully loaded (OK), in error (ER) or skipped (SK)

ERROR_FLAG

TINYINT

1

0

A flag that indicates that this batch was in error during the last synchornization attempt.

SQL_STATE

VARCHAR

10

For a status of error (ER), this is the XOPEN or SQL 99 SQL State.

SQL_CODE

INTEGER

0

X

For a status of error (ER), this is the error code from the database that is specific to the vendor.

SQL_MESSAGE

LONGVARCHAR

For a status of error (ER), this is the error message that describes the error.

LAST_UPDATE_HOSTNAME

VARCHAR

255

The host name of the process that last did work on this batch.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a process last updated this entry.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

SUMMARY

VARCHAR

255

A high level summary of what is included in a batch, often a list of table names.

IGNORE_COUNT

INTEGER

0

X

The number of times a batch was ignored.

BYTE_COUNT

BIGINT

0

X

The number of bytes that were sent as part of this batch.

LOAD_FLAG

TINYINT

1

0

A flag that indicates that this batch is part of an initial load.

EXTRACT_COUNT

INTEGER

0

X

The number of times this an attempt to extract this batch occurred.

SENT_COUNT

INTEGER

0

X

The number of times this batch was sent. A batch can be sent multiple times if an ACK is not received.

LOAD_COUNT

INTEGER

0

X

The number of times an attempt to load this batch occurred.

RELOAD_ROW_COUNT

INTEGER

0

X

The number of reloads in the batch, which include rows from any extract transforms.

OTHER_ROW_COUNT

INTEGER

0

X

The number of SQL, BSH, and DDL events in the batch, which include rows from any extract transforms.

DATA_ROW_COUNT

INTEGER

0

X

The total number of rows in the batch, which include rows from any extract transforms.

EXTRACT_ROW_COUNT

INTEGER

0

X

The total number of rows extracted from the capture log.

LOAD_ROW_COUNT

INTEGER

0

X

The total number of rows loaded into the database, which includes rows from any load transforms or filters.

DATA_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts in the batch, which include rows from any extract transforms.

DATA_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates in the batch, which include rows from any extract transforms.

DATA_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes in the batch, which include rows from any extract transforms.

EXTRACT_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts extracted from the capture log.

EXTRACT_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates extracted from the capture log.

EXTRACT_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes extracted from the capture log.

LOAD_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts loaded into the database, which includes rows from any load transforms or filters.

LOAD_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates loaded into the database, which includes rows from any load transforms or filters.

LOAD_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes loaded into the database, which includes rows from any load transforms or filters.

NETWORK_MILLIS

INTEGER

0

X

The number of milliseconds spent transfering this batch across the network.

FILTER_MILLIS

INTEGER

0

X

The number of milliseconds spent in filters processing data.

LOAD_MILLIS

INTEGER

0

X

The number of milliseconds spent loading the data into the target database.

ROUTER_MILLIS

INTEGER

0

X

The number of milliseconds spent creating this batch.

EXTRACT_MILLIS

INTEGER

0

X

The number of milliseconds spent extracting the data out of the source database.

TRANSFORM_EXTRACT_MILLIS

INTEGER

0

X

The number of milliseconds spent transforming the data on the extract side.

TRANSFORM_LOAD_MILLIS

INTEGER

0

X

The number of milliseconds spent transforming the data on the load side.

LOAD_ID

BIGINT

An id that ties multiple batches together to identify them as being part of an initial load.

COMMON_FLAG

TINYINT

1

0

A flag that indicates that the data in this batch is shared by other nodes (they will have the same batch_id). Shared batches will be extracted to a common location.

FALLBACK_INSERT_COUNT

INTEGER

0

X

The number of times an update was turned into an insert because the data was not already in the target database.

FALLBACK_UPDATE_COUNT

INTEGER

0

X

The number of times an insert was turned into an update because a data row already existed in the target database.

CONFLICT_WIN_COUNT

INTEGER

0

X

Not implemented. The number of times a conflict was detected for a row and it was resolved as the winning row.

CONFLICT_LOSE_COUNT

INTEGER

0

X

Not implemented. The number of times a conflict was detected for a row and it was resolved as the losing row.

IGNORE_ROW_COUNT

INTEGER

0

X

The number of times a row was ignored.

MISSING_DELETE_COUNT

INTEGER

0

X

The number of times a delete did not affect the database because the row was already deleted.

SKIP_COUNT

INTEGER

0

X

The number of times a batch was sent and skipped because it had already been loaded according to incoming_batch.

FAILED_ROW_NUMBER

INTEGER

0

X

This numbered data event that failed as read from the CSV.

FAILED_LINE_NUMBER

INTEGER

0

X

The current line number in the CSV for this batch that failed.

FAILED_DATA_ID

BIGINT

0

X

For a status of error (ER), this is the data_id that was being processed when the batch failed.

BULK_LOADER_FLAG

TINYINT

1

0

A flag that indicates that this batch did or did not use the bulk loader.

A.27. INCOMING_ERROR

The captured data change that is in error for a batch. The user can tell the system what to do by updating the resolve columns. Entries in data_error are created when an incoming batch encounters an error.

Table 49. INCOMING_ERROR

Name

Type

Size

Default

Keys

Not Null

Description

BATCH_ID

BIGINT

50

PK

X

The id of the outgoing_batch that is being loaded.

NODE_ID

VARCHAR

50

PK

X

The node_id of the source of the batch being loaded. A node_id of -1 means that the batch was 'unrouted'.

FAILED_ROW_NUMBER

BIGINT

PK

X

The row number in the batch that encountered an error when loading.

FAILED_LINE_NUMBER

BIGINT

0

X

The current line number in the CSV for this batch that failed.

TARGET_CATALOG_NAME

VARCHAR

255

The catalog name for the table being loaded.

TARGET_SCHEMA_NAME

VARCHAR

255

The schema name for the table being loaded.

TARGET_TABLE_NAME

VARCHAR

255

X

The table name for the table being loaded.

EVENT_TYPE

CHAR

1

X

The type of event captured by this entry. For triggers, this is the change that occurred, which is 'I' for insert, 'U' for update, or 'D' for delete. Other events include: 'R' for reloading the entire table (or subset of the table) to the node; 'S' for running dynamic SQL at the node, which is used for adhoc administration.

BINARY_ENCODING

VARCHAR

10

HEX

X

The type of encoding the source system used for encoding binary data.

COLUMN_NAMES

LONGVARCHAR

X

The column names defined on the table. The column names are stored in comma-separated values (CSV) format.

PK_COLUMN_NAMES

LONGVARCHAR

X

The primary key column names defined on the table. The column names are stored in comma-separated values (CSV) format.

ROW_DATA

LONGVARCHAR

The row data from the batch as captured from the source. The column values are stored in comma-separated values (CSV) format.

OLD_DATA

LONGVARCHAR

The old row data prior to update from the batch as captured from the source. The column values are stored in CSV format.

CUR_DATA

LONGVARCHAR

The current row data that caused the error to occur. The column values are stored in CSV format.

RESOLVE_DATA

LONGVARCHAR

The capture data change from the user that is used instead of row_data. This is useful when resolving a conflict manually by specifying the data that should load.

RESOLVE_IGNORE

TINYINT

1

0

Indication from the user that the row_data should be ignored and the batch can continue loading with the next row.

CONFLICT_ID

VARCHAR

50

Unique identifier for the conflict detection setting that caused the error

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

A.28. JOB

Defines custom jobs that can be executed in the SymmetricDS instance.

Table 50. JOB

Name

Type

Size

Default

Keys

Not Null

Description

JOB_NAME

VARCHAR

50

PK

X

The unique name of the job

JOB_TYPE

VARCHAR

10

X

The type of the job. Valid values are 'sql', 'java' and 'bsh'. The system exclusively uses the 'built_in' job type.

REQUIRES_REGISTRATION

TINYINT

1

1

X

Does the engine need to be registered for this job to run?

JOB_EXPRESSION

LONGVARCHAR

The payload of the job. For bsh (beanshell) jobs, this should be bsh script. For Java jobs, this should be Java code of a class which implements the IJob interface. For SQL jobs, this should be a sql script.

DESCRIPTION

VARCHAR

255

An optional description of the job for users of the system.

DEFAULT_SCHEDULE

VARCHAR

50

The schedule to use if no schedule parameter is found. Overridden by job.jobname.period.time.ms or job.jobname.cron.

DEFAULT_AUTO_START

TINYINT

1

1

X

Determine if this job should auto start. Overridden by start.jobname.job.

NODE_GROUP_ID

VARCHAR

50

X

Target the job at a specific node group id. To target all groups, use the value of 'ALL'.

IS_CLUSTERED

TINYINT

1

0

X

Whether to acquire a cluster lock or not.

CREATE_BY

VARCHAR

50

The user who created this entry.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.29. LOAD_FILTER

A table that allows you to dynamically define filters using bsh.

Table 51. LOAD_FILTER

Name

Type

Size

Default

Keys

Not Null

Description

LOAD_FILTER_ID

VARCHAR

50

PK

X

The id of the load filter.

LOAD_FILTER_TYPE

VARCHAR

10

X

The type of load filter. Possible values include: BSH, JAVA, SQL

SOURCE_NODE_GROUP_ID

VARCHAR

50

X

The source node group for the filter.

TARGET_NODE_GROUP_ID

VARCHAR

50

X

The destination node group for the filter.

TARGET_CATALOG_NAME

VARCHAR

255

Optional name for the catalog the configured table is in.

TARGET_SCHEMA_NAME

VARCHAR

255

Optional name for the schema a configured table is in.

TARGET_TABLE_NAME

VARCHAR

255

The name of the target table that will trigger the bsh filter.

FILTER_ON_UPDATE

TINYINT

1

1

X

Whether or not the filter should apply on an update.

FILTER_ON_INSERT

TINYINT

1

1

X

Whether or not the filter should apply on an insert.

FILTER_ON_DELETE

TINYINT

1

1

X

Whether or not the filter should apply on a delete.

BEFORE_WRITE_SCRIPT

LONGVARCHAR

The script to apply before the write is completed.

AFTER_WRITE_SCRIPT

LONGVARCHAR

The script to apply after the write is completed.

BATCH_COMPLETE_SCRIPT

LONGVARCHAR

The script to apply on batch complete.

BATCH_COMMIT_SCRIPT

LONGVARCHAR

The script to apply on batch commit.

BATCH_ROLLBACK_SCRIPT

LONGVARCHAR

The script to apply on batch rollback.

HANDLE_ERROR_SCRIPT

LONGVARCHAR

The script to apply when data cannot be processed.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

LOAD_FILTER_ORDER

INTEGER

1

X

Specifies the order in which to apply load filters if more than one target operation occurs.

FAIL_ON_ERROR

TINYINT

1

0

X

Whether we should fail the batch if the filter fails.

A.30. LOCK

Contains semaphores that are set when processes run, so that only one server can run a process at a time. Enable this feature by using the cluster.lock.enabled parameter.

Table 52. LOCK

Name

Type

Size

Default

Keys

Not Null

Description

LOCK_ACTION

VARCHAR

50

PK

X

The process that needs a lock.

LOCK_TYPE

VARCHAR

50

X

Type of lock that indicates differently locking behavior. Types include cluster, exclusive, and shared. Cluster lock is used to allow one server to run at a time, but any process from the same server can overtake the lock, which avoids stalled processing. Exclusive lock is owned by one process, regardless of which server it is on, but another process can acquire the lock after lock_time is older than exclusive.lock.timeout.ms. Shared lock allows multiple processes to use the same lock, incrementing the shared_count, but requires no exclusive lock exists and prevents an exclusive lock.

LOCKING_SERVER_ID

VARCHAR

255

The name of the server that currently has a lock. This is typically a host name, but it can be overridden using the -Druntime.symmetric.cluster.server.id=name System property.

LOCK_TIME

TIMESTAMP

The time a lock is aquired. Use the cluster.lock.timeout.ms to specify a lock timeout period.

SHARED_COUNT

INTEGER

0

X

For a lock_type of SHARED, this is the number of processes sharing the same lock. After the shared_count drops to zero, a shared lock is removed.

SHARED_ENABLE

INTEGER

0

X

For a lock_type of SHARED, this flag set to 1 indicates that more processes can share the lock. If an exclusive lock is needed, the flag is set to 0 to prevent further shared locks from accumulating.

LAST_LOCK_TIME

TIMESTAMP

Timestamp when a process last updated this entry.

LAST_LOCKING_SERVER_ID

VARCHAR

255

The server id of the process that last did work on this batch.

A.31. MONITOR

Defines monitors that will run periodically to look for problems in the system.

Table 53. MONITOR

Name

Type

Size

Default

Keys

Not Null

Description

MONITOR_ID

VARCHAR

128

PK

X

Unique identifier for a monitor.

NODE_GROUP_ID

VARCHAR

50

ALL

X

Target a specific node group to run this monitor. Target all groups, use a value of 'ALL'.

EXTERNAL_ID

VARCHAR

255

ALL

X

Target a specific node by its external ID to run this monitor. Target all nodes, use a value of 'ALL'.

TYPE

VARCHAR

50

X

Monitor type to execute. Built-in types are cpu, disk, memory, batchError, batchUnsent, dataGap, dataUnrouted, and log.

EXPRESSION

LONGVARCHAR

An expression used by the monitor to set options specific to the monitor type.

THRESHOLD

BIGINT

0

X

The minimum value returned when the monitor runs that will cause a monitor event to be recorded.

RUN_PERIOD

INTEGER

0

X

Run this monitor periodically every number of seconds.

RUN_COUNT

INTEGER

0

X

Average the value across a number of runs before checking threshold.

SEVERITY_LEVEL

INTEGER

0

X

DISPLAY_ORDER

INTEGER

0

X

Numerical ascending order of monitors to display to user.

IS_INSIGHT

TINYINT

1

0

X

Insights are optional recommendations for changing the system settings, while other monitors are for errors that must be resolved.

IS_PINNED

TINYINT

1

0

X

A pinned monitor will be displayed to the user even if there are no unresolved events for it.

ENABLED

TINYINT

1

0

X

Whether or not this monitor is enabled for execution.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.32. MONITOR_EVENT

Records an event of when a system problem occurred.

Table 54. MONITOR_EVENT

Name

Type

Size

Default

Keys

Not Null

Description

MONITOR_ID

VARCHAR

128

PK

X

Unique identifier for a monitor that caused the event.

NODE_ID

VARCHAR

50

PK

X

Unique identifier for the node that created the event.

EVENT_TIME

TIMESTAMP

PK

X

Timestamp when the event was created.

HOST_NAME

VARCHAR

60

Host name of the node that created the event.

TYPE

VARCHAR

50

X

Monitor type that detected the value recorded.

THRESHOLD

BIGINT

0

X

Minimum value for the monitor to cause an event.

EVENT_VALUE

BIGINT

0

X

Actual value detected by the monitor.

EVENT_COUNT

INTEGER

0

X

Number of times this event has occurred and been updated.

SEVERITY_LEVEL

INTEGER

0

X

Severity level configured for the monitor.

IS_RESOLVED

TINYINT

0

X

Whether an event is resolved because its value dropped below the threshold.

IS_NOTIFIED

TINYINT

0

X

Whether a notification was run.

IS_INSIGHT

TINYINT

0

X

Insights are optional recommendations for changing the system settings, while other monitor events are for errors that must be resolved.

NOT_BEFORE

TIMESTAMP

Timestamp before which the recommendation should not be displayed in the web console.

APPROVED_OPTION

INTEGER

0

X

Identifier for the option that was selected when approving the recommendation.

APPROVED_BY

VARCHAR

50

The user who approved the event.

IS_APPROVAL_PROCESSED

TINYINT

0

X

Whether the recommendation’s approval has been processed.

DETAILS

LONGVARCHAR

JSON formatted details about the event

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when the event was last updated.

A.33. NODE

Representation of an instance of SymmetricDS that synchronizes data with one or more additional nodes. Each node has a unique identifier (nodeId) that is used when communicating, as well as a domain-specific identifier (externalId) that provides context within the local system.

Table 55. NODE

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

A unique identifier for a node.

NODE_GROUP_ID

VARCHAR

50

X

The node group that this node belongs to, such as 'store'.

EXTERNAL_ID

VARCHAR

255

X

A domain-specific identifier for context within the local system. For example, the retail store number.

SYNC_ENABLED

TINYINT

1

0

Indicates whether this node should be sent synchronization. Disabled nodes are ignored by the triggers, so no entries are made in data_event for the node.

SYNC_URL

VARCHAR

255

The URL to contact the node for synchronization.

SCHEMA_VERSION

VARCHAR

50

The version of the database schema this node manages. Useful for specifying synchronization by version.

SYMMETRIC_VERSION

VARCHAR

50

The version of SymmetricDS running at this node.

CONFIG_VERSION

VARCHAR

50

The version of configuration running at this node.

DATABASE_TYPE

VARCHAR

50

The database product name at this node as reported by JDBC.

DATABASE_VERSION

VARCHAR

50

The database product version at this node as reported by JDBC.

DATABASE_NAME

VARCHAR

50

The database product name identified by SymmetricDS.

BATCH_TO_SEND_COUNT

INTEGER

0

The number of outgoing batches that have not yet been sent. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

BATCH_IN_ERROR_COUNT

INTEGER

0

The number of outgoing batches that are in error at this node. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

BATCH_LAST_SUCCESSFUL

TIMESTAMP

The last incoming or outgoing batch that was marked OK that is not on the config, heartbeat, or monitor channels.

DATA_ROWS_TO_SEND_COUNT

INTEGER

0

The number of of rows to replicated in batches. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

DATA_ROWS_LOADED_COUNT

INTEGER

0

The number of rows that has been successfully replicated since last purge. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

OLDEST_LOAD_TIME

TIMESTAMP

The oldest recorded sync time since the last purge.

MOST_RECENT_ACTIVE_TABLE

VARCHAR

255

The most active table to sync recently (time based on last 5 minutes and doubles until a match is found or until 80 min.))

PURGE_OUTGOING_LAST_RUN_MS

BIGINT

0

The number of milliseconds that the purge job took on the last run. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

PURGE_OUTGOING_LAST_FINISH

TIMESTAMP

The time the purge outgoing last finished succesfully. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

PURGE_OUTGOING_AVERAGE_MS

BIGINT

The average time the purge outgoing has been run. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

ROUTING_LAST_RUN_MS

BIGINT

The time it took for the the last routing job to finish succesfully. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

ROUTING_LAST_FINISH

TIMESTAMP

The time the routing job last finished succesfully. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

ROUTING_AVERAGE_RUN_MS

BIGINT

0

The number of milliseconds that the routing job averages. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

SYM_DATA_SIZE

BIGINT

The current size of the sym_data table. This field is updated as part of the heartbeat job if the heartbeat.update.node.with.batch.status property is set to true.

CREATED_AT_NODE_ID

VARCHAR

50

The node_id of the node where this node was created. This is typically filled automatically with the node_id found in node_identity where registration was opened for the node.

DEPLOYMENT_TYPE

VARCHAR

50

An indicator as to the type of SymmetricDS software that is running. Possible values are, but not limited to: engine, standalone, war, professional, mobile

DEPLOYMENT_SUB_TYPE

VARCHAR

50

An indicator as to the deployment sub type of the node. Possible values are, but not limited to: load-only

A.34. NODE_CHANNEL_CTL

Used to ignore or suspend a channel. A channel that is ignored will have its data_events batched and they will immediately be marked as 'OK' without sending them. A channel that is suspended is skipped when batching data_events.

Table 56. NODE_CHANNEL_CTL

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

Unique identifier for a node.

CHANNEL_ID

VARCHAR

128

PK

X

The name of the channel_id that is being controlled.

SUSPEND_ENABLED

TINYINT

1

0

Indicates if this channel is suspended, which prevents batches from being sent, although new batches can still be created.

IGNORE_ENABLED

TINYINT

1

0

Indicates if this channel is ignored, which marks batches with a status of OK like they were actually processed.

LAST_EXTRACT_TIME

TIMESTAMP

Record the last time data was extract for a node and a channel.

A.35. NODE_COMMUNICATION

This table is used to coordinate communication with other nodes.

Table 57. NODE_COMMUNICATION

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

Unique identifier for a node.

QUEUE

VARCHAR

25

default

PK

X

The queue name to use in relation to the channel.

COMMUNICATION_TYPE

VARCHAR

10

PK

X

The type of communication that is taking place with this node. Valid values are: PULL, PUSH

LOCK_TIME

TIMESTAMP

The timestamp when this node was locked

LOCKING_SERVER_ID

VARCHAR

255

The name of the server that currently has a pull lock for the node. This is typically a host name, but it can be overridden using the -Druntime.symmetric.cluster.server.id=name System property.

LAST_LOCK_TIME

TIMESTAMP

The timestamp when this node was last locked

LAST_LOCK_MILLIS

BIGINT

0

The amount of time the last communication took.

SUCCESS_COUNT

BIGINT

0

The number of successive successful communication attempts.

FAIL_COUNT

BIGINT

0

The number of successive failed communication attempts.

SKIP_COUNT

BIGINT

0

The number of skipped communication attempts.

TOTAL_SUCCESS_COUNT

BIGINT

0

The total number of successful communication attempts with the node.

TOTAL_FAIL_COUNT

BIGINT

0

The total number of failed communication attempts with the node.

TOTAL_SUCCESS_MILLIS

BIGINT

0

The total amount of time spent during successful communication attempts with the node.

TOTAL_FAIL_MILLIS

BIGINT

0

The total amount of time spent during failed communication attempts with the node.

BATCH_TO_SEND_COUNT

BIGINT

0

The number of batches this node has queued for pull.

NODE_PRIORITY

INTEGER

0

Used to order nodes when initiating a pull operation. Can be used to move a node to the top of the list to pull from it as quickly as possible.

A.36. NODE_GROUP

A category of Nodes that synchronizes data with one or more NodeGroups. A common use of NodeGroup is to describe a level in a hierarchy of data synchronization.

Table 58. NODE_GROUP

Name

Type

Size

Default

Keys

Not Null

Description

NODE_GROUP_ID

VARCHAR

50

PK

X

Unique identifier for a node group, usually named something meaningful, like 'store' or 'warehouse'.

DESCRIPTION

VARCHAR

255

A description of this node group.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.37. NODE_GROUP_CHANNEL_WND

An optional window of time for which a node group and channel will extract and send data.

Table 59. NODE_GROUP_CHANNEL_WND

Name

Type

Size

Default

Keys

Not Null

Description

NODE_GROUP_ID

VARCHAR

50

PK

X

The node_group_id that this window applies to.

CHANNEL_ID

VARCHAR

128

PK

X

The channel_id that this window applies to.

START_TIME

TIMESTAMP

2

PK

X

The start time for the active window.

END_TIME

TIMESTAMP

2

PK

X

The end time for the active window. Note that if the end_time is less than the start_time then the window crosses a day boundary.

ENABLED

TINYINT

1

0

X

Enable this window. If this is set to '0' then this window is ignored.

A source node_group sends its data updates to a target NodeGroup using a pull, push, or custom technique.

Table 60. NODE_GROUP_LINK

Name

Type

Size

Default

Keys

Not Null

Description

SOURCE_NODE_GROUP_ID

VARCHAR

50

PK FK

X

The node group where data changes should be captured.

TARGET_NODE_GROUP_ID

VARCHAR

50

PK FK

X

The node group where data changes will be sent.

DATA_EVENT_ACTION

CHAR

1

W

X

The notification scheme used to send data changes to the target node group. (P = Push, W = Wait for Pull, B = Both Push and Wait for Pull (control from channel), R = Route-Only)

SYNC_CONFIG_ENABLED

TINYINT

1

1

X

Indicates whether configuration that has changed should be synchronized to target nodes on this link. Affects all SymmetricDS configuration tables except for sym_extract_request, sym_file_snapshot, sym_monitor_event, sym_node, sym_node_host, sym_node_security, sym_table_reload_request, and sym_table_reload_status.

SYNC_SQL_ENABLED

TINYINT

1

1

X

Whether SQL/DDL events are routed and synced on this link.

IS_REVERSIBLE

TINYINT

1

0

X

Indicates if communication can work in reverse as specified on the channel. A reversible push link can be overridden to pull, and a reversible pull link can be overridden to push on the channel.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.39. NODE_HOST

Representation of an physical workstation or server that is hosting the SymmetricDS software. In a clustered environment there may be more than one entry per node in this table.

Table 61. NODE_HOST

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

A unique identifier for a node.

HOST_NAME

VARCHAR

60

PK

X

The host name of a workstation or server. If more than one instance of SymmetricDS runs on the same server, then this value can be a 'server id' specified by -Druntime.symmetric.cluster.server.id

INSTANCE_ID

VARCHAR

60

A unique identifer generated by each installation of SymmetricDS. Used to to safe-guard against multi-access to the sym tables.

IP_ADDRESS

VARCHAR

50

The ip address for the host.

OS_USER

VARCHAR

50

The user SymmetricDS is running under

OS_NAME

VARCHAR

50

The name of the OS

OS_ARCH

VARCHAR

50

The hardware architecture of the OS

OS_VERSION

VARCHAR

50

The version of the OS

AVAILABLE_PROCESSORS

INTEGER

0

The number of processors available to use.

FREE_MEMORY_BYTES

BIGINT

0

The amount of free memory available to the JVM.

TOTAL_MEMORY_BYTES

BIGINT

0

The amount of total memory available to the JVM.

MAX_MEMORY_BYTES

BIGINT

0

The max amount of memory available to the JVM.

JAVA_VERSION

VARCHAR

50

The version of java that SymmetricDS is running as.

JAVA_VENDOR

VARCHAR

255

The vendor of java that SymmetricDS is running as.

JDBC_VERSION

VARCHAR

255

The verision of the JDBC driver that is being used.

SYMMETRIC_VERSION

VARCHAR

50

The version of SymmetricDS running at this node.

TIMEZONE_OFFSET

VARCHAR

6

The time zone offset in RFC822 format at the time of the last heartbeat.

HEARTBEAT_TIME

TIMESTAMP

The last timestamp when the node sent a heartbeat, which is attempted every ten minutes by default.

LAST_RESTART_TIME

TIMESTAMP

X

Timestamp when this instance was last restarted.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

A.40. NODE_HOST_CHANNEL_STATS

Table 62. NODE_HOST_CHANNEL_STATS

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

A unique identifier for a node.

HOST_NAME

VARCHAR

60

PK

X

The host name of a workstation or server. If more than one instance of SymmetricDS runs on the same server, then this value can be a 'server id' specified by -Druntime.symmetric.cluster.server.id

CHANNEL_ID

VARCHAR

128

PK

X

The channel_id of the channel that data changes will flow through.

START_TIME

TIMESTAMP

2

PK

X

The start time for the period which this row represents.

END_TIME

TIMESTAMP

2

PK

X

The end time for the period which this row represents.

DATA_ROUTED

BIGINT

0

Indicate the number of data rows that have been routed during this period.

DATA_UNROUTED

BIGINT

0

The amount of data that has not yet been routed at the time this stats row was recorded.

DATA_EVENT_INSERTED

BIGINT

0

Indicate the number of data rows that have been routed during this period.

DATA_EXTRACTED

BIGINT

0

The number of data rows that were extracted during this time period.

DATA_BYTES_EXTRACTED

BIGINT

0

The number of bytes that were extracted during this time period.

DATA_EXTRACTED_ERRORS

BIGINT

0

The number of errors that occurred during extraction during this time period.

DATA_BYTES_SENT

BIGINT

0

The number of bytes that were sent during this time period.

DATA_SENT

BIGINT

0

The number of rows that were sent during this time period.

DATA_SENT_ERRORS

BIGINT

0

The number of errors that occurred while sending during this time period.

DATA_LOADED

BIGINT

0

The number of rows that were loaded during this time period.

DATA_BYTES_LOADED

BIGINT

0

The number of bytes that were loaded during this time period.

DATA_LOADED_ERRORS

BIGINT

0

The number of errors that occurred while loading during this time period.

DATA_LOADED_OUTGOING

BIGINT

0

The number of rows that were acknowledged as loaded by another node during this time period.

DATA_BYTES_LOADED_OUTGOING

BIGINT

0

The number of bytes that were acknowledged as loaded by another node during this time period.

DATA_LOADED_OUTGOING_ERRORS

BIGINT

0

The number of errors that occurred while loading to another node during this time period.

A.41. NODE_HOST_JOB_STATS

Table 63. NODE_HOST_JOB_STATS

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

A unique identifier for a node.

HOST_NAME

VARCHAR

60

PK

X

The host name of a workstation or server. If more than one instance of SymmetricDS runs on the same server, then this value can be a 'server id' specified by -Druntime.symmetric.cluster.server.id

JOB_NAME

VARCHAR

50

PK

X

The name of the job.

START_TIME

TIMESTAMP

2

PK

X

The start time for the period which this row represents.

END_TIME

TIMESTAMP

2

PK

X

The end time for the period which this row represents.

PROCESSED_COUNT

BIGINT

0

The number of items that were processed during the job run.

ERROR_FLAG

TINYINT

1

0

A flag that indicates that this job was in error.

ERROR_MESSAGE

LONGVARCHAR

The error message that occurred.

TARGET_NODE_ID

VARCHAR

50

The target node the job is working on.

TARGET_NODE_COUNT

INTEGER

0

The number of target nodes the job is working on.

A.42. NODE_HOST_STATS

Table 64. NODE_HOST_STATS

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK

X

A unique identifier for a node.

HOST_NAME

VARCHAR

60

PK

X

The host name of a workstation or server. If more than one instance of SymmetricDS runs on the same server, then this value can be a 'server id' specified by -Druntime.symmetric.cluster.server.id

START_TIME

TIMESTAMP

2

PK

X

The end time for the period which this row represents.

END_TIME

TIMESTAMP

2

PK

X

RESTARTED

BIGINT

0

X

Indicate that a restart occurred during this period.

NODES_PULLED

BIGINT

0

Number of nodes that pulled from this host after passing checks for authentication and concurrency.

TOTAL_NODES_PULL_TIME

BIGINT

0

Total time spent handling pull requests on this host in milliseconds.

NODES_PUSHED

BIGINT

0

Number of nodes that pushed to this host after passing checks for authentication and concurrency.

TOTAL_NODES_PUSH_TIME

BIGINT

0

Total time spent handling push requests on this host in milliseconds.

NODES_REJECTED

BIGINT

0

Number of attempts by nodes to push or pull with this host that were rejected because no concurrent workers were available or the push reservation was not found.

NODES_REGISTERED

BIGINT

0

Number of nodes that registered with this host, either a first-time registration or a subsequent re-registration.

NODES_LOADED

BIGINT

0

Number of full or partial load requests that were created on this host to send loads to target nodes.

NODES_DISABLED

BIGINT

0

Number of offline nodes that had their sync enabled flag set to false on this host by the watch dog job.

PURGED_DATA_ROWS

BIGINT

0

PURGED_DATA_EVENT_ROWS

BIGINT

0

PURGED_STRANDED_DATA_ROWS

BIGINT

0

PURGED_STRANDED_EVENT_ROWS

BIGINT

0

PURGED_EXPIRED_DATA_ROWS

BIGINT

0

PURGED_BATCH_OUTGOING_ROWS

BIGINT

0

PURGED_BATCH_INCOMING_ROWS

BIGINT

0

TRIGGERS_CREATED_COUNT

BIGINT

TRIGGERS_REBUILT_COUNT

BIGINT

TRIGGERS_REMOVED_COUNT

BIGINT

A.43. NODE_IDENTITY

After registration, this table will have one row representing the identity of the node. For a root node, the row is entered by the user.

Table 65. NODE_IDENTITY

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK FK

X

Unique identifier for a node.

A.44. NODE_SECURITY

Security features like node passwords and open registration flag are stored in the node_security table.

Table 66. NODE_SECURITY

Name

Type

Size

Default

Keys

Not Null

Description

NODE_ID

VARCHAR

50

PK FK

X

Unique identifier for a node.

NODE_PASSWORD

VARCHAR

50

X

The password used by the node to prove its identity during synchronization.

REGISTRATION_ENABLED

TINYINT

1

0

Indicates whether registration is open for this node. Re-registration may be forced for a node if this is set back to '1' in a parent database for the node_id that should be re-registred.

REGISTRATION_TIME

TIMESTAMP

The timestamp when this node was last registered.

REGISTRATION_NOT_BEFORE

TIMESTAMP

Allow registration beginning at this timestamp.

REGISTRATION_NOT_AFTER

TIMESTAMP

Allow registration until this timestamp.

INITIAL_LOAD_ENABLED

TINYINT

1

0

Indicates whether an initial load will be sent to this node.

INITIAL_LOAD_TIME

TIMESTAMP

The timestamp when an initial load was started for this node.

INITIAL_LOAD_END_TIME

TIMESTAMP

The timestamp when an initial load was completed for this node.

INITIAL_LOAD_ID

BIGINT

A reference to the load_id in outgoing_batch for the last load that occurred.

INITIAL_LOAD_CREATE_BY

VARCHAR

255

The user that created the initial load. A null value means that the system created the batch.

REV_INITIAL_LOAD_ENABLED

TINYINT

1

0

Indicates that this node should send a reverse initial load.

REV_INITIAL_LOAD_TIME

TIMESTAMP

The timestamp when this node last sent an initial load.

REV_INITIAL_LOAD_ID

BIGINT

A reference to the load_id in outgoing_batch for the last reverse load that occurred.

REV_INITIAL_LOAD_CREATE_BY

VARCHAR

255

The user that created the reverse initial load. A null value means that the system created the batch.

FAILED_LOGINS

TINYINT

0

Number of failed login attempts

CREATED_AT_NODE_ID

VARCHAR

50

The node_id of the node where this node was created. This is typically filled automatically with the node_id found in node_identity where registration was opened for the node.

A.45. NOTIFICATION

Defines what notification to send when a monitor detects a problem in the system.

Table 67. NOTIFICATION

Name

Type

Size

Default

Keys

Not Null

Description

NOTIFICATION_ID

VARCHAR

128

PK

X

Unique identifier for a notification.

NODE_GROUP_ID

VARCHAR

50

ALL

X

Target a specific node group to run this notification. To target all groups, use a value of 'ALL'.

EXTERNAL_ID

VARCHAR

255

ALL

X

Target a specific node by its external ID to run this notification. To target all nodes, use a value of 'ALL'.

SEVERITY_LEVEL

INTEGER

0

X

Look for monitor events using this severity level or above. To match all severity levels, use a value of 0.

TYPE

VARCHAR

50

X

Notification type that will send a message. Built-in types are mail and log.

EXPRESSION

LONGVARCHAR

An expression used by the notification to set options specific to the notification type.

ENABLED

TINYINT

1

0

X

Whether or not this notication is enabled.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.46. OUTGOING_BATCH

Used for tracking the sending a collection of data to a node in the system. A new outgoing_batch is created and given a status of 'NE'. After sending the outgoing_batch to its target node, the status becomes 'SE'. The node responds with either a success status of 'OK' or an error status of 'ER'. An error while sending to the node also results in an error status of 'ER' regardless of whether the node sends that acknowledgement.

Table 68. OUTGOING_BATCH

Name

Type

Size

Default

Keys

Not Null

Description

BATCH_ID

BIGINT

PK

X

A unique id for the batch.

NODE_ID

VARCHAR

50

PK

X

The node that this batch is targeted at.

CHANNEL_ID

VARCHAR

128

The channel that this batch is part of.

STATUS

CHAR

2

The current status of a batch can be routing (RT), requested to be extracted in the background (RQ), newly created and ready for replication (NE), being queried from the database (QY), sent to a node (SE), ready to be loaded (LD), resend requested because missing in staging (RS), acknowledged as successful (OK), ignored (IG), or in error (ER).

ERROR_FLAG

TINYINT

1

0

A flag that indicates that this batch was in error during the last synchornization attempt.

SQL_STATE

VARCHAR

10

For a status of error (ER), this is the XOPEN or SQL 99 SQL State.

SQL_CODE

INTEGER

0

X

For a status of error (ER), this is the error code from the database that is specific to the vendor.

SQL_MESSAGE

LONGVARCHAR

For a status of error (ER), this is the error message that describes the error.

LAST_UPDATE_HOSTNAME

VARCHAR

255

The host name of the process that last did work on this batch.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a process last updated this entry.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

SUMMARY

VARCHAR

255

A high level summary of what is included in a batch, often a list of table names.

IGNORE_COUNT

INTEGER

0

X

The number of times a batch was ignored.

BYTE_COUNT

BIGINT

0

X

The number of bytes that were sent as part of this batch.

LOAD_FLAG

TINYINT

1

0

A flag that indicates that this batch is part of an initial load.

EXTRACT_COUNT

INTEGER

0

X

The number of times this an attempt to extract this batch occurred.

SENT_COUNT

INTEGER

0

X

The number of times this batch was sent. A batch can be sent multiple times if an ACK is not received.

LOAD_COUNT

INTEGER

0

X

The number of times an attempt to load this batch occurred.

RELOAD_ROW_COUNT

INTEGER

0

X

The number of reloads in the batch, which include rows from any extract transforms.

OTHER_ROW_COUNT

INTEGER

0

X

The number of SQL, BSH, and DDL events in the batch, which include rows from any extract transforms.

DATA_ROW_COUNT

INTEGER

0

X

The total number of rows in the batch, which include rows from any extract transforms.

EXTRACT_ROW_COUNT

INTEGER

0

X

The total number of rows extracted from the capture log.

LOAD_ROW_COUNT

INTEGER

0

X

The total number of rows loaded into the database, which includes rows from any load transforms or filters.

DATA_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts in the batch, which include rows from any extract transforms.

DATA_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates in the batch, which include rows from any extract transforms.

DATA_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes in the batch, which include rows from any extract transforms.

EXTRACT_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts extracted from the capture log.

EXTRACT_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates extracted from the capture log.

EXTRACT_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes extracted from the capture log.

LOAD_INSERT_ROW_COUNT

INTEGER

0

X

The number of inserts loaded into the database, which includes rows from any load transforms or filters.

LOAD_UPDATE_ROW_COUNT

INTEGER

0

X

The number of updates loaded into the database, which includes rows from any load transforms or filters.

LOAD_DELETE_ROW_COUNT

INTEGER

0

X

The number of deletes loaded into the database, which includes rows from any load transforms or filters.

NETWORK_MILLIS

INTEGER

0

X

The number of milliseconds spent transfering this batch across the network.

FILTER_MILLIS

INTEGER

0

X

The number of milliseconds spent in filters processing data.

LOAD_MILLIS

INTEGER

0

X

The number of milliseconds spent loading the data into the target database.

ROUTER_MILLIS

INTEGER

0

X

The number of milliseconds spent creating this batch.

EXTRACT_MILLIS

INTEGER

0

X

The number of milliseconds spent extracting the data out of the source database.

TRANSFORM_EXTRACT_MILLIS

INTEGER

0

X

Not implemented. The number of milliseconds spent transforming the data on the extract side.

TRANSFORM_LOAD_MILLIS

INTEGER

0

X

Not implemented. The number of milliseconds spent transforming the data on the load side.

LOAD_ID

BIGINT

An id that ties multiple batches together to identify them as being part of an initial load.

COMMON_FLAG

TINYINT

1

0

A flag that indicates that the data in this batch is shared by other nodes (they will have the same batch_id). Shared batches will be extracted to a common location.

FALLBACK_INSERT_COUNT

INTEGER

0

X

The number of times an update was turned into an insert because the data was not already in the target database.

FALLBACK_UPDATE_COUNT

INTEGER

0

X

The number of times an insert was turned into an update because a data row already existed in the target database.

CONFLICT_WIN_COUNT

INTEGER

0

X

The number of times a conflict was detected for a row and it was resolved as the winning row.

CONFLICT_LOSE_COUNT

INTEGER

0

X

The number of times a conflict was detected for a row and it was resolved as the losing row.

IGNORE_ROW_COUNT

INTEGER

0

X

The number of times a row was ignored.

MISSING_DELETE_COUNT

INTEGER

0

X

The number of times a delete did not affect the database because the row was already deleted.

SKIP_COUNT

INTEGER

0

X

The number of times a batch was sent and skipped because it had already been loaded according to incoming_batch.

TOTAL_EXTRACT_MILLIS

INTEGER

0

X

Not implemented. The total number of milliseconds spent processing a batch on the extract side.

TOTAL_LOAD_MILLIS

INTEGER

0

X

Not implemented. The total number of milliseconds spent processing a batch on the load side.

EXTRACT_JOB_FLAG

TINYINT

1

0

A flag that indicates that this batch is going to be extracted by another job.

EXTRACT_START_TIME

TIMESTAMP

The time the batch started extracting from the source.

TRANSFER_START_TIME

TIMESTAMP

The time the batch started transfering across the network.

LOAD_START_TIME

TIMESTAMP

The time the batch started loading into the target.

FAILED_DATA_ID

BIGINT

0

X

For a status of error (ER), this is the data_id that was being processed when the batch failed.

FAILED_LINE_NUMBER

BIGINT

0

X

The current line number in the CSV for this batch that failed.

CREATE_BY

VARCHAR

255

The user that created the batch. A null value means that the system created the batch.

BULK_LOADER_FLAG

TINYINT

1

0

A flag that indicates that this batch did or did not use the bulk loader.

A.47. OUTGOING_ERROR

The captured data change that is in error for a batch. The user can tell the system what to do by updating the resolve columns. Entries in data_error are created when an outgoing batch encounters an error.

Table 69. OUTGOING_ERROR

Name

Type

Size

Default

Keys

Not Null

Description

BATCH_ID

BIGINT

50

PK

X

The id of the outgoing_batch that is being loaded.

NODE_ID

VARCHAR

50

PK

X

The node_id of the source of the batch being loaded. A node_id of -1 means that the batch was 'unrouted'.

FAILED_ROW_NUMBER

BIGINT

PK

X

The row number in the batch that encountered an error when loading.

FAILED_LINE_NUMBER

BIGINT

0

X

The current line number in the CSV for this batch that failed.

TARGET_CATALOG_NAME

VARCHAR

255

The catalog name for the table being loaded.

TARGET_SCHEMA_NAME

VARCHAR

255

The schema name for the table being loaded.

TARGET_TABLE_NAME

VARCHAR

255

X

The table name for the table being loaded.

EVENT_TYPE

CHAR

1

X

The type of event captured by this entry. For triggers, this is the change that occurred, which is 'I' for insert, 'U' for update, or 'D' for delete. Other events include: 'R' for reloading the entire table (or subset of the table) to the node; 'S' for running dynamic SQL at the node, which is used for adhoc administration.

BINARY_ENCODING

VARCHAR

10

HEX

X

The type of encoding the source system used for encoding binary data.

COLUMN_NAMES

LONGVARCHAR

X

The column names defined on the table. The column names are stored in comma-separated values (CSV) format.

PK_COLUMN_NAMES

LONGVARCHAR

X

The primary key column names defined on the table. The column names are stored in comma-separated values (CSV) format.

ROW_DATA

LONGVARCHAR

The row data from the batch as captured from the source. The column values are stored in comma-separated values (CSV) format.

OLD_DATA

LONGVARCHAR

The old row data prior to update from the batch as captured from the source. The column values are stored in CSV format.

CUR_DATA

LONGVARCHAR

The current row data that caused the error to occur. The column values are stored in CSV format.

RESOLVE_DATA

LONGVARCHAR

The capture data change from the user that is used instead of row_data. This is useful when resolving a conflict manually by specifying the data that should load.

RESOLVE_IGNORE

TINYINT

1

0

Indication from the user that the row_data should be ignored and the batch can continue loading with the next row.

CONFLICT_ID

VARCHAR

50

Unique identifier for the conflict detection setting that caused the error

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

A.48. PARAMETER

Provides a way to manage most SymmetricDS settings in the database.

Table 70. PARAMETER

Name

Type

Size

Default

Keys

Not Null

Description

EXTERNAL_ID

VARCHAR

255

PK

X

Target the parameter at a specific external id. To target all nodes, use the value of 'ALL.'

NODE_GROUP_ID

VARCHAR

50

PK

X

Target the parameter at a specific node group id. To target all groups, use the value of 'ALL.'

PARAM_KEY

VARCHAR

80

PK

X

The name of the parameter.

PARAM_VALUE

LONGVARCHAR

The value of the parameter.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.49. REGISTRATION_REDIRECT

Provides a way for a centralized registration server to redirect registering nodes to their prospective parent node in a multi-tiered deployment.

Table 71. REGISTRATION_REDIRECT

Name

Type

Size

Default

Keys

Not Null

Description

REGISTRANT_EXTERNAL_ID

VARCHAR

255

PK

X

Maps the external id of a registration request to a different parent node.

REGISTRATION_NODE_ID

VARCHAR

50

X

The node_id of the node that a registration request should be redirected to.

A.50. REGISTRATION_REQUEST

Audits when a node registers or attempts to register.

Table 72. REGISTRATION_REQUEST

Name

Type

Size

Default

Keys

Not Null

Description

NODE_GROUP_ID

VARCHAR

50

PK

X

The node group that this node belongs to, such as 'store'.

EXTERNAL_ID

VARCHAR

255

PK

X

A domain-specific identifier for context within the local system. For example, the retail store number.

STATUS

CHAR

2

X

The current status of the registration attempt. Valid statuses are RQ (requested), RR (redirected), RJ (rejected), ER (error), OK (successful).

HOST_NAME

VARCHAR

60

X

The host name of a workstation or server. If more than one instance of SymmetricDS runs on the same server, then this value can be a 'server id' specified by -Druntime.symmetric.cluster.server.id

IP_ADDRESS

VARCHAR

50

X

The ip address for the host.

ATTEMPT_COUNT

INTEGER

0

The number of registration attempts.

REGISTERED_NODE_ID

VARCHAR

50

A unique identifier for a node.

ERROR_MESSAGE

LONGVARCHAR

Record any errors or warnings that occurred when attempting to register.

CREATE_TIME

TIMESTAMP

2

PK

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

A.51. ROUTER

Configure a type of router from one node group to another. Note that routers are mapped to triggers through trigger_routers.

Table 73. ROUTER

Name

Type

Size

Default

Keys

Not Null

Description

ROUTER_ID

VARCHAR

50

PK

X

Unique description of a specific router

TARGET_CATALOG_NAME

VARCHAR

255

Optional name of catalog where a target table is located. If this field is unspecified, the catalog will be either the default catalog at the target node or the source_catalog_name from the trigger, depending on how use_source_catalog_schema is set on the router. Variables are substituted for $(sourceNodeId), $(sourceExternalId), $(sourceNodeGroupId), $(targetNodeId), $(targetExternalId), $(targetNodeGroupId), and $(none).

TARGET_SCHEMA_NAME

VARCHAR

255

Optional name of schema where a target table is located. If this field is unspecified, the schema will be either the default schema at the target node or the source_schema_name from the trigger, depending on how use_source_catalog_schema is set on the router. Variables are substituted for $(sourceNodeId), $(sourceExternalId), $(sourceNodeGroupId), $(targetNodeId), $(targetExternalId), $(targetNodeGroupId), and $(none).

TARGET_TABLE_NAME

VARCHAR

255

Optional name for a target table. Only use this if the target table name is different than the source.

SOURCE_NODE_GROUP_ID

VARCHAR

50

FK

X

Routers with this node_group_id will install triggers that are mapped to this router.

TARGET_NODE_GROUP_ID

VARCHAR

50

FK

X

The node_group_id for nodes to route data to. Note that routing can be further narrowed down by the configured router_type and router_expression.

ROUTER_TYPE

VARCHAR

50

default

X

The name of a specific type of router. Out of the box routers are 'default','column','bsh', 'subselect' and 'audit.' Custom routers can be configured as extension points.

ROUTER_EXPRESSION

LONGVARCHAR

An expression that is specific to the type of router that is configured in router_type. See the documentation for each router for more details.

SYNC_ON_UPDATE

TINYINT

1

1

X

Flag that indicates that this router should route updates.

SYNC_ON_INSERT

TINYINT

1

1

X

Flag that indicates that this router should route inserts.

SYNC_ON_DELETE

TINYINT

1

1

X

Flag that indicates that this router should route deletes.

USE_SOURCE_CATALOG_SCHEMA

TINYINT

1

1

X

Whether or not to assume that the target catalog/schema name should be the same as the source catalog/schema name. The target catalog or schema name will still override if not blank.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for router

A.52. SEQUENCE

A table that supports application level sequence numbering.

Table 74. SEQUENCE

Name

Type

Size

Default

Keys

Not Null

Description

SEQUENCE_NAME

VARCHAR

50

PK

X

Unique identifier of a specific sequence.

CURRENT_VALUE

BIGINT

0

X

The current value of the sequence.

INCREMENT_BY

INTEGER

1

X

Specify the interval between sequence numbers. This integer value can be any positive or negative integer, but it cannot be 0.

MIN_VALUE

BIGINT

1

X

Specify the minimum value of the sequence.

MAX_VALUE

BIGINT

9999999999

X

Specify the maximum value the sequence can generate.

CYCLE_FLAG

TINYINT

1

0

Indicate whether the sequence should automatically cycle once a boundary is hit.

CACHE_SIZE

INTEGER

0

X

Specify the number of sequence numbers to acquire and cache when one is requested.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

A.53. TABLE_GROUP

Group of tables to sync together as a dependent unit

Table 75. TABLE_GROUP

Name

Type

Size

Default

Keys

Not Null

Description

ID

VARCHAR

50

PK

X

A unique identifier

SOURCE_NODE_GROUP_ID

VARCHAR

50

X

The node group where data changes are captured.

TARGET_NODE_GROUP_ID

VARCHAR

50

X

The node group where data changes will be sent.

DESCRIPTION

VARCHAR

255

Optional description

WRITER_TYPE

CHAR

1

Type of writer: D=default, J=JSON

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

CREATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_BY

VARCHAR

50

The user who created this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.54. TABLE_GROUP_HIER

Relationships between tables for finding dependent data

Table 76. TABLE_GROUP_HIER

Name

Type

Size

Default

Keys

Not Null

Description

ID

VARCHAR

50

PK

X

A unique identifier

TABLE_GROUP_ID

VARCHAR

50

FK

X

Link to parent table_group

SOURCE_CATALOG_NAME

VARCHAR

255

Optional name for the catalog the configured table is in.

SOURCE_SCHEMA_NAME

VARCHAR

255

Optional name for the schema a configured table is in.

SOURCE_TABLE_NAME

VARCHAR

255

X

The name of the source table that will be grouped.

PARENT_ID

VARCHAR

50

A unique identifier of parent

RELATION_TYPE

CHAR

1

When writer type is non-default, this defines the structure of data in hierarchy when writing output: O=object, A=array

PRIMARY_COLUMN_NAMES

LONGVARCHAR

X

The primary key columns name on this table that uniquely identify a row, comma-separated

LOCAL_COLUMN_NAMES

LONGVARCHAR

The local key columns name on this table that reference a parent table, comma-separated

PARENT_COLUMN_NAMES

LONGVARCHAR

The parent column names on the parent that join with local columns, comma-separated

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

CREATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_BY

VARCHAR

50

The user who created this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

A.55. TABLE_RELOAD_REQUEST

This table acts as a means to queue up a reload of a specific table. Either the target or the source node may insert into this table to queue up a load. If the target node inserts into the table, then the row will be synchronized to the source node and the reload events will be queued up during routing.

Table 77. TABLE_RELOAD_REQUEST

Name

Type

Size

Default

Keys

Not Null

Description

TARGET_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the node to receive the table reload.

SOURCE_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the node that will be the source of the table reload.

TRIGGER_ID

VARCHAR

128

PK

X

Unique identifier for a trigger associated with the table reload. Note the trigger must be linked to the router.

ROUTER_ID

VARCHAR

50

PK

X

Unique description of the router associated with the table reload. Note the router must be linked to the trigger.

CREATE_TIME

TIMESTAMP

2

PK

X

Timestamp when this entry was created.

CREATE_TABLE

TINYINT

1

0

X

Flag that indicates that a table create script will be sent as part of the reload

DELETE_FIRST

TINYINT

1

0

X

Flag that indicates that the table will be deleted before loading.

RELOAD_SELECT

LONGVARCHAR

Overrides the initial load select.

BEFORE_CUSTOM_SQL

LONGVARCHAR

SQL Statement to run prior to loading the table

RELOAD_TIME

TIMESTAMP

The timestamp when the reload was started for this node.

LOAD_ID

BIGINT

An id that ties multiple batches together to identify them as being part of a load. The load id is generated by the system.

PROCESSED

TINYINT

1

0

X

Flag that indicates that this load was processed into batches.

CHANNEL_ID

VARCHAR

128

The channel that was specified as a source of the load.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user or the system last updated this entry.

A.56. TABLE_RELOAD_STATUS

This table holds the stats for a specific load id associated in table reload request.

Table 78. TABLE_RELOAD_STATUS

Name

Type

Size

Default

Keys

Not Null

Description

LOAD_ID

BIGINT

PK

X

An id that ties multiple batches together to identify them as being part of a load. The load id is generated by the system.

SOURCE_NODE_ID

VARCHAR

50

PK

X

Unique identifier for the node that will be the source of the table reload.

TARGET_NODE_ID

VARCHAR

50

X

Unique identifier for the node to receive the table reload.

START_TIME

TIMESTAMP

Timestamp when this entry was created.

END_TIME

TIMESTAMP

Timestamp when this entry was created.

COMPLETED

TINYINT

1

0

X

Flag that indicates that this load was completed.

CANCELLED

TINYINT

1

0

X

Flag that indicates that this load was cancelled. When cancelled it will also be marked completed

FULL_LOAD

TINYINT

1

0

X

Flag that indicates that this load is a full load for all tables, otherwise it is a partial load.

START_DATA_BATCH_ID

BIGINT

A load can be split across multiple batches. This is the first of N batches the load will be split across.

END_DATA_BATCH_ID

BIGINT

This is the last of N batches the load will be split across.

SETUP_BATCH_COUNT

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

DATA_BATCH_COUNT

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

FINALIZE_BATCH_COUNT

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

SETUP_BATCH_LOADED

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

DATA_BATCH_LOADED

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

FINALIZE_BATCH_LOADED

BIGINT

0

X

The number of batches that have been loaded for this load id. The same count will exists for all rows with a matching load id.

TABLE_COUNT

BIGINT

0

X

The number of tables associated with this load, which is generated by the system.

ROWS_LOADED

BIGINT

0

X

The number of rows associated with this load, which is generated by the system.

ROWS_COUNT

BIGINT

0

X

The number of rows associated with this load, which is generated by the system.

ERROR_FLAG

TINYINT

1

0

X

A flag that indicates that this request is in error during the last attempt to process.

ERROR_BATCH_ID

BIGINT

The batch ID that is in error when the error_flag is 1.

SQL_STATE

VARCHAR

10

For a request in error, this is the XOPEN or SQL 99 SQL State.

SQL_CODE

INTEGER

0

X

For a request in error, this is the error code from the database that is specific to the vendor.

SQL_MESSAGE

LONGVARCHAR

For a request in error, this is the error message that describes the error.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user or the system last updated this entry.

BATCH_BULK_LOAD_COUNT

BIGINT

0

X

The number of batches that were loaded with the bulk loader.

A.57. TRANSFORM_COLUMN

Defines the column mappings and optional data transformation for a data loader transformation.

Table 79. TRANSFORM_COLUMN

Name

Type

Size

Default

Keys

Not Null

Description

TRANSFORM_ID

VARCHAR

50

PK

X

Unique identifier of a specific transform.

INCLUDE_ON

CHAR

1

*

PK

X

Indicates whether this mapping is included during an insert (I), update (U), delete (D) operation at the target based on the dml type at the source. A value of * represents the fact that you want to map the column for all operations.

TARGET_COLUMN_NAME

VARCHAR

128

PK

X

Name of the target column.

SOURCE_COLUMN_NAME

VARCHAR

128

Name of the source column.

PK

TINYINT

1

0

Indicates whether this mapping defines a primary key to be used to identify the target row. At least one row must be defined as a pk for each transform_id.

TRANSFORM_TYPE

VARCHAR

50

copy

The name of a specific type of transform. Custom transformers can be configured as extension points.

TRANSFORM_EXPRESSION

LONGVARCHAR

An expression that is specific to the type of transform that is configured in transform_type. See the documentation for each transformer for more details.

TRANSFORM_ORDER

INTEGER

1

X

Specifies the order in which to apply transforms if more than one target operation occurs.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for transform_column

A.58. TRANSFORM_TABLE

Defines a data loader transformation which can be used to map arbitrary tables and columns to other tables and columns.

Table 80. TRANSFORM_TABLE

Name

Type

Size

Default

Keys

Not Null

Description

TRANSFORM_ID

VARCHAR

50

PK

X

Unique identifier of a specific transform.

SOURCE_NODE_GROUP_ID

VARCHAR

50

PK FK

X

The node group where data changes are captured.

TARGET_NODE_GROUP_ID

VARCHAR

50

PK FK

X

The node group where data changes will be sent.

TRANSFORM_POINT

VARCHAR

10

X

The point during the transport of captured data that a transform happens. Support values are EXTRACT or LOAD.

SOURCE_CATALOG_NAME

VARCHAR

255

Optional name for the catalog the configured table is in.

SOURCE_SCHEMA_NAME

VARCHAR

255

Optional name for the schema a configured table is in.

SOURCE_TABLE_NAME

VARCHAR

255

X

The name of the source table that will be transformed.

TARGET_CATALOG_NAME

VARCHAR

255

Optional name for the catalog a target table is in. Only use this if the target table is not in the default catalog.

TARGET_SCHEMA_NAME

VARCHAR

255

Optional name of the schema a target table is in. Only use this if the target table is not in the default schema.

TARGET_TABLE_NAME

VARCHAR

255

The name of the target table.

UPDATE_FIRST

TINYINT

1

0

If true, the target actions are attempted as updates first, regardless of whether the source operation was an insert or an update.

UPDATE_ACTION

VARCHAR

255

UPDATE_COL

X

An action to take upon update of a row. Possible values are: DEL_ROW, UPD_ROW, INS_ROW or NONE.

DELETE_ACTION

VARCHAR

10

X

An action to take upon delete of a row. Possible values are: DEL_ROW, UPD_ROW, or NONE.

TRANSFORM_ORDER

INTEGER

1

X

Specifies the order in which to apply transforms if more than one target operation occurs.

COLUMN_POLICY

VARCHAR

10

SPECIFIED

X

Specifies whether all columns need to be specified or whether they are implied. Possible values are SPECIFIED or IMPLIED.

CREATE_TIME

TIMESTAMP

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for transform_table

A.59. TRIGGER

Configures database triggers that capture changes in the database. Configuration of which triggers are generated for which tables is stored here. Triggers are created in a node’s database if the source_node_group_id of a router is mapped to a row in this table.

Table 81. TRIGGER

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_ID

VARCHAR

128

PK

X

Unique identifier for a trigger.

SOURCE_CATALOG_NAME

VARCHAR

255

Optional name for the catalog the configured table is in. If the name includes * then a wildcard match on the table name will be attempted. Wildcard names can include a list of names that are comma separated. The ! symbol may be used to indicate a NOT match condition.

SOURCE_SCHEMA_NAME

VARCHAR

255

Optional name for the schema a configured table is in. If the name includes * then a wildcard match on the table name will be attempted. Wildcard names can include a list of names that are comma separated. The ! symbol may be used to indicate a NOT match condition.

SOURCE_TABLE_NAME

VARCHAR

255

X

The name of the source table that will have a trigger installed to watch for data changes. If the name includes * then a wildcard match on the table name will be attempted. Wildcard names can include a list of names that are comma separated. The ! symbol may be used to indicate a NOT match condition.

CHANNEL_ID

VARCHAR

128

FK

X

The channel_id of the channel that data changes will flow through.

RELOAD_CHANNEL_ID

VARCHAR

128

reload

FK

X

The channel_id of the channel that will be used for reloads.

SYNC_ON_UPDATE

TINYINT

1

1

X

Whether or not to install an update trigger.

SYNC_ON_INSERT

TINYINT

1

1

X

Whether or not to install an insert trigger.

SYNC_ON_DELETE

TINYINT

1

1

X

Whether or not to install an delete trigger.

SYNC_ON_INCOMING_BATCH

TINYINT

1

0

X

Whether or not an incoming batch that loads data into this table should cause the triggers to capture data_events. Be careful turning this on, because an update loop is possible.

NAME_FOR_UPDATE_TRIGGER

VARCHAR

255

Override the default generated name for the update trigger.

NAME_FOR_INSERT_TRIGGER

VARCHAR

255

Override the default generated name for the insert trigger.

NAME_FOR_DELETE_TRIGGER

VARCHAR

255

Override the default generated name for the delete trigger.

SYNC_ON_UPDATE_CONDITION

LONGVARCHAR

Specify a condition for the update trigger firing using an expression specific to the database.

SYNC_ON_INSERT_CONDITION

LONGVARCHAR

Specify a condition for the insert trigger firing using an expression specific to the database.

SYNC_ON_DELETE_CONDITION

LONGVARCHAR

Specify a condition for the delete trigger firing using an expression specific to the database.

CUSTOM_BEFORE_UPDATE_TEXT

LONGVARCHAR

Specify update trigger text to execute before the SymmetricDS trigger text runs. If you need to modify data, use custom_on_update_text instead, so data is captured in order. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

CUSTOM_BEFORE_INSERT_TEXT

LONGVARCHAR

Specify insert trigger text to execute before the SymmetricDS trigger text runs. If you need to modify data, use custom_on_insert_text instead, so data is captured in order. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

CUSTOM_BEFORE_DELETE_TEXT

LONGVARCHAR

Specify delete trigger text to execute brefore the SymmetricDS trigger text runs. If you need to modify data, use custom_on_delete_text instead, so data is captured in order. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

CUSTOM_ON_UPDATE_TEXT

LONGVARCHAR

Specify update trigger text to execute after the SymmetricDS trigger text runs. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

CUSTOM_ON_INSERT_TEXT

LONGVARCHAR

Specify insert trigger text to execute after the SymmetricDS trigger text runs. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

CUSTOM_ON_DELETE_TEXT

LONGVARCHAR

Specify delete trigger text to execute after the SymmetricDS trigger text runs. This field is not applicable for H2, HSQLDB 1.x or Apachy Derby.

EXTERNAL_SELECT

LONGVARCHAR

Specify a SQL select statement that returns a single result. It will be used in the generated database trigger to populate the EXTERNAL_DATA field on the data table.

TX_ID_EXPRESSION

LONGVARCHAR

Override the default expression for the transaction identifier that groups the data changes that were committed together.

CHANNEL_EXPRESSION

LONGVARCHAR

An expression that will be used to capture the channel id in the trigger. This expression will only be used if the channel_id is set to 'dynamic.'

EXCLUDED_COLUMN_NAMES

LONGVARCHAR

Specify a comma-delimited list of columns that should not be synchronized from this table. Note that if a primary key is found in this list, it will be ignored.

INCLUDED_COLUMN_NAMES

LONGVARCHAR

Specify a comma-delimited list of columns that should only be synchronized from this table. Note that if a primary key is found in this list, it will be ignored.

SYNC_KEY_NAMES

LONGVARCHAR

Specify a comma-delimited list of columns that should be used as the key for synchronization operations. By default, if not specified, then the primary key of the table will be used.

USE_STREAM_LOBS

TINYINT

1

0

X

Specifies whether to capture lob data as the trigger is firing or to stream lob columns from the source tables using callbacks during extraction. A value of 1 indicates to stream from the source via callback; a value of 0, lob data is captured by the trigger.

USE_CAPTURE_LOBS

TINYINT

1

0

X

Provides a hint as to whether this trigger will capture big lobs data. If set to 1 every effort will be made during data capture in trigger and during data selection for initial load to use lob facilities to extract and store data in the database. On Oracle, this may need to be set to 1 to get around 4k concatenation errors during data capture and during initial load.

USE_CAPTURE_OLD_DATA

TINYINT

1

1

X

Set this to 1 to capture old data. Old data is used for conflict resolution and it also is used to calculate which columns have changed. The software will only update changed columns when old data is captured.

USE_HANDLE_KEY_UPDATES

TINYINT

1

1

X

Deprecated. Allows handling of primary key updates (SQLServer dialect only)

STREAM_ROW

TINYINT

1

0

X

Captures only the primary key when the trigger fires and creates a reload event to pull the full row during extraction.

TIME_BASED_COLUMN_NAME

VARCHAR

255

Only used for time-based capture nodes. This will be the column to use as part of the time-based capture mining.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for trigger

A.60. TRIGGER_HIST

A history of a table’s definition and the trigger used to capture data from the table. When a database trigger captures a data change, it references a trigger_hist entry so it is possible to know which columns the data represents. trigger_hist entries are made during the sync trigger process, which runs at each startup, each night in the syncTriggersJob, or any time the syncTriggers() JMX method is manually invoked. A new entry is made when a table definition or a trigger definition is changed, which causes a database trigger to be created or rebuilt.

Table 82. TRIGGER_HIST

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_HIST_ID

INTEGER

PK

X

Unique identifier for a trigger_hist entry

TRIGGER_ID

VARCHAR

128

X

Unique identifier for a trigger

SOURCE_TABLE_NAME

VARCHAR

255

X

The name of the source table that will have a trigger installed to watch for data changes.

SOURCE_CATALOG_NAME

VARCHAR

255

The catalog name where the source table resides.

SOURCE_SCHEMA_NAME

VARCHAR

255

The schema name where the source table resides.

NAME_FOR_UPDATE_TRIGGER

VARCHAR

255

The name used when the insert trigger was created.

NAME_FOR_INSERT_TRIGGER

VARCHAR

255

The name used when the update trigger was created.

NAME_FOR_DELETE_TRIGGER

VARCHAR

255

The name used when the delete trigger was created.

TABLE_HASH

BIGINT

0

X

A hash of the table definition, used to detect changes in the definition.

TRIGGER_ROW_HASH

BIGINT

0

X

A hash of the trigger definition. If changes are detected to the values that affect a trigger definition, then the trigger will be regenerated.

TRIGGER_TEMPLATE_HASH

BIGINT

0

X

A hash of the trigger text. If changes are detected to the values that affect a trigger text then the trigger will be regenerated.

COLUMN_NAMES

LONGVARCHAR

X

The column names defined on the table. The column names are stored in comma-separated values (CSV) format.

PK_COLUMN_NAMES

LONGVARCHAR

X

The primary key column names defined on the table. The column names are stored in comma-separated values (CSV) format.

IS_MISSING_PK

TINYINT

0

X

Indicates whether the source table is missing a primary key.

LAST_TRIGGER_BUILD_REASON

CHAR

1

X

The following reasons for a change are possible: New trigger that has not been created before (N); Schema changes in the table were detected (S); Configuration changes in Trigger (C); Trigger was missing (T), Trigger template changed (E), Forced rebuild (F).

ERROR_MESSAGE

LONGVARCHAR

Record any errors or warnings that occurred when attempting to build the trigger.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

INACTIVE_TIME

TIMESTAMP

The date and time when a trigger was inactivated.

A.61. TRIGGER_ROUTER

Map a trigger to a router.

Table 83. TRIGGER_ROUTER

Name

Type

Size

Default

Keys

Not Null

Description

TRIGGER_ID

VARCHAR

128

PK FK

X

The id of a trigger.

ROUTER_ID

VARCHAR

50

PK FK

X

The id of a router.

ENABLED

TINYINT

1

1

X

Indicates whether this trigger router is enabled or not.

INITIAL_LOAD_ORDER

INTEGER

1

X

Order sequence of this table when an initial load is sent to a node. If this value is the same for multiple tables, then SymmetricDS will attempt to order the tables according to FK constraints. If this value is set to a negative number, then the table will be excluded from an initial load.

INITIAL_LOAD_SELECT

LONGVARCHAR

Optional expression that can be used to pare down the data selected from a table during the initial load process.

INITIAL_LOAD_DELETE_STMT

LONGVARCHAR

The expression that is used to delete data when an initial load occurs. If this field is empty, no delete will occur before the initial load. If this field is not empty, the text will be used as a sql statement and executed for the initial load delete.

PING_BACK_ENABLED

TINYINT

1

0

X

When enabled, the node will route data that originated from a node back to that node. This attribute is only effective if sync_on_incoming_batch is set to 1.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

DESCRIPTION

LONGVARCHAR

Optional notes and comments for trigger_router

DATA_REFRESH_TYPE

VARCHAR

50

If this table should be refreshed when the refresh job runs. Usually for tables where typical trigger, log, or time-based capture is not available. Possible values are: 'auto' (attempts difference only refresh unless there are more than the threshold and will switch automatically to a full), 'differences' (Uses the compare feature to send only differences), and 'full' (remove all data at the target and fully load it again using a reload)

A.62. TRIGGER_ROUTER_GROUPLET

Deprecated in 3.9. This tables defines what grouplets are associated with what trigger routers. The existence of the grouplet for a trigger_router enables nodes associated with the grouplet and at the same time it disables the trigger router for all other nodes.

Table 84. TRIGGER_ROUTER_GROUPLET

Name

Type

Size

Default

Keys

Not Null

Description

GROUPLET_ID

VARCHAR

50

PK FK

X

Unique identifier for the grouplet.

TRIGGER_ID

VARCHAR

128

PK FK

X

The id of a trigger.

ROUTER_ID

VARCHAR

50

PK FK

X

The id of a router.

APPLIES_WHEN

CHAR

1

PK

X

Indicates the side that a grouplet should be applied to. Use 'T' for target and 'S' for source and 'B' for both source and target.

CREATE_TIME

TIMESTAMP

X

Timestamp when this entry was created.

LAST_UPDATE_BY

VARCHAR

50

The user who last updated this entry.

LAST_UPDATE_TIME

TIMESTAMP

X

Timestamp when a user last updated this entry.

Appendix B: Parameter List

There are two kinds of parameters that can be used to configure the behavior of SymmetricDS: Startup Parameters and Runtime Parameters . Startup Parameters are required to be in a system property or a property file, while Runtime Parameters can also be found in the Parameter table from the database. Parameters are re-queried from their source at a configured interval and can also be refreshed on demand by using the JMX API.

The following table shows the source of parameters and the hierarchy of precedence.

Table 85. Parameter Discovery Precedence
Location Required Description

symmetric-default.properties

Y

Packaged inside symmetric-core jar file. This file has all the default settings along with descriptions.

conf/symmetric.properties

N

Changes to this file in the conf directory of a standalone install apply to all engines in the JVM.

symmetric-override.properties

N

Changes to this file, provided by the end user in the JVM’s classpath, apply to all engines in the JVM.

engines/*.properties

N

Properties for a specific engine or node that is hosted in a standalone install.

Java System Properties

N

Any SymmetricDS property can be passed in as a -D property to the runtime. It will take precedence over any properties file property.

Parameter table

N

A table which contains SymmetricDS parameters. Parameters can be targeted at a specific node group and even at a specific external id. These settings will take precedence over all of the above.

IParameterFilter

N

An extension point which allows parameters to be sourced from another location or customized. These settings will take precedence over all of the above.

B.1. Startup Parameters

Startup parameters are read once from properties files and apply only during start up. The following properties are used:

allow.updates.with.results

When executing DML statements during data load, this controls whether executeUpdate or execute is used on the PreparedStatement. executeUpdate is used by default. execute() allows for unusual situations like when an application trigger generates a result set during an update statement.

Default: false

auto.config.database

If this is true, when symmetric starts up it will try to create the necessary tables.

Default: true

auto.config.database.fast

If both auto.config.database and this parameter are true when symmetric starts up, it will only try to create the necessary tables if the current software version is different from the version used during the last start up (stored on sym_node.symmetric_version).

Default: true

auto.config.registration.svr.sql.script

Provide the path to a SQL script that can be run to do initial setup of a registration server. This script will only be run on a registration server if the node_identity cannot be found.

Default:

auto.insert.registration.svr.if.not.found

If this is true, then node, group, security and identity rows will be inserted if the registration.url is blank and there is no configured node identity.

Default: true

auto.resolve.unique.index.ignore.null.values

If this is true, when a batch receives a unique index violation, the blocking rows for each unique index will only be deleted if the unique index has a value that is not null.

Default: true

auto.sync.config.after.upgrade

If this is true, then check if configuration should be pulled from registration server after upgrading. If the config version in the database does not match the software version, it will pull config.

Default: true

auto.update.node.values.from.properties

Update the node row in the database from the local properties during a heartbeat operation.

Default: true

cache.table.time.ms

This is the amount of time table meta data will be cached before re-reading it from the database

Default: 3600000

cluster.server.id

Set this if you want to give your server a unique name to be used to identify which server did what action. Typically useful when running in a clustered environment. This is currently used by the ClusterService when locking for a node.

Default:

console.explore.show

Display the explore tab.

Default: true

console.log.slow.sql.threshold.millis

For queries related to the web console, defines the number of milliseconds before logging that a query is slow.

Default: 4000

db.connection.properties

These are settings that will be passed to the JDBC driver as connection properties. Suggested settings by database are as follows: Oracle db.connection.properties=oracle.net.CONNECT_TIMEOUT=300000;oracle.jdbc.ReadTimeout=300000;SetBigStringTryClob=true

Default:

db.delimited.identifier.mode

Determines whether delimited identifiers are used or normal SQL92 identifiers (which may only contain alphanumerical characters and the underscore, must start with a letter and cannot be a reserved keyword).

Default: true

db.driver

Specify your database driver

Default: org.h2.Driver

db.init.sql

Specify one or more SQL statements that will be run when a database connection is created. Multiple SQL statements must be separated by semicolons. If a statement has a semicolon in the value, then put two semicolons in its place to allow the specification of keeping the semicolon in the value. For example: db.init.sql=SET OPTION 'host=localhost;;user=user1;;password=pwd1';SET OPTION TWO This will be split into two statements that are sent to the database server: SET OPTION 'host=localhost;user=user1;password=pwd1' SET OPTION TWO

Default:

db.jdbc.bulk.execute.batch.override

This override any database specific bulk loader if the channel data loader algorithm is set to bulk and a specific bulk loader is available. For example is a PostgreSQL database is setup for bulk loading and this parameter is true the JdbcBatchBulkDatabaseWriter will be used instead of the PostgreSQL bulk loader. If a dialect does not have a designated bulk loader and the channel is set for bulk loading this will be used automatically even if the the parameter is false. Essentially the parameter only controls "overriding" a dialect specific bulk loader.

Default: false

db.jdbc.bulk.execute.batch.size

This is the default number of rows that will be sent to the database as a batch when SymmetricDS uses the JDBC batch API for bulk data load types. Designed for loads that are setup with a channel data loader algorithm of batch.

Default: 25

db.jdbc.execute.batch.size

This is the default number of rows that will be sent to the database as a batch when SymmetricDS uses the JDBC batch API. Currently, only routing uses JDBC batch. The data loader does not.

Default: 100

db.jdbc.isolation.level

Override the JDBC isolation level. The isolation level is detected by platform and automatically set, but it can be overridden here. Most platforms need at least read committed level to prevent phantom reads. (0=none, 1=read uncommitted, 2=read committed, 4=repeatable read, 8=serializable)

Default:

db.jdbc.streaming.results.fetch.size

This is the default fetch size for streaming result sets.

Default: 100

db.jndi.name

Name of a JNDI data source to use instead of using SymmetricDS’s connection pool. When this is set the db.url is ignored. Using a JNDI data source is relevant when deploying to an application server.

Default:

db.metadata.ignore.case

Indicates that case should be ignored when looking up references to tables using the database’s metadata api.

Default: true

db.native.extractor

Name of class that can extract native JDBC objects and interact directly with the driver. Spring uses this to perform operations specific to database, like handling LOBs on Oracle.

Default: org.springframework.jdbc.support.nativejdbc.CommonsDbcpNativeJdbcExtractor

db.password

Specify your database password

Default:

db.pool.initial.size

The initial size of the connection pool

Default: 5

db.pool.max.active

The maximum number of connections that will be allocated in the pool The http.concurrent.workers.max value should be less than half of this value.

Default: 50

db.pool.max.idle

The maximum number of connections that can remain idle in the pool, without extra ones being released

Default: 20

db.pool.max.wait.millis

This is how long a request for a connection from the datasource will wait before giving up.

Default: 30000

db.pool.min.evictable.idle.millis

This is how long a connection can be idle before it will be evicted.

Default: 120000

db.pool.min.idle

The minimum number of connections that can remain idle in the pool, without extra ones being created

Default: 5

db.read.strings.as.bytes

If set to true forces database columns that contain character data to be read as bytes (bypassing JDBC driver character encoding) so the raw values be encoded using the system default character set (usually UTF8). This property was added to bypass MySQL character encoding so the raw data can be converted to utf8 directly.

Default: false

db.sql.query.timeout.seconds

Number of seconds a query can execute before being interrupted using the statement query timeout. The timeout can avoid hanging if a query is slow to return because of an un-optimized plan or if the system is overloaded. Set to zero for no limit. (Make sure socket/read timeout in db.url is set higher than query timeout.)

Default: 3600

db.url

Specify your database URL

Default: jdbc:h2:mem:setme

db.user

Specify your database user

Default: please set me

db.validation.query

This is the query to validate the database connection in Connection Pool. It is database specific. The following are example statements for different databases. MySQL db.validation.query=select 1 Oracle db.validation.query=select 1 from dual DB2 db.validation.query=select max(1) from syscat.datatypes

Default:

db2.capture.transaction.id

Turn on the capture of transaction id for DB2 systems that support it.

Default: false

db2.zseries.version

Use to map the version string a zseries jdbc driver returns to the 'zseries' dialect

Default: DSN08015

engine.name

This is the engine name. This should be set if you have more than one engine running in the same JVM. It is used to name the JMX management bean. Please do not use underscores in this name.

Default: SymmetricDS

external.id

The external id for this SymmetricDS node. The external id is usually used as all or part of the node id.

Default: please set me

group.id

The node group id that this node belongs to

Default: please set me

hsqldb.initialize.db

If using the HsqlDbDialect, this property indicates whether Symmetric should setup the embedded database properties or if an external application will be doing so.

Default: true

http.concurrent.reservation.timeout.ms

This is the amount of time the host will keep a concurrent connection reservation after it has been attained by a client node while waiting for the subsequent reconnect to push.

Default: 20000

http.session.expire.seconds

How long in seconds an authenticated node can keep a session before it expires and the node must be authenticated again. A value of zero or less means never expire.

Default: 14400

http.session.max.count

Maximum number of authenticated sessions to keep in memory before removing the oldest. Normally, this won’t be reached unless something is mis-configured, like a cluster that is not using a sticky session load balancer.

Default: 15000

http.use.header.security.token

When authenticating to a server node, send the security token in the request header instead of using a URL parameter. Using the request header avoids accidentally logging the security token. The transport uses the remote node’s version to determine if it should use header or parameter. Set this to false to force usage of the older parameter style of authentication.

Default: true

http.use.session.auth

When authenticating a client node, create a session on the server and give a session ID to the client that can be sent with subsequent requests. This avoids authenticating every request and limits how often the security token is sent.

Default: true

jmx.line.feed

Specify the type of line feed to use in JMX console methods. Possible values are: text or html.

Default: text

job.log.miner.period.time.ms

How often to run Log Miner in milliseconds

Default: 10000

job.random.max.start.time.ms

When starting jobs, symmetric attempts to randomize the start time to spread out load. This is the maximum wait period before starting a job.

Default: 10000

log.miner.use.incoming.transactions

Use incoming batches to collect transaction IDs and their source node IDs. When mining changes, it will ignore these transactions to prevent update loops. If trigger.sync_on_incoming is enabled, it will mine the change but include the source node ID so routing won’t send a batch back to that node. Some miners have their own built-in loop prevention and don’t need this. Supported on the following platforms: PostgreSQL, Oracle, MySql, MariaDB.

Default: true

log.slow.sql.threshold.millis

Defines the number of milliseconds before logging that a query is slow.

Default: 20000

log.sql.parameters.inline

Defines whether the logging of SQL statements include the values inline or not.

Default: true

mysql.bulk.load.local

Whether or not files are local to client only, so we must send the file to MySQL to load. If client is running on same server as MySQL, then this can be set to false to have MySQL read file directly.

Default: true

mysql.bulk.load.max.bytes.before.flush

Maximum number of bytes to write to file before running with 'LOAD DATA INFILE' to MySQL

Default: 1000000000

mysql.bulk.load.max.rows.before.flush

Maximum number of rows to write to file before running with 'LOAD DATA INFILE' to MySQL

Default: 100000

node.password.encrypt.enable

Encrypt the node_password in the sym_node_security table when saving it. Decryption is always done on reading if the node_password starts with "enc:" prefix.

Default: true

oracle.bulk.load.ezconnect

For bulk loading with SQL*Loader, specify how to connect to the database with an ezconnect name. If blank, the connection is determined using the db.url parameter.

Default:

oracle.bulk.load.sqlldr.cmd

Path to the sqlldr executable for running Oracle SQL*Loader. If blank, it will check for ORACLE_HOME environment variable and find sqlldr there. Otherwise, it will run "sqlldr" and expect the operating system to find it.

Default:

oracle.bulk.load.sqlldr.infile.charset

For bulk loading with SQL*Loader, specify input file character set when needed to support characters other than US-ASCII characters. For example: oracle.bulk.load.sqlldr.infile.charset=UTF8 LENGTH SEMANTICS CHAR This will generate the following string in the control file after the LOAD DATA line: CHARACTERSET UTF8 LENGTH SEMANTICS CHAR If blank, this line will not show up.

Default:

oracle.bulk.load.sqlldr.options

Options passed to Oracle SQL*Loader.

Default: silent=(header,discards) direct=false readsize=4096000 bindsize=4096000 rows=2000 errors=0

oracle.jdbc.lob.handling

This indicates how to process LOB values. "plain" is the default, which uses the JDBC interface directly. "createtemporarylob" uses the creation of CLOB/BLOB temporary objects to process the data. "streamlob" uses the stream interface of the JDBC interface (part of JDBC 4.0).

Default: plain

postgres.convert.infinity.date.to.null

Enable this parameter to convert infinite (positive or negative) values of date and timestamp (with/without time zones) types to null. DatabaseOverrideable: false

Default: true

registration.max.time.between.retries

The maximum time (in seconds) between node registration retries.

Default: 30

registration.url

This is the URL this node will use to register and pull it’s configuration. If this is the root server, then this may remain blank and the configuration should be inserted directly into the database

Default: please set me

routing.gaps.use.transaction.view

Find the earliest starting time of any transaction in the database, and expire any data gaps that were created before that time. When enabled, it can expire gaps sooner in some cases (before the routing.stale.dataid.gap.time.ms timeout), and in other cases, it can prevent expiring gaps that reach the timeout when there are still open transactions.

Default: false

s3.load.format

The file format in which S3 output will be written

Default: csv

security.service.class.name

The class name for the Security Service to use for encrypting and decrypting database passwords. Leave blank for default service.

Default:

staging.dir

This is the location the staging directory will be put. If it isn’t set the staging directory will be located according to java.io.tmpdir.

Default:

staging.low.space.threshold.megabytes

Staging directory low disk space threshold in megabytes. When free space goes below the threshold, requests to create staging files will receive an exception. Data loading and extracting will stop and log an error until disk space becomes available. To disable free space checks, set to zero or negative.

Default: 100

staging.manager.class

Staging Manager

Default: com.jumpmind.symmetric.stage.EnhancedStagingManager

start.compare.job

Whether the compare job is started.

Default: true

start.data.refresh.job

Whether the data refresh job is started.

Default: false

start.heartbeat.job

Whether the heartbeat job is enabled for this node. The heartbeat job simply inserts an event to update the heartbeat_time column on the node_host table for the current node.

Default: true

start.initial.load.extract.job

Whether the background initial load extractor job is started.

Default: true

start.log.miner.job

Log Miner job to find changes from a database archive log

Default: false

start.monitor.job

Whether the monitor job is started.

Default: true

start.pull.job

Whether the pull job is enabled for this node.

Default: true

start.purge.incoming.job

Whether the incoming purge job is enabled for this node.

Default: true

start.purge.outgoing.job

Whether the outgoing purge job is enabled for this node.

Default: true

start.push.job

Whether the push job is enabled for this node.

Default: true

start.refresh.cache.job

Whether the refresh cache job is enabled for this node.

Default: false

start.routing.job

Whether the routing job is enabled for this node.

Default: true

start.stage.management.job

Whether the stage management job is enabled for this node.

Default: true

start.stat.flush.job

Whether the statistic flush job is enabled for this node.

Default: true

start.synctriggers.job

Whether the sync triggers job is enabled for this node.

Default: true

start.watchdog.job

Whether the watchdog job is enabled for this node.

Default: true

statistic.manager.class

Statistics Manager

Default: com.jumpmind.symmetric.statistic.ThroughputStatisticManager

statistic.record.count.threshold

Controls whether statistics are recorded to the sym_node_host_job_stats table. Statistics will only be recorded if the number of items that were processed exceeds this threshold. A value of -1 (or any negative value) disables these statistics.

Default: 0

statistic.record.enable

Control whether statistics are recorded.

Default: true

stream.to.file.enabled

Save data to the file system before transporting it to the client or loading it to the database if the number of bytes is past a certain threshold. This allows for better compression and better use of database and network resources. Statistics in the batch tables will be more accurate if this is set to true because each timed operation is independent of the others.

Default: true

sync.table.prefix

When symmetric tables are created and accessed, this is the prefix to use for the tables.

Default: sym

sync.trigger.prefix

When symmetric triggers are created and accessed, this is the prefix to use for the tables. If this is not set, it will use the sync.table.prefix as the trigger prefix

Default:

sync.triggers.expand.table.clone

When sync triggers expands a table name to include node external ID, use the first table found and clone it for all other nodes to improve performance.

Default: true

sync.triggers.reg.svr.install.without.config

Whether or not sync triggers job will install triggers on a registration server before configuration is created. When true, triggers are installed as soon as possible, right after configuration tables are created. When false, triggers are installed when the configuration includes one or more group links where this node is the source. A default of true avoids race conditions associated with registering nodes and performing operations before configuration is complete and triggers are installed.

Default: true

sync.url

The url that can be used to access this SymmetricDS node. The default setting of http://$(hostName):31415/sync should be valid of the standalone launcher is used with the default settings The tokens of $(hostName) and $(ipAddress) are supported for this property.

target.db.connection.properties

These are settings that will be passed to the target JDBC driver as connection properties. Suggested settings by database are as follows: Oracle db.connection.properties=oracle.net.CONNECT_TIMEOUT=300000;oracle.jdbc.ReadTimeout=300000;SetBigStringTryClob=true

Default:

target.db.driver

Specify your target database driver. This applies for load only, extract only, and log based data capture.

Default:

target.db.init.sql

Specify one or more SQL statements that will be run when a database connection is created. Multiple SQL statements must be separated by semicolons. If a statement has a semicolon in the value, then put two semicolons in its place to allow the specification of keeping the semicolon in the value. For example: db.init.sql=SET OPTION 'host=localhost;;user=user1;;password=pwd1';SET OPTION TWO This will be split into two statements that are sent to the database server: SET OPTION 'host=localhost;user=user1;password=pwd1' SET OPTION TWO

Default:

target.db.jdbc.execute.batch.size

This is the default number of rows that will be sent to the target database as a batch when SymmetricDS uses the JDBC batch API. Currently, only routing uses JDBC batch. The data loader does not.

Default:

target.db.jdbc.isolation.level

Override the target JDBC isolation level. The isolation level is detected by platform and automatically set, but it can be overridden here. Most platforms need at least read committed level to prevent phantom reads. (0=none, 1=read uncommitted, 2=read committed, 4=repeatable read, 8=serializable)

Default:

target.db.jdbc.streaming.results.fetch.size

This is the default fetch size for target streaming result sets.

Default:

target.db.password

Specify your target database password. This applies for load only, extract only, and log based data capture.

Default:

target.db.pool.initial.size

The initial size of the target connection pool

Default:

target.db.pool.max.active

The maximum number of target connections that will be allocated in the pool The http.concurrent.workers.max value should be less than half of this value.

Default:

target.db.pool.max.idle

The maximum number of target connections that can remain idle in the pool, without extra ones being released

Default:

target.db.pool.max.wait.millis

This is how long a request for a target connection from the datasource will wait before giving up.

Default:

target.db.pool.min.evictable.idle.millis

This is how long a target connection can be idle before it will be evicted.

Default:

target.db.pool.min.idle

The minimum number of target connections that can remain idle in the pool, without extra ones being created

Default:

target.db.read.strings.as.bytes

If set to true forces target database columns that contain character data to be read as bytes (bypassing JDBC driver character encoding) so the raw values be encoded using the system default character set (usually UTF8). This property was added to bypass MySQL character encoding so the raw data can be converted to utf8 directly.

Default:

target.db.sql.query.timeout.seconds

Number of seconds a query can execute before being interrupted using the statement query timeout. The timeout can avoid hanging if a query is slow to return because of an un-optimized plan or if the system is overloaded. Set to zero for no limit. (Make sure socket/read timeout in db.url is set higher than query timeout.)

Default: 3600

target.db.test.on.borrow

This property determines whether or not the target pool will validate objects before they are borrowed from the pool.

Default:

target.db.test.on.return

This property determines whether or not the target pool will validate objects before they are returned to the pool.

Default:

target.db.test.while.idle

This property determines whether or not the target idle object evictor will validate connections.

Default:

target.db.url

Specify your target database URL. This applies for load only, extract only, and log based data capture.

Default:

target.db.user

Specify your target database user. This applies for load only, extract only, and log based data capture.

Default:

target.db.validation.query

This is the query to validate the target database connection in Connection Pool. It is database specific. The following are example statements for different databases. MySQL db.validation.query=select 1 Oracle db.validation.query=select 1 from dual DB2 db.validation.query=select max(1) from syscat.datatypes

Default:

target.log.slow.sql.threshold.millis

Defines the number of milliseconds before logging that a target query is slow.

Default:

target.log.sql.parameters.inline

Defines whether the logging of target SQL statements include the values inline or not.

Default:

tibero.bulk.load.dname

For bulk loading with tbLoader, specify the database name. If blank, the database name is determined using the db.url parameter.

Default:

tibero.bulk.load.tbloader.cmd

Path to the tbloader executable for running Tibero tbLoader. If blank, it will check for TB_HOME environment variable and find tbLoader there. Otherwise, it will run "tbloader" and expect the operating system to find it.

Default:

tibero.bulk.load.tbloader.options

Options passed to Tibero tbLoader

Default: direct=N dpl_parallel=1 disable_idx=N readsize=2097152 bindsize=2097152 rows=2000 errors=0

transport.type

Specify the transport type. Supported values currently include: http, file, internal.

Default: http

treat.binary.as.lob.enabled

Whether binary fields should be treated as lobs

Default: false

update.service.class

Update service

Default: com.jumpmind.symmetric.console.service.impl.ConsoleUpdateService

B.2. Runtime Parameters

Runtime parameters are read periodically from properties files or the database. The following properties are used:

as400.cast.clob.to

Specify the database type to cast clob values to

Default: DBCLOB

as400.journal.library

The library that contains the journal for log based capture on the AS400

Default: SYM

as400.journal.name

The journal that will be used for log based capture on the AS400

Default: QSQJRN

auto.create.channels.cleanup

When auto-creating channels, this option enables removal of reload channels that weren’t used when rebalancing tables across reload channels

Default: true

auto.registration

If this is true, registration is opened automatically for nodes requesting it.

Default: false

auto.reload

If this is true, a reload is automatically sent to nodes when they register and it’s either their first time or they haven’t received an initial load before. (A first-time registration occurs when a node contacts the registration URL because it is missing its node identity. The status of the initial load for a node is recorded on sym_node_security as the initial_load_time.) This parameter needs set at the registration node.

Default: false

auto.reload.reverse

If this is true, a reload is automatically sent from a source node to all target nodes after the source node has registered for the first time. This parameter needs set at the registration node.

Default: false

auto.reload.use.config

If this is true, a reload will use configuration to determine which nodes to send a load to. Default will send to the load to the node it regiestered

Default: false

auto.resolve.capture.delete.missing.rows

If this is true, when a delete affects zero rows, if the table is configured for sync on incoming, then the auto resolver will write the deletes into the capture log.

Default: false

auto.resolve.foreign.key.violation

If this is true, when a batch receives a foreign key violation, the missing data will be automatically sent to resolve it. The resolution is done at the source node by sending reload batches when it receives the acknowledgement of the batch error.

Default: true

auto.resolve.foreign.key.violation.delete

If this is true, when a batch receives a foreign key violation due to the parent row being updated or deleted, all existing child rows will be deleted.

Default: true

auto.resolve.foreign.key.violation.reverse

If this is true, when a batch receives a foreign key violation, the missing data will be automatically sent to resolve it. The resolution is done at the target node by sending a script that requests reload batches.

Default: false

auto.resolve.foreign.key.violation.reverse.peers

When resolving a foreign key violation for a batch in error at a node, also send the missing data to peers in a multi-master configuration who may also be missing the same data.

Default: false

auto.resolve.foreign.key.violation.reverse.reload

If this is true, when a reload batch receives a foreign key violation, the missing data will be automatically sent to resolve it. The resolution is done at the target node by sending a script that requests reload batches.

Default: true

auto.resolve.primary.key.violation

If this is true, when a batch receives a primary key violation during an update, the blocking row will be replaced with the updated row.

Default: true

auto.resolve.unique.index.violation

If this is true, when a batch receives a unique index violation, the blocking rows for each unique index will be deleted.

Default: true

auto.start.engine

This indicates whether this node engine should be started when the instance is restarted

Default: true

auto.sync.config.at.startup

If this is true, then check if configuration should be pulled from registration server at startup. If the config version in the database does not match the software version, it will pull config.

Default: true

auto.sync.configuration

Capture and send SymmetricDS configuration changes to client nodes.

Default: true

auto.sync.configuration.on.incoming

Whether triggers should fire when changes sync into the node that this property is configured for.

Default: true

auto.sync.triggers

If this is true, triggers will be created or dropped to match configuration during the sync triggers process.

Default: true

auto.sync.triggers.after.config.change

If this is true, when a configuration change is detected during routing, symmetric will make sure all triggers in the database are up to date.

Default: true

auto.sync.triggers.after.config.loaded

If this is true, when a configuration change is detected while being loaded onto a target node, symmetric will make sure all triggers in the database are up to date.

Default: true

auto.sync.triggers.at.startup

If this is true, then run the sync triggers process at startup

Default: false

auto.sync.triggers.at.startup.force

If this is true, then force rebuild of all triggers at startup

Default: false

batch.screen.max.to.select

The maximum number of batches to display on the batch screens.

Default: 0

batch.screen.use.batchps

Enable use of batchps command over remote status for outgoing batches screen, which uses current list of batches being processed. When disabled, it will use batchstatus command with a list of batches to track over remote status.

Default: true

batch.window.max.data.length

The maximum number of characters to display in the data grid in the batch window.

Default: 100

bsh.extension.global.script

BeanShell script to include at the beginning of all scripts used in extensions

Default:

bsh.load.filter.handles.missing.tables

This parameter can be used to indicate that bean shell load filters will handle missing tables. Useful for the case where you want to make, for example, global catalog or schema changes at the destination in the case where the catalog, schema, or table doesn’t exist but the BSH will handle it.

Default: false

bsh.transform.global.script

BeanShell script to include at the beginning of all scripts used in transforms

Default:

cache.channel.common.batches.time.ms

This is the amount of time the routing service will cache the common batch status of channels.

Default: 600000

cache.channel.default.router.time.ms

This is the amount of time the routing service will cache the default router status of channels.

Default: 600000

cache.channel.time.ms

This is the amount of time channel entries will be cached before re-reading them from the database.

Default: 600000

cache.conflict.time.ms

This is the amount of time conflict setting entries will be cached before re-reading them from the database.

Default: 600000

cache.grouplets.time.ms

This is the amount of time grouplet entries will be cached before re-reading them from the database.

Default: 600000

cache.load.filter.time.ms

This is the amount of time load filter entries will be cached before re-reading them from the database.

Default: 600000

cache.monitor.time.ms

This is the amount of time monitor entries will be cached before re-reading them from the database.

Default: 60000

cache.node.group.link.time.ms

This is the amount of time node group links entries will be cached before re-reading them from the database.

Default: 600000

cache.node.security.time.ms

This is the amount of time node security entries will be cached before re-reading them from the database.

Default: 600000

cache.node.time.ms

This is the amount of time node entries will be cached before re-reading them from the database.

Default: 600000

cache.notification.time.ms

This is the amount of time notification entries will be cached before re-reading them from the database.

Default: 60000

cache.transform.time.ms

This is the amount of time transform entries will be cached before re-reading them from the database.

Default: 600000

cache.trigger.router.time.ms

This is the amount of time trigger entries will be cached before re-reading them from the database.

Default: 600000

check.software.updates

Check for a newer version of SymmetricDS and let the user know when one is available.

Default: true

cloud.bulk.codepage

The cloud based codepage to use for bulk insert if supported by vendor and version

Default:

cloud.bulk.field.quote

The cloud based character used in bulk loading to be used when quoting fields

Default:

cloud.bulk.field.terminator

The cloud based terminator used in bulk loading to separate each field

Default: |

cloud.bulk.load.azure.account.key

The Azure storage account key

Default:

cloud.bulk.load.azure.account.name

The Azure storage account name

Default:

cloud.bulk.load.azure.blob.container

The Azure storage account blob container name

Default: symmetricds

cloud.bulk.load.azure.sas.token

The Azure storage account SAS Token (Shared Access Signature)

Default:

cloud.bulk.load.max.bytes.before.flush

Maximum number of bytes to write to file before copying to cloud storage

Default: -1

cloud.bulk.load.max.rows.before.flush

Maximum number of rows to write to file before copying to cloud storage

Default: -1

cloud.bulk.load.s3.access.key

The AWS access key ID (aws_access_key_id) to use as credentials for uploading to S3

Default:

cloud.bulk.load.s3.bucket

The S3 bucket where bulk load files will be uploaded

Default:

cloud.bulk.load.s3.endpoint

The endpoint for the s3 bucket. If not set it will use the default endpoint.

Default:

cloud.bulk.load.s3.region

The AWS S3 region

Default: us-east-1

cloud.bulk.load.s3.secret.key

The AWS secret access key (aws_secret_access_key) to use as credentials for uploading to S3

Default:

cloud.bulk.row.terminator

The cloud based terminator used in bulk loading to be used for a new row

Default:

cluster.lock.enabled

Enables clustering of jobs across multiple nodes. This feature is available in SymmetricDS Pro.

Default: false

cluster.lock.refresh.ms

Period of time that certain locks will get refreshed while long processing happens. This value should be a small fraction of cluster.lock.timeout.ms

Default: 1200000

cluster.lock.timeout.ms

Indicate that this node is being run on a farm or cluster of servers and it needs to use the database to 'lock' out other activity when actions are taken.

Default: 7200000

cluster.staging.enabled

Set this to true if you are sharing your staging directory with multiple instances of SymmetricDS

Default: false

compare.immediate.if.request.found

If true, the job will continue running as long as uncompleted requests exist. See compare.immediate.time.ms for the delay between checking and executing requests.

Default: true

compare.immediate.time.ms

The delay between querying for uncompleted requests when compare.immediate.if.request.found is true.

Default: 2000

compare.lock.timeout.ms

The amount of time a single compare worker node_communication lock will timeout after.

Default: 7200000

compare.max.inflight.count

Maximum number of outstanding changes for a table that is allowed for comparison. When the maximum is exceeded, the table request will be ignored.

Default: 10000

compare.queue.per.request.count

Number of threads to use for each compare request. See also compare.thread.per.server.count for overall thread pool.

Default: 4

compare.repair.batch.csv

If true, the repair script will be batch CSV and use the default data loader. The data loader is tolerant of different case for catalog, schema, and table name. If false, the repair script will be SQL with unquoted table names that is run on the database.

Default: true

compare.thread.per.server.count

The number of threads created that will be used to compare data concurrently on one node in the cluster.

Default: 10

compression.level

Set the compression level this node will use when compressing synchronization payloads. @see java.util.zip.Deflater NO_COMPRESSION = 0 BEST_SPEED = 1 BEST_COMPRESSION = 9 DEFAULT_COMPRESSION = -1

Default: -1

compression.strategy

Set the compression strategy this node will use when compressing synchronization payloads. @see java.util.zip.Deflater FILTERED = 1 HUFFMAN_ONLY = 2 DEFAULT_STRATEGY = 0

Default: 0

conflict.default.pk.with.fallback

When enabled, the default conflict detection uses the primary key and the resolution uses fallback, which was the original default in version 3.11 and older. This is a convenient way to go back to the old default without having to configure a conflict on each group link.

Default: false

console.admin.session.timeout

Timeout Administrator Sessions in Minutes

Default: 30

console.auth.client.cert

Allow console to authenticate user from a client certificate (X.509)

Default: false

console.auth.client.cert.subject.regex

Regular expression to find the username from the distinguished name (DN) subject of a client certificate (X.509). The first group in parenthesis will be used for the username. The default regex will match on the common name (CN) field of the subject.

Default: .CN=(.?),.*

console.auth.config.client

Allow configuration changes to occur on remote non-registration nodes.

Default: true

console.auth.fail.delay.ms

Delay in milliseconds between login prompts following a failed login attempt

Default: 5000

console.auth.ldap.admin.password

The password for the user specified by the console.auth.ldap.adminDN parameter.

Default:

console.auth.ldap.adminDN

The DN of the user to log in as while determining whether a user belongs to one of the groups specified by the console.auth.ldap.role.group.map parameter. If blank, the credentials from the login screen will be used.

Default:

console.auth.ldap.baseDN

The LDAP base DN to search for a user or group.

Default:

console.auth.ldap.host

The LDAP server host name.

Default:

console.auth.ldap.memberAttribute

The attribute of an LDAP group to check to determine whether or not a user is a member of the group. This only has an effect if the console.auth.ldap.role.group.map parameter is set and the console.auth.ldap.use.memberOf parameter is set to false.

Default: member

console.auth.ldap.role.group.map

Map of LDAP groups to console role IDs. Allows LDAP users to log in without having a corresponding console user if they are a member of one of the specified groups.

Default:

console.auth.ldap.searchAttribute

LDAP attribute to search for.

Default:

console.auth.ldap.use.memberOf

Whether or not the LDAP server maintains a "memberOf" attribute. This only has an effect if the console.auth.ldap.role.group.map parameter is set.

Default: true

console.auth.saml.groupAttribute

The attribute of an SAML response to check to determine whether or not a user is a member of a SAML group. This only has an effect if the console.auth.saml.role.group.map parameter is set.

Default: group

console.auth.saml.idp.cert

Path to the public x509 certificate of the SAML IdP.

Default:

console.auth.saml.idp.entityid

Identifier of the SAML IdP entity (must be a URI).

Default:

console.auth.saml.idp.sso.service.url

SSO endpoint info of the SAML IdP. URL target of the IdP where the SP will send the authentication request message.

Default:

console.auth.saml.organization.display.name

Display name of the organization within the SAML SP’s metadata

Default:

console.auth.saml.organization.language

Language of the organization within the SAML SP’s metadata

Default:

console.auth.saml.organization.name

Name of the organization within the SAML SP’s metadata

Default:

console.auth.saml.organization.url

URL of the organization within the SAML SP’s metadata

Default:

console.auth.saml.role.group.map

Map of SAML groups to console role IDs. Allows SAML users to log in without having a corresponding console user if they are a member of one of the specified groups.

Default:

console.auth.saml.sp.nameid.format

Specifies constraints on the name identifier to be used to represent the requested subject during SAML authentication.

Default: urn:oasis:names:tc:SAML:1.1:nameid-format:unspecified

console.auto.create.table.group.hiers.depth.limit

The maximum depth that the Auto Create Table Group Hierarchies dialog will go to when searching for child tables.

Default: 5

console.auto.create.table.group.hiers.table.limit

The maximum number of child tables that the Auto Create Table Group Hierarchies dialog will search for.

Default: 25

console.events.purge.retention.period

This is the retention time for the number of minutes that a console event will be retained.

Default: 43200

console.help.show

Display help text next to entry forms on screens by default.

Default: true

console.ip.addr.filter

Filters access by IP address using a regular expression match. For example, filter for local addresses is ^(10\.|192\.168|127).* Gives user a blank page. Set to "none" to disable all access.

Default:

console.log.slow.refresh.threshold.millis

The number of milliseconds before logging that the refresh of a screen is slow.

Default: 5000

console.node.connection.diagnostic.download.bandwidth.payloads

Console node connection diagnostic download bandwidth payloads

Default: 100000,1000000,10000000

console.node.connection.diagnostic.upload.bandwidth.payloads

Console node connection diagnostic upload bandwidth payloads

Default: 100000,1000000,10000000

console.password.encrypt.method

Password encryption type (SHA-1 or SHA-512)

Default: SHA-512

console.password.expire.days

Password expiration duration in days (0 = never expire)

Default: 0

console.password.failed.attempts

Number of failed login attempts before user gets disabled (0 = never disabled)

Default: 5

console.password.min.change.chars

The minimum number of characters that must be changed when changing a password

Default: 0

console.password.min.length

Minimum password length for new passwords (must be greater than 0)

Default: 6

console.password.otp

Required two-factor authentication for login

Default: false

console.password.prohibit.common.words

Prohibit commonly used words in passwords

Default: true

console.password.prohibit.previous

The number of recent passwords to prohibit when creating a new password (0 = no history is kept, -1 = infinite)

Default: 0

console.password.require.alphanum

Require an alphanumeric password

Default: true

console.password.require.mixed.case

Require mixed case in password

Default: true

console.password.require.symbol

Require a symbol in password

Default: true

console.password.salt

Use salt in password encryption

Default: true

console.password.verificationcode.timeout

Timeout for Verification Codes in Minutes

Default: 15

console.registration.method

Change the functionality of the registration of a node in the manage tab. The possible values are: wait-accept and pre-auth

Default: wait-accept

console.remote.meta.data

Retrieve meta data for catalog, schema, table per node group

Default: true

console.remote.meta.data.thread.count

Number of threads to be used for fetching meta data from remote nodes

Default: 5

console.report.as.offline.minutes

Setting that defines when a Node should be considered "offline." The offlineNodes, batchUnsent, and batchUnsentOffline monitors use this setting.

Default: 1440

console.sessions.max

The maximum number of sessions that can be active at a time (0 = unlimited)

Default: 20

console.sessions.per.user

The maximum number of sessions that can be active at a time per user (0 = unlimited)

Default: 1

console.user.disable.inactive.days

Number of consecutive days a user can be logged out before getting disabled (0 = unlimited)

Default: 0

console.user.session.timeout

Timeout User Sessions in Minutes

Default: 30

console.web.enabled

Determines if console is available or if a splash screen is displayed. Does not affect synchronization. Gives user a message about the web console is disabled. Requires restart on change.

Default: true

console.web.hide.system.info

Hide system information such as heartbeat and config information

Default: true

create.index.convert.unique.to.nonunique.when.columns.not.required

If set to true, then convert unique indexes to non-unique when one of the columns is defined as not required (defined with the NOT NULL constraint).

Default: true

create.table.not.null.columns.supported

If set to true, when a table’s schema is sent to the target database it will use NOT NULL statements to match the source. If this is false NOT NULL will not be included in the sql

Default: true

create.table.without.defaults

If set to true, when a table’s schema is sent to the target database default values will not be included.

Default: false

create.table.without.foreign.keys

If set to true, when a table’s schema is sent to the target database foreign keys will not be included.

Default: false

create.table.without.indexes

If set to true, when a table’s schema is sent to the target database, indexes will not be included.

Default: false

create.table.without.pk.if.source.without.pk

If set to true, when a table’s schema is sent to the target database it will not have all columns set as the primary key if the source does not have any primary keys.

Default: true

dashboard.refresh.rate.ms

The configured refresh rate of the dashboard screen.

Default: 10000

data.create_time.timezone

The timezone to use for create_time in the data table, e.g. +00:00. By default, when this is not set, trigger templates will normally call a now() type function.

Default:

data.flush.jdbc.batch.size

Flush size for JDBC batch mode used by services to save configuration.

Default: 10000

data.id.increment.by

This is the expected increment value for the data_id in the data table. This is useful if you use auto_increment_increment and auto_increment_offset in MySQL. Note that these settings require innodb_autoinc_lock_mode=0, otherwise the increment and offset are not guaranteed.

Default: 1

data.refresh.create.first

Whether or not to set create first flag on reload requests from the data refresh job. See also initial.load.defer.create.constraints which defaults to true. If setting both this and initial.load.defer.create.constraints to true, the reload request will remove constraints before loading the data, and add back the constraints afterwards.

Default: false

dataextractor.enable

Disable the extraction of all channels with the exception of the config channel

Default: true

dataextractor.text.column.expression

Provide an expression that will be used in the trigger templates, and in the initial load and the sym_data extraction SQL for all text based column values (like varchar, char, nvarchar, clob and nchar columns). The expression can be used to make scenario based casts. For example, if the data in the database was inserted under a different character set that the default character set on Oracle, then a helpful expression might be something like this: convert($(columnName), 'AR8ISO8859P6', 'AR8MSWIN1256')

Default:

dataloader.apply.changes.only

Indicates that old data should be used to create the update statement. If old data is equal to the new data and this property is set to true, then no update statement will be run.

Default: true

dataloader.create.table.alter.to.match.db.case

Whether to alter the case of the database tables that are created by the SymmetricDS data loader to match the default case of the target database.

Default: true

dataloader.create.table.without.defaults.on.error

If set to true, when a table creation fails on a database platform that is different than the source database, try to create the table without default values.

Default: true

dataloader.enable

Disable the loading of all channel with the exception of the config channel. This property can be set to allow all changes to be extracted without introducing other changes in order to allow maintenance operations.

Default: true

dataloader.error.save.curval

Indicates that the current value of the row should be recorded in the incoming_error table

Default: false

dataloader.fit.to.column

Indicate that the data loader should truncate data that is bigger than the target columns can handle. This applies to text-based columns only.

Default: false

dataloader.ignore.missing.tables

Tables that are missing at the target database will be ignored. This should be set to true if you expect that in some clients a table might not exist. If set to false, the batch will fail.

Default: false

dataloader.ignore.sql.event.errors

Indicate that the data loader should ignore errors while loading a SQL event and the execution of the statement fails.

Default: false

dataloader.log.sql.params.on.error

Indicate that the data loader should log SQL parameter values when a batch fails, which can be helpful for debugging. Since SQL parameters will contain application data, some sites may need to turn this off for policy compliance.

Default: true

dataloader.max.rows.before.commit

This is the maximum number of rows that will be supported in a single transaction. If the database transaction row count reaches a size that is greater than this number then the transaction will be auto committed. The default value of -1 indicates that there is no size limit.

Default: 10000

dataloader.sleep.time.after.early.commit

Amount of time to sleep before continuing data load after dataloader.max.rows.before.commit rows have been loaded. This is useful to give other application threads a chance to do work before continuing to load.

Default: 5

dataloader.text.column.expression

Provide a SQL expression that will be used by the data loader in DML statements for all text based column values (like varchar, char, nvarchar, clob and nchar columns). The expression can be used to make scenario based casts. For example, if the data in the database should be converted to a different character set that the default character set on Oracle, then a helpful expression might be something like this: convert($(columnName), 'AR8MSWIN1256', 'AR8ISO8859P6')

Default:

dataloader.use.primary.keys.from.source

Indicates that the database writer should use the primary keys from the source. Flip this to false if you want update and deletes to be based on the primary key as defined by the target table

Default: true

datareload.batch.insert.transactional

Indicate whether the process of inserting data, data_events and outgoing_batches for a reload is transactional. The only reason this might be marked as false is to reduce possible contention while multiple nodes connect for reloads at the same time.

Default: true

db.master.collation

For Sql Server, work around "Implicit conversion of varchar" issues by explicitly collating varchar columns in the database trigger. Relevant when the default database collation does not match the collation of the varchar columns of a table

Default:

db.treat.date.time.as.varchar.enabled

This is a setting that instructs the data capture and data load to treat JDBC TIME, DATE, and TIMESTAMP columns as if they were VARCHAR columns. This means that the columns will be captured and loaded in the form that the database stores them. Setting this to true on MySQL will allow datetime columns with the value of '0000-00-00 00:00:00' to be synchronized.

Default: false

dbf.router.validate.header

Determines if the *.DBF file headers should be validated when using the DBF Router

Default: true

default.values.to.leave.unquoted

A list of default values in CSV format that should not be quoted when synchronized from another node.

Default:

default.values.to.translate

A list of pairs of default values in CSV format that tells SymmetricDS to translate the first value to the second value when synchronized from another node.

Default:

extensions.xml

Spring xml configuration for extension points. This property enables maintaining Spring extension point configuration in the database. After changing this property a server restart is required.

Default:

<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
           xmlns:context="http://www.springframework.org/schema/context"
           xmlns:util="http://www.springframework.org/schema/util"
           xsi:schemaLocation="http://www.springframework.org/schema/beans
           http://www.springframework.org/schema/beans/spring-beans-3.0.xsd
           http://www.springframework.org/schema/util
           http://www.springframework.org/schema/util/spring-util-3.0.xsd
           http://www.springframework.org/schema/context
           http://www.springframework.org/schema/context/spring-context-3.0.xsd">
</beans>
external.id.is.unique.enabled

null

Default: true

extract.check.row.size

Determines if the size of a LOB value should be checked before extracting to prevent a JVM crash that can occur if the size of a LOB is bigger than the max size of a java array 2^31 - 1

Default: false

extract.row.capture.time

Extract the capture time of each row and put it in the batch to be used by the conflict manager for picking the winning row during a conflict. Enable for the best precision in resolution, which includes a unix timestamp for each occurrence of insert, update, and delete. Disable to include a single unix timestamp to represent the entire batch, when accuracy is less important or conflict management isn’t needed.

Default: true

extract.row.max.length

Used when the extract.check.row.size is true as an upper limit to check against. If the size exceeds the limit the row will be skipped and logged as a warning.

Default: 1000000000

file.compression.level

Set the compression level this node will use when compressing file synchronization payloads. @see java.util.zip.Deflater NO_COMPRESSION = 0 BEST_SPEED = 1 BEST_COMPRESSION = 9 DEFAULT_COMPRESSION = -1

Default: -1

file.pull.lock.timeout.ms

null

Default: 7200000

file.pull.period.minimum.ms

null

Default: 0

file.pull.thread.per.server.count

null

Default: 1

file.push.lock.timeout.ms

null

Default: 7200000

file.push.period.minimum.ms

null

Default: 0

file.push.thread.per.server.count

null

Default: 1

file.sync.azure.account.name

Azure Blob Storage Account Name for file sync

Default:

file.sync.azure.sas.token

Azure Blob Storage SAS token for file sync

Default:

file.sync.delete.ctl.file.after.sync

If the ctl file is used to control file triggers this will allow the system to remove the ctl file after sync but leave the source file.

Default: false

file.sync.delete.zip.file.after.sync

If set to true, this will delete the zip file from staging immediately after attempting to sync it to the target. If synchronization fails due to a timeout or an error, the zip file will have to be re-created during the next sync attempt. If set to false, this will leave the zip file in staging until it is purged by the Stage Management job.

Default: false

file.sync.enable

Enables File Synchronization capabilities

Default: false

file.sync.fast.scan

For Unix, fast scan will look for files that were modified since the last run of file sync tracker and commit their changes using the data loader max commit row setting. When it finds modified directories, it compares to the file snapshot to find changes. For a large file system, this is faster and more efficient than the normal tracker. This setting works on most Unix systems, but reportedly not working on some Windows systems.

Default: false

file.sync.lock.wait.ms

How long file sync should wait in millis for the exclusive lock used by file tracker or the shared lock used by file sync push/pull jobs.

Default: 300000

file.sync.prevent.ping.back

Record each file received in the sym_incoming_file table, which is checked when syncing outgoing files to prevent a "ping back" where the same file change is sent back and forth during bi-directional sync. If you aren’t using bi-directional sync, turn this off for better performance.

Default: true

file.sync.retry.count

How many times the file sync bean shell script will be attempted at the target node. The bean shell script will be attempted more than once if the exception java.nio.file.FileSystemException is thrown during the execution of the script.

Default: 2

file.sync.retry.delay.ms

How long to pause in milliseconds between the attempts of the bean shell script that is run at the target node when the execution of the script fails.

Default: 5000

file.sync.s3.account.key

S3 Access Key for Bucket Storage

Default:

file.sync.s3.region

S3 Region for Bucket Storage

Default:

file.sync.s3.secret.key

S3 Secret Key for Bucket Storage

Default:

file.sync.use.crc

Calculate a checksum for each file (using CRC32), which is used to detect a file collision if the target file has a different checksum. If you don’t need to detect conflicts, turn this off for better performance.

Default: true

file.sync.use.ctl.as.file.ext

If the ctl file is used to control file triggers this will look for a control file with the same name but .ctl replacing the existing extension. Example: temp.txt would need control file called temp.ctl instead of temp.txt.ctl

Default: false

filesynctracker.max.rows.before.commit

Maximum number of rows to put into sym_file_snapshot in a single database transaction. This will help keep sym_data freed up for other processing to occur.

Default: 10000

firebird.extract.varchar.row.old.pk.data

On Firebird database, the varchar sizes to use during extracting the row_data, old_data, and pk_data from sym_data. Specify the values as comma-separated for row, old, and pk respectively. By casting to varchar and using small sizes, performance of extract is improved. The entire row size must be under 64K or you will get a "block size exceeds implementation restriction" error. If you need to extract larger sizes, then enable the contains_big_lobs on the channel.

Default: 20000,20000,1000

google.bigquery.location

Google BigQuery location.

Default: US

google.bigquery.max.rows.per.rpc

Google BigQuery parameter to control the number of rows sent through the insertAll command while sending data into BigQuery.

Default: 100

google.bigquery.project.id

Google BigQuery project id.

Default:

google.bigquery.security.credentials.path

Google BigQuery security credentials path for the json file containing the credentials to connect to Big Query.

Default:

grouplet.enable

Enables the Grouplet functionality

Default: false

heartbeat.sync.on.push.enabled

Specify whether to push node_host records to configured push clients. If this is true the node for this instance and the node_host rows for all children instances will be pushed to all nodes that this node is configured to push to.

Default: true

heartbeat.sync.on.push.period.sec

This is the number of seconds between when the sym_node_host table’s heartbeat_time column is updated. This property depends on the frequency of the heartbeat job. If the heartbeat job is set to run every 10 minutes and this property is set to 10 seconds, then the heartbeat will only update every 10 minutes.

Default: 0

heartbeat.sync.on.startup

When this property is set to true the heartbeat process will run at server startup. Prior to 3.4 the heartbeat always happened at startup.

Default: false

heartbeat.update.node.with.batch.status

When this is set to true, SymmetricDS will update fields in the sym_node table that indicate the number of outstanding errors and/or batches it has pending

Default: false

http.compression

Whether or not to use compression over HTTP connections. Currently, this setting only affects the push connection of the source node. Compression on a pull is enabled using a filter in the web.xml for the PullServlet. @see web.compression.disabled to enable/disable the filter

Default: true

http.concurrent.workers.max

This is the number of HTTP concurrent push/pull requests SymmetricDS will accept. This is controlled by the NodeConcurrencyFilter. The number is per servlet the filter is applied to. The db.pool.max.active value should be twice this value plus some overhead (like 10 more than twice this value).

Default: 20

http.connect.timeout.ms

Sets the connection timeout on the internal HttpUrlConnection

Default: 90000

http.push.stream.output.enabled

The HTTP client connection, during a push, buffers the entire outgoing pay-load locally before sending it. Set this to true if you are getting heap space errors during a push. Note that basic auth may not work when this is turned on.

Default: true

http.push.stream.output.size

When HTTP chunking is turned on, this is the size to use for each chunk.

Default: 30720

http.timeout.ms

Sets the read timeout on the internal HttpUrlConnection

Default: 90000

http.transport.manager.class

Custom http transport manager supporting remote status.

Default: com.jumpmind.symmetric.console.remote.RemoteStatusHttpTransportManager

https.verified.server.names

During SSL handshaking, if the URL’s hostname and the server’s identification hostname mismatch, the verification mechanism will check this comma separated list of server names to see if the cert should be accepted (see javax.net.ssl.HostnameVerifier.) Set this value equal to 'all' if all server names should be accepted. Set this value to blank if a valid SSL cert is required.

Default:

hybrid.push.pull.buffer.status.updates

This controls whether or not the ReportStatus job will buffer its status updates. When buffered, status updates will only be sent when a channel’s batch to send count goes from 0 to non-zero.

Default: true

hybrid.push.pull.enabled

Enable hybrid push/pull functionality. Allows for pull configuration, but also allows clients to report their outgoing batch status so that the nodes that have data can be prioritized for pulling.

Default: false

hybrid.push.pull.timeout.ms

When hybrid.push.pull.enabled=true, how much time in millis has to go by to pull from a node that hasn’t reported pending batches.

Default: 3600000

incoming.batches.record.ok.enabled

Indicates whether batches that have loaded successfully should be recorded in the incoming_batch table. Note that if this is set to false, then duplicate batches will NOT be skipped because SymmetricDS will have no way of knowing that a batch has already loaded. This parameter can be set to false to reduce contention on sym_incoming_batch for systems with many clients.

Default: true

incoming.batches.skip.duplicates

This instructs symmetric to attempt to skip duplicate batches that are received. Symmetric might be more efficient when recovering from error conditions if this is set to true, but you run the risk of missing data if the batch ids get reset (on one node, but not another) somehow (which is unlikely in production, but fairly likely in lab or development setups).

Default: true

initial.load.after.sql

This is SQL that will run on the client after an initial load finishes. The default delimiter for these lines is a semicolon. To override, include a single line that starts with delimiter and is followed by the new delimiter, then the old For example a line that reads: delimiter $; would change sql lines to have a delimiter of $ for subsequent lines. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default:

initial.load.before.sql

This is SQL that will run on the client before an initial load starts. The default delimiter for these lines is a semicolon. To override, include a single line that starts with delimiter and is followed by the new delimiter, then the old For example a line that reads: delimiter $; would change sql lines to have a delimiter of $ for subsequent lines. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default:

initial.load.block.channels

Initial load and reload events should normally block other channels to ensure each table is loaded first followed by changes captured during the initial load. Setting this to false will allow all channels to load in priority order even when reload events or an initial load is running.

Default: true

initial.load.concat.csv.in.sql.enabled

Indicates that the SQL used to extract data from a table for an initial load should concatenate the data using the same SQL expression that a trigger uses versus concatenating the data in code.

Default: false

initial.load.create.first

Set this if tables should be created prior to an initial load. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default: false

initial.load.defer.create.constraints

If tables are created as part of the initial load, it will defer the creation of foreign keys and indexes to improve performance. After data is loaded, the constraints will be added after tables have reached consistency. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default: true

initial.load.defer.table.logging

If tables are created as part of the initial load, it will defer setting up table-level logging to improve performance. Applies to loads only. After data is loaded, the table-level logging will be switched on. This parameter needs set for the node that will send the initial load, not the node receiving it. Support for table-level logging varies by database dialect and is ignored where not applicable. Currently this is only supported by the PostgreSQL UNLOGGED feature. Oracle’s NOLOGGING feature might be implemented in the future.

Default: true

initial.load.delete.first

Set this if tables should be purged prior to an initial load. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default: false

initial.load.delete.first.sql

This is the SQL statement that will be used for purging a table during an initial load. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default: delete from %s

initial.load.extract.thread.per.server.count

The number of threads available for concurrent extracts of initial load batches.

Default: 20

initial.load.extract.timeout.ms

The number of milliseconds to wait until the lock will be broken on an initial load extract job.

Default: 7200000

initial.load.extract.use.two.pass.lob

Some dialects can extract small LOBs faster using a substring function. When the dialect supports it and the parameter is enabled, the initial load will extract tables with LOBs using two passes. The first pass gets rows with LOBs under the character limit, and the second pass gets all other rows. Currently implemented for Oracle/Tibero.

Default: true

initial.load.purge.stage.immediate.threshold.rows

If the number of rows in a reload batch is greater than or equal to this threshold, it will immediately purge the staging file after it is successfully loaded. Set this to -1 to disable and keep the staging files.

Default: 100

initial.load.recursion.self.fk

For tables with self-referencing foreign keys, the initial load will attempt to query each level of the rows starting with parent rows, as a way to preserve the order of rows.

Default: true

initial.load.reload.select.override

If set overrides the reload_select of the row in sym_table_reload_request when sending a load via the Load Data Wizard. The wizard will prompt the user to input any variables prefixed with a colon (:) that will be replaced in the final SQL.

Default:

initial.load.reverse.after.sql

This is SQL that will run on the server after a reverse initial load finishes.

Default:

initial.load.reverse.before.sql

This is SQL that will run on the server before a reverse initial load starts.

Default:

initial.load.reverse.first

Indicate that if both the initial load and the reverse initial load are requested, then the reverse initial load should take place first.

Default: true

initial.load.schema.dump.command

Specify a system command that writes the structure of the database to system.out to be captured and sent to the node that is being initial loaded. Used in conjunction with initial.load.schema.load.command. An example is: pg_dump --dbname=server --schema=my_schema --schema-only --clean

Default:

initial.load.schema.load.command

Specify a system command that will take the content captured by initial.load.schema.dump.command and apply it to the database. The content is passed to the system command via system.in. An example is: psql --output=output.log --dbname=client

Default:

initial.load.transport.max.bytes.to.sync

This is the number of maximum number of bytes to synchronize in one connect during an initial load.

Default: 524288000

initial.load.unblock.channels.on.error

Allow other channels to load when initial.load.block.channels is true and the reload channel goes into error. When the initial load runs while changes are being made, it can lead to foreign key errors that might resolve when the other channels are allowed to sync.

Default: false

initial.load.use.column.templates.enabled

The initial load SQL to extract a table will wrap each column with a template of SQL functions to format the data using the database. When disabled, each column is selected directly, and data is formatted by the server instead, which is faster. Currently, the server can format text, numeric, date, time, timestamp, and binary types, and it will fall back on using templates for other types.

Default: false

initial.load.use.estimated.counts

Estimate the row count for tables to generate the number of batches needed for initial load. It is faster to estimate the row count than to get an actual row count.

Default: true

initial.load.use.extract.job.enabled

Indicate that the extract job job should be used to extract reload batches

Default: true

initial.load.use.reload.channel

Indicate that the initial load events should be put on the reload channel. If this is set to false each table will be put on it’s assigned channel during the reload. This parameter needs set for the node that will send the initial load, not the node receiving it.

Default: true

job.compare.period.time.ms

null

Default: 10000

job.data.refresh.cron

null

Default: 0 0 0 * * *

job.file.sync.pull.period.time.ms

null

Default: 60000

job.file.sync.push.period.time.ms

null

Default: 60000

job.file.sync.tracker.cron

null

Default: 0 0/5 * * * *

job.heartbeat.period.time.ms

This is how often the heartbeat job runs. Note that this doesn’t mean that a heartbeat is performed this often. See heartbeat.sync.on.push.period.sec to change how often the heartbeat is sync’d

Default: 900000

job.initial.load.extract.period.time.ms

This is how often the initial load extract job will run in the background

Default: 10000

job.initial.load.queue.period.time.ms

This is how often the initial load queue job will run in the background

Default: 10000

job.monitor.period.time.ms

null

Default: 60000

job.offline.pull.period.time.ms

This is how often the offline pull job will be run to schedule offline reading of batch files from nodes.

Default: 60000

job.offline.push.period.time.ms

This is how often the offline push job will be run to schedule offline writing of batch files for nodes.

Default: 60000

job.pull.period.time.ms

This is how often the pull job will be run to schedule pulls of nodes.

Default: 10000

job.purge.first.pass

Enables a first pass purge for sym_data and sym_data_event that quickly purges the beginning of the table that precedes outstanding batches. These delete statements don’t use joins, so they run quicker.

Default: true

job.purge.first.pass.outstanding.batches.threshold

The maximum number of outstanding batches allowed for running the first pass purge. If there are too many outstanding batches, it will take too long to find their first data_id, so it shouldn’t be run.

Default: 100000

job.purge.incoming.cron

This is how often the incoming batch purge job will be run.

Default: 0 0 */4 * * *

job.purge.max.data.gaps.read

The maximum number of expired data gaps to check individually before using a single query to detect any expired data that was missed. The single query gets a range from sym_data before the starting gap that does not exist in sym_data_event, and then queries only the expired data gaps that are within that range.

Default: 100

job.purge.max.lingering.batches.read

The number of lingering batch IDs to read into memory at once before purging them from the database. A lingering batch is a batch that had a non-OK status when it became old enough to purge. The purge job runs a special routine to purge lingering batches after their status becomes OK, which allows the normal routine to move forward efficiently.

Default: 100000

job.purge.max.num.batches.to.delete.in.tx

This is the number of batches that will be purged in one database transaction.

Default: 5000

job.purge.max.num.data.event.batches.to.delete.in.tx

This is the number of batches that will be purged from the data_event table in one database transaction.

Default: 5

job.purge.max.num.data.to.delete.in.tx

This is the number of data ids that will be purged in one database transaction.

Default: 5000

job.purge.outgoing.cron

This is how often the outgoing batch and data purge job will be run.

Default: 0 0 */4 * * *

job.push.period.time.ms

This is how often the push job will be run to schedule pushes to nodes.

Default: 10000

job.refresh.cache.cron

This is when the refresh cache job will run.

Default: 0/30 * * * * *

job.report.status.cron

This is how often a client will push its status to the root server. Used in conjuction with hybrid.push.pull.

Default: 0 0/5 * * * *

job.routing.period.time.ms

This is how often the router will run in the background

Default: 10000

job.stage.management.cron

This is when the stage management job will run.

Default: 0 0 * * * *

job.stat.flush.cron

This is how often accumulated statistics will be flushed out to the database from memory.

Default: 0 0/5 * * * *

job.sync.config.cron

This is when the sync config job will run.

Default: 0 0/10 1 * * *

job.synctriggers.cron

This is when the sync triggers job will run.

Default: 0 0 0 * * *

job.watchdog.period.time.ms

null

Default: 3600000

jobs.synchronized.enable

If jobs need to be synchronized so that only one job can run at a time, set this parameter to true

Default: false

license.cache.time.ms

How long in milliseconds to cache license key. Useful for re-reading license key that was just installed to a cluster.

Default: 60000

lock.timeout.ms

The amount of time a thread can hold a shared or exclusive lock before another thread can break the lock. The timeout is a safeguard in case an unexpected exception causes a lock to be abandoned. Restarting the service will clear all locks.

Default: 1800000

lock.wait.retry.ms

While waiting for a lock to be released, how often should we check the lock status in the sym_lock table in the database.

Default: 10000

log.conflict.resolution

Whether logging is enabled for conflict resolution

Default: false

log.miner.flush.commit.size

Log miner commit size for saving change data capture rows.

Default: 1000

log.miner.mariadb.directory

Specifies the name and path for the directory that contains MariaDB’s binary log files. When blank, SymmetricDS will use each log file’s path as it appears in MariaDB’s binary log index file.

Default:

log.miner.mariadb.index

Specifies the name and path for MariaDB’s binary log index file. When blank, SymmetricDS will use MariaDB’s log_bin_index system variable.

Default:

log.miner.mssql.sort.by.transaction.id

Specifies that the order of the SQL Server log mining output should be ordered by transaction ID. If set to false, then the output is ordered by foreign key constraints.

Default: true

log.miner.mssql.sort.in.memory

For SQL-Server log miner, collect changes from all tables into memory, then sort them by change number and foreign keys. This attempts to preserve the order of changes across the database. When the same row is updated multiple times, SQL-Server will merge and return the last change number for it, which prevents accurate ordering. When disabled, the log miner collects and syncs changes from each table ordered by change number, processing each table in order by foreign keys.

Default: true

log.miner.mssql.sort.in.memory.max.rows

The maximum number of rows to sort in memory. When the number of changes exceeds the maximum, it will process like the sort in memory parameter is disabled.

Default: 1000000

log.miner.mssql.use.tsql.detect.changes

Use a single T-SQL statement that returns which tables have changes waiting. The single round trip with SQL-Server can save time. When disabled, a statement is run for each table to check for changes.

Default: true

log.miner.mysql.directory

Specifies the name and path for the directory that contains MySQL’s binary log files. When blank, SymmetricDS will use each log file’s path as it appears in MySQL’s binary log index file.

Default:

log.miner.mysql.index

Specifies the name and path for MySQL’s binary log index file. When blank, SymmetricDS will use MySQL’s log_bin_index system variable.

Default:

log.miner.xid.queue.threshold

Log miner number of changes per transaction that will be stored in memory before switching to writing them to disk instead. Set to -1 to always use memory.

Default: 100000

login.confirm.banner

Text displayed within a pop-up banner that must be accepted on login

Default:

login.intro.text

Intro text displayed at the top of the login dialog

Default:

mongodb.use.mongo.ids

If set assumes all tables replicating with MongoDB will have an _id column which is what MongoDB utilizes.

Default: false

monitor.events.capture.enabled

Enable capturing of monitor events and syncing to other nodes. When disabled, monitor events are still generated, just not synced to other nodes.

Default: true

mssql.allow.only.row.level.locks.on.runtime.tables

Automatically update data, data_event and outgoing_batch tables to allow only row level locking.

Default: true

mssql.auto.create.trigger.options

When auto-creating table triggers, some options will be detected and applied for the user, such as excluding ROWVERSION columns.

Default: true

mssql.bulk.load.bcp.cmd

Specifies the path to the SQL Server bulk copy program utility (bcp) executable.

Default:

mssql.bulk.load.bcp.keep.null.values

Specifies if null values should remain null when using the SQL Server bcp bulk loader to load into nullable columns that have default values.

Default: false

mssql.bulk.load.bcp.quoted.identifiers

Specifies that identifiers like database, schema, and tables names need to be quoted because they may contain special characters like spaces or dashes.

Default: true

mssql.bulk.load.bcp.trust.server.cert

Specifies if server cert should be trusted when connecting bcp bulk loader to SQL Server.

Default: false

mssql.bulk.load.bcp.use.instance

Specifies if the server instance should be used on the SQL Server bulk copy program utility (bcp) executable.

Default: true

mssql.bulk.load.field.terminator

Specify the field terminator used by the SQL Server bulk loader. Pick something that does not exist in the data in your database.

Default: |}

mssql.bulk.load.fire.triggers

Whether or not triggers should be allowed to fire when bulk loading data.

Default: false

mssql.bulk.load.max.rows.before.flush

Maximum number of rows to write to file before running with "BULK INSERT" to SQL-Server

Default: 100000

mssql.bulk.load.row.terminator

Specify the line terminator used by the SQL Server bulk loader. Pick something that does not exist in the data in your database.

Default: |>\n

mssql.bulk.load.unc.path

Specify a UNC network path to the tmp\bulkloaddir directory for SQL Server to access bulk load files. Use this property with bulk loader when SymmetricDS is on a separate server from SQL Server.

Default:

mssql.bulk.load.use.bcp

Enables the use of the SQL Server bulk copy program utility (bcp) as the bulk loader.

Default: false

mssql.include.catalog.in.triggers

Includes the catalog/database name within generated triggers (catalog.schema.table). May need turned off to support backup processes such as creating a bacpac file

Default: true

mssql.lock.escalation.disabled

Disables lock escalation and turns off page level locking. May need turned off to support backup processes such as creating a bacpac file

Default: true

mssql.trigger.execute.as

Specify the user the SymmetricDS triggers should execute as. Possible values are { CALLER | SELF | OWNER | 'user_name' }

Default: caller

mssql.trigger.order.first

Set the order of triggers to 'First' using sp_settriggerorder after creating triggers. This is needed when the user has existing custom triggers that modify data. The SymmetricDS triggers need to fire first and capture the first change so that order of changes is preserved. If the user has a trigger set as 'First', it will be changed to 'None'.

Default: false

mssql.use.ntypes.for.sync

Use ntext for the data capture columns and cast to nvarchar(max) in the trigger text so that nvarchar, ntext and nchar double byte data isn’t lost when the database collation for char types isn’t compatible with n char types.

Default: false

mssql.use.snapshot.isolation

Specifies if snapshot isolation should be automatically turned on.

Default: false

mssql.use.varchar.for.lob.in.sync

Use varchar(max) or nvarchar(max) when a column in SymmetricDS tables is set to long or nlong. This is for example necessary when using a _UTF8 of _SC collation.

Default: false

mysql.bulk.load.replace

Whether or not to replace rows that already exist, based on primary key or unique key. If set to false, duplicates will be skipped.

Default: true

mysql.tinyint.ddl.to.boolean

Allows MySQL columns of type tinyint to be sent to other platforms as boolean in ddl statements

Default: false

node.copy.mode.enabled

If the copy mode is enabled and the node starts up with an identity that does not match the configured external id, then the node will register with a special parameter that indicates the registration server should copy outgoing batch to the new node id.

Default: false

node.id.creator.script

This is a bean shell script that will be used to generate the node id for a registering node. This script is run on the registration server, not the registering node. The following variable are available for use: node, hostname, remoteHost, remoteAddress, and log. You can get the node group id by calling node.getNodeGroupId()

Default:

node.offline

Set the node to offline mode so that outgoing and incoming data are written to and read from a local directory instead of being sent over the network.

Default: false

node.offline.archive.dir

For a node operating in offline mode, specify the local directory where incoming data files should be moved to after successfully loading them. If this parameter is empty, files are removed after loading. The $(nodeGroupId) and $(nodeId) variables are useful when running multiple engines in the same server.

Default:

node.offline.error.dir

For a node operating in offline mode, specify the local directory where incoming data files should be moved to when they encounter an error during loading. To guarentee order of data loading, this parameter should be left empty so the file is not moved. The $(nodeGroupId) and $(nodeId) variables are useful when running multiple engines in the same server.

Default:

node.offline.incoming.accept.all

Accept batch data files from any node in the incoming directory, regardless of whether or not it is considered offline.

Default: true

node.offline.incoming.dir

For a node operating in offline mode, specify the local directory where data files should be read from. The $(nodeGroupId) and $(nodeId) variables are useful when running multiple engines in the same server.

Default: tmp/$(nodeGroupId)-$(nodeId)/offline/incoming

node.offline.outgoing.dir

For a node operating in offline mode, specify the local directory where data files should be written to. The $(nodeGroupId) and $(nodeId) variables are useful when running multiple engines in the same server.

Default: tmp/$(nodeGroupId)-$(nodeId)/offline/outgoing

node.password.failed.attempts

Number of failed login attempts by a node before lockout (0 = never lockout, -1 = never lockout or record)

Default: 5

node.screen.max.children.to.display

The maximum number of children to display under each node on the node screen before dividing them into smaller groups.

Default: 100

num.of.ack.retries

This is the number of times we will attempt to send an ACK back to the remote node when pulling and loading data.

Default: 5

offline.node.detection.period.minutes

This is the number of minutes that a node has been offline before taking action A value of -1 (or any negative value) disables the feature.

Default: 86400

offline.node.detection.restart.minutes

This is the number of minutes after a node has been restarted that it will begin considering remote nodes as offline. This gives remote nodes a chance to send their heartbeats.

Default: 5

offline.pull.lock.timeout.ms

The amount of time a single offline pull worker node_communication lock will timeout after.

Default: 7200000

offline.pull.thread.per.server.count

The number of threads created that will be used to read incoming offline batch data files

Default: 1

offline.push.lock.timeout.ms

The amount of time a single offline push worker node_communication lock will timeout after.

Default: 7200000

offline.push.thread.per.server.count

The number of threads created that will be used to write outgoing offline batch data files

Default: 1

opensearch.load.aws.access.key

The AWS secret access key (aws_secret_access_key) to use as credentials for uploading to S3

Default:

opensearch.load.aws.secret.key

The AWS secret access key (aws_secret_access_key) to use as credentials for uploading to S3

Default:

oracle.auto.create.trigger.options

When auto-creating table triggers, some options will be detected and applied for the user, including enabling stream LOB for LONG columns and excluded ROWID columns.

Default: true

oracle.bulk.load.field.terminator

Specify the field terminator used by the Oracler SQL*Loader. Pick something that does not exist in the data in your database.

Default: |}

oracle.bulk.load.line.terminator

Specify the line terminator used by the Oracler SQL*Loader. Pick something that does not exist in the data in your database.

Default: |>

oracle.load.query.hint.parallel.count

For initial load extracting data to specify the number of parallel processes to use while selecting data from a table

Default: 1

oracle.sequence.noorder

On Oracle RAC, an ordered sequence for sym_data must be coordinated across RAC nodes, which has wait overhead. By setting this to true, a no-order sequence is used instead, which performs better for high throughput. Because the sequence is no longer ordered, sym_data is queried using an order by of create_time and data_id. You will need to restart after changing this parameter to get DDL applied to the sequence and sym_data.

Default: false

oracle.sequence.noorder.nextvalue.db.urls

For Oracle RAC in no-order mode, this parameter provides two methods for managing data gaps across multiple nodes in the cluster. When left blank, routing will use gv$_sequences to manage gaps. Or, use this parameter to specify a comma-separated list of database URLs to connect to during the heartbeat, which ensures each RAC node has periodic activity within its gap to prevent the gap from expiring.

Default:

oracle.template.precision

This is the precision that is used in the number template for oracle triggers

Default: *,38

oracle.template.precision.text.minimum

Use the text minimum format model for capturing changes to number data types. When enabled, numbers are converted with to_char('TM'), which can capture up to 40 digits. When disabled, numbers are converted with cast to number(*,38), which can capture up to 38 digits.

Default: false

oracle.transaction.view.clock.sync.threshold.ms

Requires access to gv$transaction. This is the threshold by which clock can be off in an oracle rac environment. It is only applicable when oracle.use.transaction.view is set to true.

Default: 60000

oracle.use.ntypes.for.sync

Use nclob for the data capture columns and use to_nclob() in the trigger text so that double byte data isn’t lost when the database collation for char types isn’t compatible with n char types.

Default: false

oracle.use.transaction.view

Requires access to gv$transaction

Default: false

outgoing.batches.copy.to.incoming.staging

When sending an outgoing batch, copy directly from the outgoing staging to the incoming staging when both nodes are on the same server. This also requires the staging to be enabled (stream.to.file.enabled=true). The HTTP transport is still used to send a batch "retry" instruction that causes the target node to read from staging.

Default: true

outgoing.batches.max.to.select

The maximum number of unprocessed outgoing batch rows for a node that will be read into memory for the next data extraction.

Default: 50000

outgoing.batches.peek.ahead.batch.commit.size

This is the number of data events that will be batched and committed together while building a batch. Note that this only kicks in if the prospective batch size is bigger than the configured max batch size.

Default: 10

outgoing.batches.update.status.data.count

Update the outgoing batch status to QY (querying) and SE (sending) when the data event count is larger than this threshold. This can improve performance overhead on small batches by avoiding status updates.

Default: 1000

outgoing.batches.update.status.millis

Update the outgoing batch status to QY (querying) and SE (sending) only when the last update to the batch is in the past by at least the specified number of milliseconds. This can improve performance overhead on small batches by avoiding status updates.

Default: 10000

parameter.reload.timeout.ms

The number of milliseconds parameters will be cached by the ParameterService before they are reread from the file system and database.

Default: 600000

postgres.security.definer

Postgres triggers default to "security invoker" with permissions based on caller. Enable this parameter to use "security definer" with permissions based on owner.

Default: false

previous.login.info.notification.enabled

Enable previous login information notification

Default: true

pull.immediate.if.data.found

If true, the pull job will re-pull immediately after it receives data.

Default: true

pull.lock.timeout.ms

The amount of time a single pull worker node_communication lock will timeout after.

Default: 7200000

pull.period.minimum.ms

This is the minimum time that is allowed between pulls of a specific node.

Default: 0

pull.thread.per.server.count

The number of threads created that will be used to pull nodes concurrently on one server in the cluster.

Default: 10

purge.compare.request.retention.minutes

This is the retention time for how long a compare request will be retained

Default: 14400

purge.expired.data.gap.retention.minutes

This is the retention time for how long an expired data gap will be retained. The purge job will watch for data to commit inside of expired data gaps, and try to resend the data. Knowing about expired data gaps lets the purge use a fast range-based delete but avoid purging any ranges that could include expired data gaps. (Checking for stranded data before running a purge isn’t sufficient because it leaves open a race condition of the data arriving during the deletes.)

Default: 1440

purge.extract.request.retention.minutes

This is the retention time for how long a extract request will be retained

Default: 14400

purge.log.summary.retention.minutes

This is the retention for how long log summary messages will be retained in memory.

Default: 60

purge.monitor.event.retention.minutes

This is the retention time for how long to keep monitor events before purging them.

Default: 43200

purge.registration.request.retention.minutes

This is the retention time for how long a registration request will be retained

Default: 7200

purge.retention.minutes

This is the retention for how long synchronization data will be kept in the symmetric synchronization tables. Note that data will be purged only if the purge job is enabled.

Default: 120

purge.snapshot.file.retention.minutes

This is the retention time for how long support snapshot files will be retained

Default: 14400

purge.stats.retention.minutes

This is the retention for how long statistic data will be kept in the symmetric stats tables. Note that data will be purged only if the statistics flush job is enabled.

Default: 14400

purge.trigger.hist.retention.minutes

This is the retention time for how long an inactive trigger history will be retained

Default: 86400

push.immediate.if.data.found

If true, the push job will re-push immediately after it sends data.

Default: true

push.lock.timeout.ms

The amount of time a single push worker node_communication lock will timeout after.

Default: 7200000

push.period.minimum.ms

This is the minimum time that is allowed between pushes to a specific node.

Default: 0

push.thread.per.server.count

The number of threads created that will be used to push to nodes concurrently on one server in the cluster.

Default: 10

redshift.append.to.copy.command

The value of this property will be appended to the end of the copy command when the redshift data loader is enabled.

Default:

redshift.bulk.load.max.bytes.before.flush

Maximum number of bytes to write to file before copying to S3 and running with COPY statement

Default: 1000000000

redshift.bulk.load.max.rows.before.flush

Maximum number of rows to write to file before copying to S3 and running with COPY statement

Default: 100000

redshift.bulk.load.s3.access.key

The AWS access key ID (aws_access_key_id) to use as credentials for uploading to S3

Default:

redshift.bulk.load.s3.bucket

The S3 bucket where bulk load files will be uploaded to before bulk loading into Redshift

Default:

redshift.bulk.load.s3.endpoint

The endpoint for the s3 bucket. If not set it will use the default endpoint.

Default:

redshift.bulk.load.s3.secret.key

The AWS secret access key (aws_secret_access_key) to use as credentials for uploading to S3

Default:

registration.auto.create.group.link

When this is set to true a group link will be created by default between two groups even if the user does not explicitly set one up.

Default: true

registration.number.of.attempts

This is the number of times registration will be attempted before being aborted. The default value is -1 which means an endless number of attempts. This parameter is specific to the node that is trying to register, not the node that is providing registration.

Default: -1

registration.push.config.allowed

When group link is configured to push to clients, allow registration to work over push. Useful when client cannot reach server, so server pushes registration to client.

Default: true

registration.reopen.use.same.password

Indicates that if registration is reopened if the same password should be used. If set to false then a new password will be generated.

Default: true

registration.require.initial.load

When this is set to true a node will have to have had an initial load prior to allowing other nodes to register with it

Default: true

registration.secret

Send a shared secret as part of the registration request. If the client and server secrets are non-blank and match, the registration is automatically allowed.

Default:

remote.status.connect.timeout.ms

Amount of time a thread will wait for a remote SymmetricDS node to connect.

Default: 900000

remote.status.enabled

Enables the use of web sockets to get the remote status of nodes for the web user interface

Default: true

remote.status.follow.group.link.direction

Determines whether remote status calls will be limited to the group link configuration.

Default: false

remote.status.follow.group.link.fallback.count

Number of failed connections to remote node before falling back to follow group link direction = true

Default: 10

remote.status.idle.timeout.ms

Idle timeout of a remote status web socket connection

Default: 300000

remote.status.max.threads

Max number of threads allowed to be used by Remote Status requests.

Default: 20

remote.status.send.timeout.ms

Send timeout of a remote status web socket connection

Default: 60000

rest.api.enable

Enables the REST API

Default: false

rest.api.heartbeat.on.pull

Enables the REST API to update the heartbeat when pulling data

Default: false

right.trim.char.values

Whether char fields should be right trimmed

Default: false

route.on.extract

Whether the routing job will start manually when a push or pull is started

Default: false

routing.collect.stats.unrouted

Enable to collect unrouted data statistics into the stat tables for graphs.

Default: false

routing.data.reader.into.memory.enabled

Router will read all unrouted data into memory, then perform sorting. Enable this option if sorting is expensive for the database to perform.

Default: false

routing.data.reader.order.by.gap.id.enabled

Use the order by clause to order sym_data when selecting data for routing. Most databases order the data naturally and might even have better performance when the order by clause is left off.

Default: true

routing.data.reader.threshold.gaps.to.use.greater.than.query

Select data to route from sym_data using a simple > start_gap_id query if the number of gaps in sym_data_gap are greater than the following number

Default: 0

routing.data.reader.use.multiple.queries

Routing reader may run multiple queries for data, with each query including the maximum number of data gaps as specified by the routing.max.gaps.to.qualify.in.sql parameter. This method attempts to use the table’s index for quick results and avoid wasting time on filtering rows that were already routed (as used by the greater.than.query method).

Default: true

routing.detect.invalid.gaps

Run checks for duplicate, invalid range, overlapping, and large gaps while processing each gap. This can be used to log information and catch problems with gap detection, but it incurs additional overhead.

Default: true

routing.flush.batches.jdbc.batch.size

JDBC batch size for sym_outgoing_batch

Default: 5000

routing.flush.jdbc.batch.size

JDBC batch size for sym_data_event

Default: 50000

routing.gaps.transaction.view.clock.sync.threshold

When using the earliest transaction time to expire gaps (routing.gaps.use.transaction.view=true), subtract the given number of milliseconds from the transaction time. This may be needed in a clustered environment where the time on each database master is not perfectly in sync.

Default: 10000

routing.immediate.if.max.channel

Start routing again immediately when that last run of routing reached the max data to route for a channel.

Default: true

routing.largest.gap.size

This is the maximum number of data that will be routed during one run. It should be a number that well exceeds the number rows that will be in a transaction.

Default: 50000000

routing.lock.timeout.ms

The amount of time a single routing worker node_communication lock will timeout after.

Default: 7200000

routing.log.stats.on.batch.error

Enable to collect routing statistics for each batch and log the statistics when a batch goes into error.

Default: false

routing.max.batch.size.exceed.percent

The percentage of the channel’s max batch size that a batch can exceed when seeking a transaction boundary using the default batch algorithm. Use zero to indicate that the batch size can grow as large as needed to include the complete transaction. For example, a setting of 100 percent with a channel’s max batch size of 1000 allows a batch size of 2000 to be routed before it will be forced as complete. This setting protects from a large transaction that causes batch sizes that far exceed the channel’s max size and have trouble loading on the target database.

Default: 100

routing.max.gap.changes

This is the maximum number of changes that can be applied to the data_gap table. If the gap detection exceeds this number of changes, it will record the minimal gaps to the table and keep the rest in memory. This setting only applies to non-clustered systems.

Default: 200

routing.max.gaps.to.qualify.in.sql

This is the number of gaps that will be included in the SQL that is used to select data from sym_data. If there are more gaps than this number, then the last gap will in the SQL will use the end id of the last gap.

Default: 100

routing.peek.ahead.memory.threshold.percent

When reading data to route, if a lot of data has been committed in a single transaction, then the peek ahead queue size can cause out of memory errors. This setting instructs the routing reader to disperse with all "non" active transactions if the peek ahead queue grows to be a certain percentage of the overall allocated heap size.

Default: 50

routing.peek.ahead.window.after.max.size

This is the maximum number of events that will be peeked at to look for additional transaction rows after the max batch size is reached. The more concurrency in your db and the longer the transaction takes the bigger this value might have to be.

Default: 2000

routing.query.channels.first

Enable to query for which channels have data waiting, and then only route for those channels.

Default: true

routing.stale.dataid.gap.time.ms

This is the time that any gaps in data_ids will be considered stale and skipped.

Default: 1200000

routing.stale.gap.busy.expire.time.ms

For a busy system, how often to run checks on sym_data in order to expire gaps. Normally the routing reads all data and gap expiration can run without checking sym_data. But when the system is busy, then not all data is read, and gap expiration must query each gap from the sym_data table, which is expensive.

Default: 7200000

routing.thread.per.server.count

The number of threads created that will be used to route channels concurrently on one server in the cluster.

Default: 5

routing.use.channel.threads

When enabled, use a thread per channel for parallel routing.

Default: false

routing.use.common.groups

Use common mode for groups of target nodes, regardless of channel configuration. A batch in common mode has only one data_event for each data row, and it extracts once for all nodes that receive it.

Default: true

routing.use.non.common.for.incoming

When using common batch mode for groups (and sync on incoming enabled for table triggers), if a change comes from a remote node, then use non-common batches instead. When a server routes each incoming remote change back to all nodes minus the source node, there are as many common groups as nodes. As the number of nodes increases, the number of small batches increases, and it can be more efficient to use non-common mode for routing incoming remote changes into a small number of large batches.

Default: true

routing.wait.for.data.timeout.seconds

null

Default: 330

schema.version

This is hook to give the user a mechanism to indicate the schema version that is being synchronized.

Default: ?

send.ack.keepalive.ms

After a push or pull HTTP connection has been idle for this many milliseconds, a small partial acknowledgement or partial batch is sent to keep the connection alive.

Default: 30000

send.usage.stats

Send usage statistics to the SymmetricDS project to help improve the software.

Default: true

server.engine.uri.interceptors

A comma-seperated list of custom interceptors which wrap URI handlers.

Default: com.jumpmind.symmetric.console.remote.RemoteStatusInterceptor

single.store.audit.log.dir

The path to the auditlogs dir for SingleStore

Default: /var/lib/memsql/{cluster-id}/auditlogs

smtp.allow.untrusted.cert

Whether or not to accept an untrusted certificate for SSL/TLS when connecting to the mail server.

Default: false

smtp.auth

Whether or not to authenticate with the mail server.

Default: false

smtp.from

The email address to use in the "from" header when sending email.

Default: symmetricds@localhost

smtp.host

The hostname of the mail server

Default: localhost

smtp.password

When authenticating with the mail server, the password to use.

Default:

smtp.port

The port number of the mail server

Default: 25

smtp.starttls

Whether or not to use TLS after connecting to the mail server.

Default: false

smtp.transport

The transport type to use when connecting to mail server, either smtp or smtps.

Default: smtp

smtp.user

When authenticating with the mail server, the username to use.

Default:

snapshot.file.include.hostname

Prefix snapshot filename with hostname to indicate which environment it came from

Default: false

snapshot.max.batches

Max number of batches to write to statistics listing for support snapshot.

Default: 10000

snapshot.max.files

Max number of files to write in directory listing for support snapshot.

Default: 50000

snapshot.max.node.channels

Max number of nodes and channels for batch statistics, after which it will group by node only.

Default: 5000

snapshot.operation.timeout.ms

Max time for a snapshot operation to complete, such as gathering table definitions, before it will be interrupted so the snapshot completes in a reasonable amount of time.

Default: 30000

snowflake.internal.stage.name

The snowflake managed stage name for internal storage

Default:

snowflake.staging.type

The snowflake staging type to use for bulk loading Valid values : SNOWFLAKE_INTERNAL, AWS_S3, AZURE

Default:

spatial.data.types.enabled

Determines if spatial data type functions will be installed. By default they will be installed but can be set to false to not install the additional spatial functions if they are not needed.

Default: true

sql.explorer.auto.commit

Whether to auto commit sql executed in the sql explorer

Default: true

sql.explorer.delimiter

Delimiter to use when executing sql

Default: ;

sql.explorer.max.results

Max number of rows to pull back as part of a query

Default: 100

sql.explorer.result.as.text

Show the results of a query in as text

Default: false

sql.explorer.show.row.numbers

Show the row numbers in a table

Default: true

staging.screen.edit.max.bytes

The maximum size of a file that can be edited via the staging area screen.

Default: 25000000

start.offline.pull.job

Whether the offline pull job is enabled for this node.

Default: false

start.offline.push.job

Whether the offline push job is enabled for this node.

Default: false

start.sync.config.job

Whether the sync config job is enabled for this node. This job checks that the configuration version matches the software version, otherwise it will pull the latest configuration from the registration server.

Default: true

stream.to.file.min.ttl.ms

If stream.to.file.enabled is false and staging is purged based on the database, then this is the minimum amount of time a staging file will be retained after it is purged from the database

Default: 1800000

stream.to.file.purge.on.ttl.enabled

When this is set to false, then batches in the staging area will only be purged after they have been purged from the database. If this is set to true, then batches will be purged based on the stream.to.file.ttl.ms setting.

Default: false

stream.to.file.threshold.bytes

If stream.to.file.enabled is true, then the threshold number of bytes at which a file will be written is controlled by this property. Note that for a synchronization the entire payload of the synchronization will be buffered in memory up to this number (at which point it will be written and continue to stream to disk.)

Default: 0

stream.to.file.ttl.ms

If stream.to.file.enabled is true, then this is how long a file will be retained in the staging directory after it has been marked as done.

Default: 3600000

sybase.allow.only.row.level.locks.on.runtime.tables

Automatically alter data, data_event and outgoing_batch tables to allow only row level locking.

Default: true

sybase.ase.convert.unitypes.for.sync

Converts unitypes in Sybase ASE to the corresponding type in the target database.

Default: false

sybase.change.identity.gap.on.runtime.tables

Automatically change attribute on data, data_event and outgoing_batch tables to set the identity gap to the number provided. This prevents skipping a large number of identities that can cause routing to stop. Use 0 to disable.

Default: 1000

sync.triggers.fix.duplicate.active.trigger.histories

Whether or not duplicate active trigger histories will be detected and fixed so that the most recent will end up being the only active trigger history.

Default: true

sync.triggers.thread.count.per.server

Number of threads to use for creating triggers and removing old ones.

Default: 1

sync.triggers.timeout.in.seconds

Time out value in milliseconds for sync triggers. This timeout value will be used to time out the inactivate step and the create or update step.

Default: 3600

target.treat.binary.as.lob.enabled

Whether target binary fields should be treated as lobs

Default:

tibero.bulk.load.field.terminator

Specify the field terminator used by the Tibero tbLoader. Pick something that does not exist in the data in your database.

Default: |}

tibero.bulk.load.line.terminator

Specify the line terminator used by the Tibero tbLoader. Pick something that does not exist in the data in your database.

Default: |>

time.based.capture

Enables time based capture miner when the log miner job is running to select changes out of tables based on a date time column or numeric primary key

Default: false

time.between.ack.retries.ms

This is the amount of time to wait between trying to send an ACK back to the remote node when pulling and loading data.

Default: 5000

topbar.system.text

System identification text shown in top bar in right corner

Default:

transport.max.bytes.to.sync

This is the number of maximum number of bytes to synchronize in one connect.

Default: 104857600

transport.max.error.millis

Networks errors will be logged at INFO level since they are retried. After the maximum number of millis for network errors that continue in succession, the logging switches to WARN level.

Default: 300000

trigger.allow.create.or.replace

Enable or disabled use of create or replace syntax on Oracle and MS-SQL 2016 SP1 and newer.

Default: true

trigger.capture.ddl.changes

Feature to install a DDL trigger to capture any schema changes, including tables, views, triggers, functions, and stored procedures, which are synced to all nodes on configured group links. Supported on MS SQL-Server, Oracle and Postgres only.

Default: false

trigger.capture.ddl.check.trigger.hist

Disable this property to capture all DDL changes regardless of whether the associated table has a corresponding row in sym_trigger_hist. MS SQL-Server, Oracle and Postgres only. See: trigger.capture.ddl.changes

Default: true

trigger.capture.ddl.delimiter

The delimiter to use when capturing changes from a DDL trigger. MS SQL-Server, Oracle and Postgres only. See: trigger.capture.ddl.changes

Default: $

trigger.create.before.initial.load.enabled

Disable this property to prevent table triggers from being created before initial load has completed.

Default: true

trigger.update.capture.changed.data.only.enabled

Enable this property to force a compare of old and new data in triggers. If old=new, then don’t record the change in the data capture table. This is currently supported by the following dialects: mysql, oracle, db2, postgres, sql server

Default: false

trigger.use.insert.delete.for.primary.key.changes

For Sybase and SQL Server, if a primary key is changed, this parameter determines if a delete followed by an insert is captured for the row change instead of an update. When set to true, an update will be created if no primary keys are changed or if only one row is changed when a primary key change occurs, otherwise a delete followed by an insert is created for each row updated when a primary key change occurs and more than one row is updated by the SQL statement. If set to false, it will always create an update. This parameter, when changed, requires a restart of the SymmetricDS instance, followed by a rebuild of the triggers.

Default: true

web.compression.disabled

Disable compression from occurring on Servlet communication. This property only affects the outbound HTTP traffic streamed by the PullServlet and PushServlet.

Default: false

B.3. Server Configuration

Server configuration is read from conf/symmetric-server.conf for settings needed by the server before the parameter system has been initialized.

host.bind.name

Specify the hostname/IP address to bind to. (Default 0.0.0.0 will bind to all interfaces.)

Default: 0.0.0.0

http.enable

Enable synchronization over HTTP.

Default: true

http.port

Port number for synchronization over HTTP.

Default: 31415

https.allow.self.signed.certs

Use a trust manager that allows self-signed server SSL certificates.

Default: true

https.enable

Enable synchronization over HTTPS (HTTP over SSL).

Default: false

https.need.client.auth

Requires client authentication with SSL certificate

Default: false

https.port

Port number for synchronization over HTTPS (HTTP over SSL).

Default: 31417

https.verified.server.names

List host names that are allowed for server SSL certificates.

Default: all

https.want.client.auth

Accepts client authentication with SSL certificate

Default: false

https2.enable

Enable HTTPS/2 for multiplexing and resistance to protocol attacks.

Default: false

server.cookie.name

Custom cookie name for session (defaults to JSESSIONID_{port})

Default:

server.http.cookies.enabled

Accept cookies if load balancer requires it for clustering

Default: false

Appendix C: Database Notes

This section describes specific settings and notes for using each supported database platform.

C.1. Compatibility

Each database management system has its own characteristics that results in feature coverage in SymmetricDS. The following table shows which features are available by database.

Table 86. Support by Database
Database Versions Transaction Identifier Data Capture Log Mining Conditional Sync Update Loop Prevention BLOB Sync CLOB Sync Transactional DDL Supports Bulk Loading Supports Bulk Replace

BigQuery *

All

DB2

9.5

DB2

10,11

DB2 for IBM i *

6

DB2 for IBM z/OS *

10

Derby

10.3.2.1

Elasticsearch *

7.13.2

Firebird

2.0

Greenplum

8.2.15 and above

H2

1.x, 2.1, 2.2

HSQLDB

1.8

HSQLDB

2.0

Informix

11

Ingres

11 and above

Interbase

9.0

MySQL

5.0.2 and above

MariaDB

5.1 and above

NuoDB

2.6 and above

OpenEdge

Tested on 12.2

Oracle

10g and above

✔ 11g and above

PostgreSQL

8.2.5 and above

✔ 8.3 and above

✔ 9.4 and above

Redshift

1.0

Snowflake *

6.13 and above

SingleStore *

All

SQL Anywhere

9

SQL Server

2005 and above

✔ 2008 and above

SQL Server Azure

Tested on 11.00.2065

SQLite

3.x

Sybase ASE

12.5

Teradata

15.10 and above

Tibero

6 and above

Database *

Asterisk (*) next to database indicates the database support is available in SymmetricDS Pro only.

Transaction Identifier

A transaction identifier is recorded in the SYM_DATA table along with changes, which allows changes in the same transaction to be grouped together for commit within a batch.

Data Capture

Changes to tables can be captured using database triggers.

Log Mining

Changes to tables can be captured using log mining. This feature is available in SymmetricDS Pro only

Conditional Sync

Conditions can be specified on SYM_TRIGGER, which are compiled into the trigger to decide if a change should be captured.

Update Loop Prevention

The remote node is recorded on data that is captured, so the system can prevent the changes from being sent back to the same node.

BLOB Sync

Binary large object data can be captured or streamed from the database.

CLOB Sync

Character large object data can be captured or streamed from the database.

C.2. Catalog and Schema

A relational database may be divided into catalogs that contain sub-databases called schemas, which contain tables. Each database management system can implement the concepts of catalog and schema differently or not at all. When locating a table, SymmetricDS uses the default catalog and schema unless the user specifies one.

Table 87. Catalog and Schema Support by Database
Database Version Catalog Support Catalog Default Schema Support Schema Default

BigQuery

(dataset name)

DB2

values current schema

Derby

values current schema

Elasticsearch

Firebird

Greenplum

select current_schema()

H2

select database()

select schema()

HSQLDB

1.0

HSQLDB

2.0

select value from information_schema.system_sessioninfo where key = 'CURRENT SCHEMA'

select value from information_schema.system_sessioninfo where key = 'CURRENT SCHEMA'

Informix

select trim(user) from sysmaster:sysdual

Ingres

select dbmsinfo('username')

Interbase

MySQL

select database()

MariaDB

select database()

NuoDB

select current_schema from system.dual

OpenEdge

select user() from sysprogress.syscalctable

Oracle

select sys_context('USERENV', 'CURRENT_SCHEMA') from dual

PostgreSQL

select current_schema()

Snowflake

select current_database()

select current_schema()

SingleStore

select database()

SQL Anywhere

select db_name()

select user_name()

SQL Server

2000

select db_name()

select 'dbo'

SQL Server

2005+

select db_name()

select schema_name()

SQL Server

select db_name()

select schema_name()

SQL Server Azure

select db_name()

select schema_name()

SQLite

Sybase ASE

select db_name()

select user_name()

Redshift

select current_schema()

Tibero

select sys_context('USERENV', 'CURRENT_SCHEMA') from dual

C.3. Google BigQuery

Use symadmin module install bigquery to install driver files, or copy your own files into the lib sub-directory.

Send changes from your relational database to Google’s BigQuery.

C.3.1. Setup

BigQuery is only supported as a load only node in SymmetricDS. See Load Only Node for details on setting up a load only node in SymmetricDS.

Setup the BigQuery node by using the Connect Database wizard and selecting BigQuery as the type.

bigquery database settings

After hitting next you can setup advanced options for your BigQuery node.

C.3.2. Loading Data

BigQuery Authentication

Create a service account within your project. This account will need to be given the roles/bigquery.admin role. Create a JSON credentials file through this service account.

Provide this file path on the advanced settings while setting up a BigQuery node. The advanced settings also requires that you provide a project ID and location for your BigQuery project.

You will need to use your Google BigQuery dataset name in the target schema of the router that is used to route data to BigQuery.
bigquery advanced settings

C.4. DB2

Use symadmin module install db2 to install driver files, or copy your own files into the lib sub-directory.

The IBM DB2 Dialect uses global variables to enable and disable node and trigger synchronization. These variables are created automatically during the first startup. The DB2 JDBC driver should be placed in the "lib" folder.

Currently, the DB2 Dialect for SymmetricDS does not provide support for transactional synchronization. Large objects (LOB) are supported, but are limited to 16,336 bytes in size. The current features in the DB2 Dialect have been tested using DB2 9.5 on Linux and Windows operating systems.

There is currently a bug with the retrieval of auto increment columns with the DB2 9.5 JDBC drivers that causes some of the SymmetricDS configuration tables to be rebuilt when auto.config.database=true. The DB2 9.7 JDBC drivers seem to have fixed the issue. They may be used with the 9.5 database.

A system temporary tablespace with too small of a page size may cause the following trigger build errors:

SQL1424N Too many references to transition variables and transition table
columns or the row length for these references is too long. Reason
code="2". LINE NUMBER=1. SQLSTATE=54040

Simply create a system temporary tablespace that has a bigger page size. A page size of 8k will probably suffice.

CREATE BUFFERPOOL tmp_bp PAGESIZE 8k;

CREATE SYSTEM TEMPORARY TABLESPACE tmp_tbsp
     PAGESIZE 8K
     MANAGED BY SYSTEM
     USING ('/home/db2inst1/tmp_tbsp')
         BUFFERPOOL tmp_bp
Table 88. Supported Data Types
Data Type Supported?

Char, VarChar, Long VarChar

Yes

Graphic, VarGraphic, Long VarGraphic

Yes

SmallInt, Integer, BigInt

Yes

Double

Yes

Decimal

Yes

Date, Time, TimeStamp

Yes

Blob, Clob, DBClob

Yes

DecFloat

No

Binary, VarBinary

No

By default DB2 will not capture the transaction id associated with the captured data. This can be turned on with the following parameter.

db2.capture.transaction.id=false

C.5. DB2 for IBM i

Use symadmin module install db2 to install driver files, or copy your own files into the lib sub-directory.

The DB2 for IBM i dialect can connect to a database on IBM iSeries (AS/400) machines. It was tested with the jt400 JDBC driver, which is already included in the SymmetricDS download. Here is an example JDBC URL:

jdbc:as400://hostname/myschema

The "libraries" parameter may be used in some cases to resolve unqualified object names:

jdbc:as400://hostname/;libraries=myschema
The tables created by SymmetricDS must have journaling enabled for commitment control.

C.5.1. Auto Journaling

The SymmetricDS library will be automatically journaled if it is created using the CREATE SCHEMA or CREATE COLLECTION SQL commands.

Otherwise, journaling can be enabled for new tables automatically by creating a default journal named QSQJRN in the library. The following steps add automatic journaling to the "sym" library (change it to your library) using the OS/400 command line:

  • Create the journal receiver object:

CRTJRNRCV JRNRCV(sym/symjrnrcv)
  • Create the journal object:

CRTJRN JRN(sym/QSQJRN) JRNRCV(sym/symjrnrcv)

C.5.2. Manual Journaling

Using automatic journaling for the SymmetricDS library is the preferred method, but journaling can also be enabled for each table manually. After starting SymmetricDS for the first time, it will connect to the database and create the required tables. Then it will log an error message that journaling needs to be enabled for its tables. The following steps add journaling to the "sym" library (change it to your library) using the OS/400 command line:

  • Create a journal receiver object:

CRTJRNRCV JRNRCV(sym/symjrnrcv)
  • Create a journal object:

CRTJRN JRN(sym/symjrn) JRNRCV(sym/symjrnrcv)
  • Start journaling:

STRJRNPF FILE(sym/SYM_C00001) JRN(sym/symjrn)

This step needs to be repeated for each physical file (table) created by SymmetricDS. A single command can be run for all tables at once, like this:

CALL QCMD
<hit F11 for more lines>

STRJRNPF FILE(sym/SYM_C00001 sym/SYM_C00002 sym/SYM_C00003 sym/SYM_C00004 sym/SYM_C00005 sym/SYM_C00006 sym/SYM_D00001 sym/SYM_D00002 sym/SYM_DATA sym/SYM_E00001 sym/SYM_E00002 sym/SYM_F00001 sym/SYM_F00002 sym/SYM_F00003 sym/SYM_F00004 sym/SYM_G00001 sym/SYM_G00002 sym/SYM_I00005 sym/SYM_I00008 sym/SYM_L00001 sym/SYM_LOCK sym/SYM_M00001 sym/SYM_M00002 sym/SYM_N00001 sym/SYM_N00002 sym/SYM_N00003 sym/SYM_N00004 sym/SYM_N00005 sym/SYM_N00006 sym/SYM_N00007 sym/SYM_N00008 sym/SYM_N00009 sym/SYM_N00010 sym/SYM_N00011 sym/SYM_N00012 sym/SYM_NODE sym/SYM_O00001 sym/SYM_P00001 sym/SYM_R00001 sym/SYM_R00002 sym/SYM_ROUTER sym/SYM_S00001 sym/SYM_T00001 sym/SYM_T00002 sym/SYM_T00003 sym/SYM_T00004 sym/SYM_T00005 sym/SYM_T00006 sym/SYM_T00007) JRN(sym/symjrn)
Table 89. Supported Data Types
Data Type Supported?

Char, VarChar, Long VarChar

Yes

Graphic, VarGraphic, Long VarGraphic

Yes

SmallInt, Integer, BigInt

Yes

Double

Yes

Decimal

Yes

Date, Time, TimeStamp

Yes

Blob, Clob, DBClob

Yes

DecFloat

No

Binary, VarBinary

No

C.5.3. Capturing Old Data

In order for SymmetricDS to capture old data for updates, the IMAGES(*BOTH) journaling attribute must be specified. One way to specify this attribute is by running the following command:

CHGJRNOBJ OBJ((sym/SYM_C00001 *FILE)) ATR(*IMAGES) IMAGES(*BOTH)

C.6. DB2 for IBM z/OS

Use symadmin module install db2 to install driver files, or copy your own files into the lib sub-directory.

To access DB2 z/OS, you will need the respective DB2 Connect license activated on the mainframe.

C.7. Derby

Use symadmin module install derby to install driver files, or copy your own files into the lib sub-directory.

The Apache Derby database can be run as an embedded database that is accessed by an application or a standalone server that can be accessed from the network. This dialect implementation creates database triggers that make method calls into Java classes. This means that the supporting JAR files need to be in the classpath when running Derby as a standalone database, which includes symmetric-ds.jar and commons-lang.jar.

C.8. Elasticsearch

Use symadmin module install elasticsearch to install driver files, or copy your own files into the lib sub-directory.

Send changes from your relational database to Elasticsearch in a variety of formats. An Elasticsearch node can be setup as a Load Only Node to receive changes from another node that is capturing changes.

Setup the Elasticsearch node by using the Connect Database wizard and selecting Elasticsearch as the type. The URL will be the connection point to Elasticsearch. If your Elasticsearch database has security enabled, please enter your username and password. When selecting an Elasticsearch User, make sure that the minimum permissions given to the user are all of the Index level permissions. This can be done by creating a custom role with those permissions, and assigning that role to the user.

elasticsearch node setup

C.8.1. Bulk Loading

Setup reload channels for bulk loading.

Update any reload channels that will be used on the table triggers that will capture changes and send them to Elasticsearch by setting the column data_loader_type to 'bulk'. It is also recommended to increase the batch size so that larger CSV files will be processed instead of the default size on reloads of 10,000 rows.

C.9. Firebird

Use symadmin module install firebird to install driver files, or copy your own files into the lib sub-directory.

The Firebird Dialect may require the installation of a User Defined Function (UDF) library in order to provide functionality needed by the database triggers. SymmetricDS includes the required UDF library, called SYM_UDF, in both source form (as a C program) and as pre-compiled libraries for both Windows and Linux. For Firebird 2.0 and earlier, the UDF is needed for capturing character and BLOB types, so the dialect will not allow startup if the UDF is missing. For Firebird 2.1 and later, the UDF is only needed for capturing BLOB types, so installation may not be necessary and the dialect does not check for it.

Download the SymmetricDS UDF Library if it is required for your use case. The SYM_UDF library is copied into the UDF folder within the Firebird installation directory.

For Linux users:

cp firebird/sym_udf.so /opt/firebird/UDF

For Windows users:

copy firebird\sym_udf.dll C:\Program Files\Firebird\Firebird_X_Y\UDF

The following limitations currently exist for this dialect:

  1. The outgoing batch does not honor the channel size, and all outstanding data events are included in a batch.

  2. Syncing of Binary Large Object (BLOB) is limited to 16KB per column.

  3. Syncing of character data is limited to 32KB per column. The overall row size of a resultset cannot exceed 64KB. For change capture, the row_data and old_data are limited to 10KB and and the pk_data is limited to 500 bytes for performance reasons. If you get the error of "arithmetic exception, numeric overflow, or string truncation" during extraction of a batch, set the contains_big_lob to true for the channel.

Firebird 3 is supported, however legacy authentication must be enabled in order to connect. Please refer to the Firebird 3.0 documentation for instructions on enabling this feature.

Table 90. Supported Data Types
Data Type Supported?

SmallInt

Yes

Integer

Yes

BigInt

Yes

Char

Yes

VarChar

Yes

Float

Yes

Decimal

Yes

Numeric

Yes

Double Precision

Yes

Date

Yes

Time

Yes

TimeStamp

Yes

Blob

No

C.9.1. Bulk Loading

SymmetricDS has bulk loading capability available for Firebird, when registering a Firebird database with SymmetricDS, bulk loading will be checked by default. SymmetricDS specifies data loader types on a channel by channel basis. Once you have your database registered, to utilize bulk loading versus straight JDBC insert, specify the Bulk Loader ("bulk") in the data_loader_type column of sym_channel table.

C.10. Greenplum

Greenplum is a data warehouse based on PostgreSQL. It is supported as a target platform in SymmetricDS.

SymmetricDS has bulk loading capability available for Greenplum. SymmetricDS specifies data loader types on a channel by channel basis. To utilize Greenplum Bulk loading versus straight JDBC insert, specify the Postgres Bulk Loader ("postgres_bulk") in the data_loader_type column of sym_channel.

C.11. H2

The H2 database allows only Java-based triggers. Therefore the H2 dialect requires that the SymmetricDS jar file be in the database’s classpath.

Table 91. Supported Data Types
Data Type Supported?

Int, TinyInt, SmallInt, BigInt

Yes

Boolean

Yes

Decimal

Yes

Double, Real

Yes

Time, Date, Timestamp

Yes

Binary, Blob

Yes

C.12. HBase

Use symadmin module install hbase to install driver files, or copy your own files into the lib sub-directory.

C.12.1. Empty HBase

If you are setting up replication to HBase and the tables are not already present in Hbase SymmetricDS can create them through the phoenix JDBC driver. This driver maintains some additional meta data about the tables so that they can be accessed using SQL through the JDBC driver.

This configuration is setup as a Load Only Node in SymmetricDS. It does require the phoenix jdbc driver though to utilize it. This driver should be downloaded and placed in the /lib folder of SymmetricDS and restarted.

C.12.2. Existing HBase

If you are setting up replication to an HBase database that already has tables present you will need to follow the steps below.

Setup a new H2 node that will contain all the SymmetricDS runtime tables. To do this go through the Add Node setup and select type H2 and provide a name for the database (it will create a new one locally if not present). This will allow SymmetricDS to create tables such as incoming_batch etc to maintain the replication.

Next you will need to setup a channel (or use the default channel) and set the data_loader_type to hbase.

Finally setup a parameter that contains the path of your hbase-site.xml file.

hbase.site.xml.path

All changes captured will now use the HBase data loader to load into an existing HBase table.

C.13. HSQLDB

Use symadmin module install hsqldb to install driver files, or copy your own files into the lib sub-directory.

HSQLDB was implemented with the intention that the database be run embedded in the same JVM process as SymmetricDS. Instead of dynamically generating static SQL-based triggers like the other databases, HSQLDB triggers are Java classes that re-use existing SymmetricDS services to read the configuration and insert data events accordingly.

The transaction identifier support is based on SQL events that happen in a 'window' of time. The trigger(s) track when the last trigger fired. If a trigger fired within X milliseconds of the previous firing, then the current event gets the same transaction identifier as the last. If the time window has passed, then a new transaction identifier is generated.

C.14. Apache Ignite

Use symadmin module install ignite to install driver files, or copy your own files into the lib sub-directory.

Since SymmetricDS is trigger based and there are not triggers in Apache Ignite, data can only be loaded to an Apache Ignite instance. The runtime SymmetricDS tables will also need to be installed in a full relational database to support integration with Apache Ignite.

See Load Only Node for more details on the general concepts of a load only node.

The following steps explain how to configure a SymmetricDS instance using Apache Ignite as a destination node:

  • Configure and start an Apache Ignite cluster.

  • Copy the Apache Ignite JDBC driver (ignite-core-VERSION.jar) to the "lib" directory of the SymmetricDS installation.

  • Start SymmetricDS and configure a master node with the desired source database.

  • Configure the desired node groups, group links, and routers.

  • Create a target node and database that will contain the SymmetricDS runtime tables for the Apache Ignite instance.

The simplest solution to support Ignite is to add a new node (see Add Node) that is connected to an H2 database to store all the SYM_* runtime tables.
  • Stop your SymmetricDS instance and edit the .properties file for the target node in the engines directory of the SymmetricDS installation.

  • Set the following properties in the engine file:

load.only=true
target.db.url=jdbc:ignite:thin://localhost
target.db.driver=org.apache.ignite.IgniteJdbcThinDriver
target.db.user=
target.db.password=
target.create.table.not.null.columns.supported=false
  • Update the jdbc url, username, and password to the desired Apache Ignite instance.

  • Restart SymmetricDS.

  • Create Table Triggers and Table Routers for the desired source tables to sync.

Keep in mind that SymmetricDS currently only supports syncing to the "PUBLIC" schema of an Apache Ignite instance.
  • (Optional) Perform an initial load from the source to the target node and/or send the table definitions to the Apache Ignite instance.

C.15. Informix

Use symadmin module install informix to install driver files, or copy your own files into the lib sub-directory.

The Informix Dialect was tested against Informix Dynamic Server 11.50, but older versions may also work. You need to download the Informix JDBC Driver (from the IBM Download Site) and put the ifxjdbc.jar and ifxlang.jar files in the SymmetricDS lib folder.

Make sure your database has logging enabled, which enables transaction support. Enable logging when creating the database, like this:

CREATE DATABASE MYDB WITH LOG;

Or enable logging on an existing database, like this:

ondblog mydb unbuf log
ontape -s -L 0

Make sure your user has permission to connect and create objects in the database:

GRANT CONNECT TO SYMMETRIC;
GRANT RESOURCE TO SYMMETRIC;

The following features are not yet implemented:

  1. Syncing of Binary and Character Large Objects (LOB) is disabled.

  2. There is no transaction ID recorded on data captured, so it is possible for data to be committed within different transactions on the target database. If transaction synchronization is required, either specify a custom transaction ID or configure the synchronization so data is always sent in a single batch. A custom transaction ID can be specified with the tx_id_expression on TRIGGER. The batch size is controlled with the max_batch_size on CHANNEL. The pull and push jobs have runtime properties to control their interval.

C.16. Ingres

This section describes Ingres specific SymmetricDS details.

C.16.1. Ingres Driver

The Ingres Driver is a licensed library, so the download of the driver can not be performed by SymmetricDS. You will have to provide the driver and install it in the lib sub-directory of the SymmetricDS installation.

C.16.2. Supported Data Types

Table 92. Supported Data Types
Data Type Supported?

Char

Yes

NChar

Yes

VarChar

Yes

NVarChar

Yes

Long VarChar

Yes

Long NVarChar

Yes

C

No

Text

No

TinyInt

Yes

SmallInt

Yes

Integer

Yes

BigInt

Yes

Float

Yes

Float4

No

AnsiDate

No

Time Without Time Zone

Yes

Time With Time Zone

Yes

Time With Local Time Zone

Yes

Timestamp Without Time Zone

Yes

Timestamp With Time Zone

Yes

Timestamp With Local Time Zone

Yes

Interval Year To Month

Yes

Interval Day To Second

Yes

IngresDate

No

Money

No

Logical_Key

No

IPV4

No

IPV6

No

UUID

Yes

Byte

Yes

VarByte

Yes

Long Byte

Yes

Boolean

Yes

Spatial

No

C.17. Interbase

Use symadmin module install interbase to install driver files, or copy your own files into the lib sub-directory.

The Interbase Dialect requires the installation of a User Defined Function (UDF) library in order to provide functionality needed by the database triggers. SymmetricDS includes the required UDF library, called SYM_UDF, in both source form (as a C program) and as pre-compiled libraries for both Windows and Linux.

Download the SymmetricDS UDF Library. The SYM_UDF library is copied into the UDF folder within the Interbase installation directory.

For Linux users:

cp interbase/sym_udf.so /opt/interbase/UDF

For Windows users:

copy interbase\sym_udf.dll C:\CodeGear\InterBase\UDF

The Interbase dialect currently has the following limitations:

  1. Data capture is limited to 4 KB per row, including large objects (LOB).

  2. There is no transaction ID recorded on data captured. Either specify a tx_id_expression on the TRIGGER table, or set a max_batch_size on the CHANNEL table that will accommodate your transactional data.

C.18. Kafka

Use symadmin module install kafka to install driver files, or copy your own files into the lib sub-directory.

Send changes from your relational database to Kafka in a variety of formats. A Kafka node can be setup as a Load Only Node to receive changes from another node that is capturing changes.

Setup the Kafka node by using the Connect Database wizard and selecting Kafka as the type. The URL will be the connection point to Kafka. User and password are not needed (or used).

kafka node setup

After hitting next you can setup advanced options for your Kafka node.

kafka advanced settings

C.18.1. Output Message Format

Set the output message format with the following property
kafka.format=JSON|XML|AVRO|CSV
JSON

Json formatted output message

{
  "table name": {
    "eventType": "INSERT|UPDATE|DELETE",
    "data": {
      "column name": "value",....
    }
  }
}
XML

Xml formatted output message

<row entity="table name" dml="INSERT|UPDATE|DELETE">
        <data key="column name">value</data>
        ...
</row>
AVRO

Apache Avro output message (Avro Schema Below)

{
  "type": "record",
  "name": "cdc",
  "fields": [
    {
      "name": "table",
      "type": "string"
    },
    {
      "name": "eventType",
      "type": "string"
    },
    {
      "name": "data",
      "type": {
        "type": "array",
        "items": {
          "name": "column",
          "type": "record",
          "fields": [
            {
              "name": "name",
              "type": "string"
            },
            {
              "name": "value",
              "type": [
                "null",
                "string"
              ]
            }
          ]
        }
      }
    }
  ]
}
CSV

CSV formatted output message

TABLE,table name,EVENT,INSERT|UPDATE|DELETE,column name,value, ...

C.18.2. Setting the Topic

Set the topic using the following property
kafka.topic.by=CHANNEL|TABLE
CHANNEL

This will send to a topic based on the channel of the batch that is being sent.

TABLE

This will send to a topic based on the table name of the change.

C.18.3. Setting Messages By

Set following property to determine how messages will be sent.
kafka.message.by=BATCH|ROW
BATCH

This will send one message for each batch containing all changes.

ROW

This will send one messsage for each change captured.

C.18.4. Setting The Producer

Set following property to specify the producer of the messages.
kafka.producer=myapplication

Provide a value for the producer of the message.

C.18.5. Using a Confluent AVRO Schema Registry

The message format must be AVRO for this option to work.
Set following property to a Confluent registry.
kafka.confluent.registry.url
Set following property to the base java package that contains the java POJOs that implement the org.apache.avro.generic.IndexedRecord.
kafka.avro.java.package
The jar file containing the AVRO generated POJO java beans must be placed in the /lib or /patches folder of SymmetricDS and then a restart is required.

C.18.6. Using Authentication

Any engine property prefixed with kafkaclient will be passed through to the Kafka client producer. Here is an example of authentication over SSL.

kafkaclient.security.protocol=SASL_SSL
kafkaclient.sasl.mechanism=PLAIN
kafkaclient.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username="myuser" password="mypassword";

C.19. MariaDB

Use symadmin module install mariadb to install driver files, or copy your own files into the lib sub-directory.

See MySQL notes. You can use either the MySQL or MariaDB driver for this dialect.

Table 93. Supported Data Types
Data Type Supported?

TinyInt, SmallInt, MediumInt, Int, BigInt

Yes

Decimal, Numeric

Yes

Float, Double

Yes

Bit

Yes

Char, Varchar

Yes

Binary, VarBinary

Yes

TinyBlob

No

Blob, MediumBlob, Longblob

Yes

TinyText, Text, MediumText, LongText

Yes

Enum

No

Set

No

Date, Time, DateTime, TimeStamp, Year

Yes

Point, LineString, Polygon, MultiPoint, MultiLinestring, MultiPolygon, GeometryCollection, Geometry

No

C.19.1. Bulk Loading

When registering a Maria database with SymmetricDS, bulk loading will be checked by default.

mysql enable bulk loading

To succesfully bulk load, once your MySQL database is registered with SymmetricDS, go to the top of the screen and select the MySQL node you want to bulk load from the drop down list of your nodes.

Then, go to the Manage tab, and go to Startup Parameters and set mysql.bulk.load.local to true, and restart SymmetricDS.

mysql bulk load local true

After completing this step, your Maria database should be configured to bulk load.

C.20. MongoDB

Use symadmin module install mongodb to install driver files, or copy your own files into the lib sub-directory.

MongoDB can be setup as a load only or a log based node.

Load Only - Select this mode at setup if you only need to load data into Mongo (no capture). Log Based - Select this mode if you need to capture changes in Mongo to be sent out to other nodes.

C.20.1. Setup

Obtain the connection url from MongoDB. On their dashboard select "Connect".

mongodb setup 1

Select the "Connect your application"

mongodb setup 2

Choose the Java driver to produce the proper connection string for driver used by SymmetricDS

mongodb setup 3

Add a new node to SymmetricDS

mongodb setup 4

On the advanced options you can setup bulk loading and a database to use. If no database is used it will create and use one named "symmetricds".

mongodb setup 5

C.20.2. Supported Operations for CDC

SymmetricDS uses the change streams feature of MongoDB to monitor ("watch") for changes that occur in MongoDB. The change stream API though does not support old data.

Updates - In the case of updates only the final updated values are provided to the change stream. This is sufficient for most replications as long as conflict detection is not needed.

Deletes - When a delete occurs the only value provided through the change stream is the _id that was deleted. This represents MongoDBs identifier for this row (similar to a rowid on other platforms).

In order to fully support deletes all replicated tables must contain an _id column and the following parameter must be turned on.

mongodb.use.mongo.ids=true

C.21. MySQL

MySQL supports several storage engines for different table types. However, SymmetricDS requires a storage engine that handles transactions. The recommended storage engine is InnoDB, which is included by default in MySQL 5.0 distributions. Either select the InnoDB engine during installation or modify your server configuration. To make InnoDB the default storage engine, modify your MySQL server configuration file (my.ini on Windows, my.cnf on Unix):

default-storage_engine = innodb

Alternatively, you can convert tables to the InnoDB storage engine with the following command:

alter table t engine = innodb;

On MySQL 5.0, the SymmetricDS user needs the SUPER privilege in order to create triggers.

grant super on *.* to symmetric;

On MySQL 5.1, the SymmetricDS user needs the TRIGGER, PROCESS, and CREATE ROUTINE privileges in order to create triggers and functions.

grant trigger on *.* to symmetric;
grant create routine on *.* to symmetric;
grant alter routine on *.* to symmetric;
grant process on *.* to symmetric;

Starting in MySQL 5.7.6, the "PROCESS" privilege is also required for the MySQL user that is modifying the application tables. This is required to look up the transaction id. Internally, the trigger will submit this query during an insert/update/delete:

select TRX_ID from INFORMATION_SCHEMA.INNODB_TRX where TRX_MYSQL_THREAD_ID = CONNECTION_ID();

grant process on *.* to db_user;

MySQL allows '0000-00-00 00:00:00' to be entered as a value for datetime and timestamp columns. JDBC cannot deal with a date value with a year of 0. In order to work around this SymmetricDS can be configured to treat date and time columns as varchar columns for data capture and data load. To enable this feature set the db.treat.date.time.as.varchar.enabled property to true.

If you are using UTF-8 encoding in the database, you might consider using the characterEncoding parameter in the JDBC URL.

jdbc:mysql://hostname/databasename?tinyInt1isBit=false&characterEncoding=utf8

If you are using Amazon Web Services to run the MySQL instance, Amazon by default uses binary log replication and disables the creation of triggers. In order to allow SymmetricDS to function correctly, the parameter 'log_bin_trust_function_creators' needs to be set to a value of '1'. You can set this value by logging into the database as root and enter the following command:

mysql -u USERNAME -p
set global log_bin_trust_function_creators=1;

If you can not log into the server as root, then you can set the parameter in a new parameter group on the Relational Database Services (RDS) web console. See the documentation from RDS for details.

If you are using log-based capture, then you must start MySQL using the --log-bin option. In addition, you must set the following system variables:

binlog_format=ROW
enforce_gtid_consistency=ON
gtid_mode=ON

Note that there are four possible values for gtid_mode and it must be set one step at a time. If gtid_mode is set to "OFF", then it must be set to "OFF_PERMISSIVE" and then "ON_PERMISSIVE" before it can be set to "ON". Also note that in versions earlier than MySQL 5.7.6, the gtid_mode variable must be set by starting the server with the --gtid-mode=ON option. Prior to MySQL 5.7.5, the --log-slave-updates option must also be used at startup.

Table 94. Supported Data Types
Data Type Supported?

TinyInt, SmallInt, Int, MediumInt, BigInt

Yes

Decimal, Numeric

Yes

Float, Double

Yes

Bit

Yes

Date, DateTime, TimeStamp, Time, Year

Yes

Char, Varchar

Yes

Binary, VarBinary

Yes

TinyBlob, Blob, MediumBlob, BigBlob

Yes

TinyText, Text, MediumText, BigText

Yes

Enum

No

Set

No

Geometry, Point, LineString, Polygon, GeometryCollection, MultiPoint, MultiLinestring, MultiPolygon

No

C.21.1. Bulk Loading

When registering a MySQL database with SymmetricDS, bulk loading will be checked by default.

mysql enable bulk loading

To succesfully bulk load, once your MySQL database is registered with SymmetricDS, go to the top of the screen and select the MySQL node you want to bulk load from the drop down list of your nodes.

mysql selecting node 1
mysql selecting node 2

Then, go to the Manage tab, and go to Startup Parameters and set mysql.bulk.load.local to true, and restart SymmetricDS.

mysql bulk load local true

In order to bulk load, your MySQL database must have the local_infile variable set to ON. In order to set this, you can run the query below to set the variable and have it remain on, even upon the restarting of your database.

SET PERSIST local_infile=1;

After completing these steps, your MySQL database should be configured to bulk load.

C.22. Progress OpenEdge

The Progress OpenEdge JDBC driver is not publicly available, so it is not possible to install the driver files via the symadmin command. Instead, to install the driver, copy your openedge.jar file into the lib sub-directory.

Make sure your user has the DBA privilege, which allows the user to create, access, modify and delete database objects:

GRANT DBA TO SYMMETRIC;

C.23. OpenSearch

Use symadmin module install opensearch to install driver files, or copy your own files into the lib sub-directory.

Send changes from your relational database to OpenSearch in a variety of formats. An OpenSearch node can be setup as a Load Only Node to receive changes from another node that is capturing changes.

Setup the OpenSearch node by using the Connect Database wizard and selecting OpenSearch as the type. The URL will be the connection point to OpenSearch. User and password are not needed (or used).

opensearch node setup

After hitting next you can setup advanced options for your OpenSearch node.

opensearch advanced settings

C.23.1. Bulk Loading

Setup reload channels for bulk loading.

Update any reload channels that will be used on the table triggers that will capture changes and send them to OpenSearch by setting the column data_loader_type to 'bulk'. It is also recommended to increase the batch size so that larger CSV files will be processed instead of the default size on reloads of 10,000 rows.

C.24. Oracle

This section describes Oracle specific SymmetricDS details.

C.24.1. Permissions

The SymmetricDS database user generally needs privileges for connecting and creating tables (including indexes), triggers, sequences, and procedures (including packages and functions). For change data capture, use the following permissions:

GRANT CONNECT TO SYMMETRIC;
GRANT RESOURCE TO SYMMETRIC;
GRANT CREATE ANY TRIGGER TO SYMMETRIC;
GRANT EXECUTE ON UTL_RAW TO SYMMETRIC;
Log Miner Permissions

When using Log Miner instead of triggers, the SYSDBA should put the database into archive log mode:

shutdown immediate;
startup mount;
alter database archivelog;
alter database open;
alter database add supplemental log data (all) columns;

If using Oracle 11g without CDB, use the following permissions:

grant create session, alter session, select any dictionary, select any transaction, execute_catalog_role to symmetric;

If using Oracle 12c without CDB, use the following permissions:

grant create session, alter session, select any dictionary, execute_catalog_role, logmining to symmetric;

If using Oracle 12c with PDB, use the following permissions:

grant create session, alter session, set container, select any dictionary, logmining, execute_catalog_role to c##symmetric CONTAINER=all;
alter user c##symmetric set container_data = (cdb$root, <PDB name>) container=current;

C.24.2. Known Limitations

  • The LONG data type is not supported with change capture, since LONG columns cannot be accessed from triggers. The data in LONG columns can still be kept in sync by enabling the use_stream_lobs feature in TRIGGER.

  • The global precision of numeric columns is controlled by the oracle.template.precision parameter. It defaults to a precision of 30,10

  • With the default settings a database row cannot exceed 4k. If the error 'ORA-01489: result of string concatenation is too long' occurs then set use_capture_lobs to 1 in the TRIGGER table and contains_big_lobs to 1 on the assigned CHANNEL. Triggers will need to be synchronized. By enabling use_capture_lobs, the concatenated varchar string is cast to a clob which allows a length of more than 4k. By enabling contains_big_lobs, the extraction of sym_data is cast to a clob which prevents truncation at 4k. There is overhead for both of these settings

  • When multiple triggers are defined on the same table, then the order in which the triggers occur appears to be arbitrary

C.24.3. Bulk Loading

SymmetricDS has bulk loading capability available for Oracle. SymmetricDS specifies data loader types on a channel by channel basis.

To utilize bulk loading with Oracle, you should create a user that is not the default system user of the database. Whatever user you end up using, make sure it also has the permissions mentioned above in the Database Permission section, as well as the ones mentioned below. It’s important to note that SQL Loader (what Oracle uses to bulk load) will not work if either your user name or password contains a '@' symbol, so make sure to avoid this in both your username and password.

create user yourusername identified by "yourpassword";

Once your user is created, make sure to add the following permissions to the user to make sure that Symmetric has the correct permissions to bulk load with your database. (Again, make sure your user has the permissions mentioned in the Database Permission section as well)

GRANT CREATE SESSION TO yourusername;
GRANT CREATE TABLE TO yourusername;
alter user yourusername quota unlimited on system;
GRANT CONNECT TO yourusername;
GRANT RESOURCE TO yourusername;
GRANT CREATE ANY TRIGGER TO yourusername;
GRANT EXECUTE ON UTL_RAW To yourusername;

Register your Oracle database with Symmetric using the user you’ve just created. Once you have done that, when you get to the Bulk Loading screen on the Connect Database Wizard you will need to provide a path to your SQL Loader installation. If you do not have SQL Loader installed, there is a link provided in the Connect Database Wizard where you can go to get this installed.

Once you have your database registered, to utilize Oracle Bulk loading versus straight JDBC insert, specify the Oracle Bulk Loader ("bulk") in the data_loader_type column of sym_channel table.

Once these steps are complete your database should be configured to bulk load with SymmetricDS.

C.24.4. Optional - Partitioning

Partitioning the DATA table by channel can help insert, routing and extraction performance on concurrent, high throughput systems. TRIGGERs should be organized to put data that is expected to be inserted concurrently on separate CHANNELs. The following is an example of partitioning. Note that both the table and the index should be partitioned. The default value allows for more channels to be added without having to modify the partitions.

CREATE TABLE SYM_DATA
(
    data_id INTEGER NOT NULL ,
    table_name VARCHAR2(50) NOT NULL,
    event_type CHAR(1) NOT NULL,
    row_data CLOB,
    pk_data CLOB,
    old_data CLOB,
    trigger_hist_id INTEGER NOT NULL,
    channel_id VARCHAR2(20),
    transaction_id VARCHAR2(1000),
    source_node_id VARCHAR2(50),
    external_data VARCHAR2(50),
    create_time TIMESTAMP
) PARTITION BY LIST (channel_id) (
PARTITION P_CONFIG VALUES ('config'),
PARTITION P_CHANNEL_ONE VALUES ('channel_one'),
PARTITION P_CHANNEL_TWO VALUES ('channel_two'),
...
PARTITION P_CHANNEL_N VALUES ('channel_n'),
PARTITION P_DEFAULT VALUES (DEFAULT));
CREATE UNIQUE INDEX IDX_D_CHANNEL_ID ON SYM_DATA (DATA_ID, CHANNEL_ID)  LOCAL
(
 PARTITION I_CONFIG,
 PARTITION I_CHANNEL_ONE,
 PARTITION I_CHANNEL_TWO,
 ...
 PARTITION I_CHANNEL_N,
 PARTITION I_DEFAULT
);

C.24.5. Supported Data Types

Table 95. Supported Data Types
Data Type Supported?

Char

Yes

NChar

Yes

VarChar2

Yes

NVarChar2

Yes

Long

No

Number

Yes

Binary_Float

Yes

Binary_Double

Yes

Date

Yes

Timestamp

Yes

Timestamp With Time Zone

Yes

Timestamp With Local Time Zone

Yes

Interval Year to Month

Yes

Interval Day to Second

Yes

Raw

Yes

Long Raw

No

RowID

Yes

URowID

No

Clob

Yes

NClob

Yes

Blob

Yes

BFile

No

C.25. PostgreSQL

Starting with PostgreSQL 8.3, SymmetricDS supports the transaction identifier. Binary Large Object (BLOB) replication is supported for both byte array (BYTEA) and object ID (OID) data types.

In order to function properly, SymmetricDS needs to use session variables. Before PostgreSQL 9.2, session variables are enabled using a custom variable class. Add the following line to the postgresql.conf file of PostgreSQL server on versions before 9.2:

custom_variable_classes = 'symmetric'

This setting is required on versions before 9.2, and SymmetricDS will log an error and exit if it cannot set session variables. PostgreSQL versions 9.2 or later do not require this setting.

Before database triggers can be created by in PostgreSQL, the plpgsql language handler must be installed on the database. If plpgsql is not already installed, the following statements can be run by the administrator on the database:

CREATE FUNCTION plpgsql_call_handler() RETURNS language_handler AS
    '$libdir/plpgsql' LANGUAGE C;

CREATE FUNCTION plpgsql_validator(oid) RETURNS void AS
    '$libdir/plpgsql' LANGUAGE C;

CREATE TRUSTED PROCEDURAL LANGUAGE plpgsql
    HANDLER plpgsql_call_handler
    VALIDATOR plpgsql_validator;

If you want SymmetricDS to install into a schema other than public you can alter the database user to set the default schema.

alter user {user name} set search_path to {schema name};

You will also need to set privileges on the schema so that the SymmetricDS user can use and write to the schema.

grant usage on schema {schema name} to {user name};
grant create on schema {schema name} to {user name};

C.25.1. Permissions

If SymmetricDS is connected into a different schema than the application tables and a different user account is used to write to the the application tables the following permissions need to be granted to the application user to access the SymmetricDS schema

GRANT SELECT, INSERT, UPDATE, DELETE ON ALL TABLES IN SCHEMA {SYMMETRIC_SCHEMA} TO  {APPLICATION USER}
GRANT ALL ON SCHEMA {SYMMETRIC_SCHEMA} TO {APPLICATION USER}
GRANT ALL ON ALL SEQUENCES IN SCHEMA {SYMMETRIC_SCHEMA} TO  {APPLICATION USER}
GRANT ALL ON ALL FUNCTIONS IN SCHEMA {SYMMETRIC_SCHEMA} TO  {APPLICATION USER}

C.25.2. Log Miner

When using Log Miner instead of triggers, it will use the logical decoding feature introduced in PostgreSQL version 9.4. The WAL level should be set to "logical" on the Postgres server. Edit postgresql.conf and restart Postgres after changing:

wal_level = logical
max_replication_slots = 10

If the SymmetricDS user is not a superuser, the following role attribute needs to be added:

ALTER USER {SYMMETRICDS USER} REPLICATION

Replication of updates and deletes work as expected for tables with a primary key. For tables without a primary key, the user needs to set the REPLICA IDENTITY on the table to either USING INDEX to record columns from the named index or FULL to record all columns of the row. In Postgres 10 and newer, it will prevent updates and deletes with an error about setting REPLICA IDENTITY.

alter table mytable replica identity full;
Table 96. supported data types
Data Type Supported?

SmallInt, Integer, BigInt

Yes

Decimal, Numeric

Yes

Real, Double Precesion

Yes

Serial, BigSerial

Yes

Char, Varchar, Text

Yes

Money

No

Timestamp, Date, Time, Interval

Yes

Enum

No

Point, Lseg, Box, Path, Polygon, Circle

Yes

C.25.3. Bulk Loading

SymmetricDS has bulk loading capability available for Postgres, when registering a PostgreSQL database with SymmetricDS, bulk loading will be checked by default. SymmetricDS specifies data loader types on a channel by channel basis. To utilize Postgres Bulk loading versus straight JDBC insert, specify the Postgres Bulk Loader ("postgres_bulk") in the data_loader_type column of sym_channel.

C.26. RabbitMQ

Send changes from your relational database to RabbitMQ in a variety of formats. A RabbitMQ node can be setup as a Load Only Node to receive changes from another node that is capturing changes.

Setup the RabbitMQ node by using the Connect Database wizard and selecting RabbitMQ as the type. The URL will be the connection point to RabbitMQ.

rabbitmq node setup

After hitting next you can setup advanced options for your RabbitMQ node.

rabbitmq advanced settings

C.26.1. Output Message Format

Set the output message format with the following property
rabbitmq.format=JSON|XML|CSV
JSON

Json formatted output message

{
  "table name": {
    "eventType": "INSERT|UPDATE|DELETE",
    "data": {
      "column name": "value",....
    }
  }
}
XML

Xml formatted output message

<row entity="table name" dml="INSERT|UPDATE|DELETE">
        <data key="column name">value</data>
        ...
</row>
CSV

CSV formatted output message

TABLE,table name,EVENT,INSERT|UPDATE|DELETE,column name,value, ...

C.26.2. Setting Messages By

Set following property to determine how messages will be sent.
rabbitmq.message.by=BATCH|ROW
BATCH

This will send one message for each batch containing all changes.

ROW

This will send one messsage for each change captured.

C.26.3. Setting Queue By

Set following property to determine what queues will be used to load data to.
rabbitmq.queue.by=TABLE|CHANNEL|USER
TABLE

A queue with the name of the table will be created for each table in the batch. For example if you are sending information from a table called "my_table", a queue called "my_table" will be created and the messages will be sent there.

CHANNEL

A queue with the name of the channel will be created for the name of the channel of the batch. For example, if the batch is sent on the "reload" channel, a queue called "sym_reload" will be created and the messages will be sent there.

USER

A queue with the name of the user will be created to load data to.

C.26.4. Queue Name

Enter the queue name you wish to configure with SymmetricDS. If you enter a value for this field, it will override the "Queue By" selection. If you leave this field blank, queue(s) will be created based on the "Queue By" field.

C.27. Redshift

Use symadmin module install redshift to install driver files, or copy your own files into the lib sub-directory.

Redshift is a managed data warehouse in the cloud from Amazon. Version 1.0 of Redshift is based on PostgreSQL 8.0, with some features modified or removed. SymmetricDS supports Redshift as a target platform where data can be loaded, but it does not support data capture. However, the initial load and reload functions are implemented, so it is possible to query rows from Redshift tables and send them to another database.

While Redshift started with PostgreSQL 8.0, there are some important differences from PostgreSQL. Redshift does not support constraints, indexes, functions, triggers, or sequences. Primary keys, foreign keys, and unique indexes can be defined on tables, but they are informational metadata that are not enforced by the system. When using the default data loader with SymmetricDS, it will enforce primary keys, either defined in the database or with the sync keys features, by checking if a row exists before attempting an insert. However, the bulk loader does not perform this check. The data types supported are smallint, integer, bigint, decimal, real, double precision, boolean, char, varchar, date, and timestamp.

A data loader named "redshift_bulk" is a bulk loader that can be set for a channel to improve loading performance. Instead of sending individual SQL statements to the database, it creates a comma separated value (CSV) file, uploads the object to Amazon S3, and uses the COPY statement to load it. The COPY command appends the new data to any existing rows in the table. If the target table has any IDENTITY columns, the EXPLICIT_IDS option is enabled to override the auto-generated values and load the incoming values. The following parameters (see Appendix B) can be set for bulk loader:

redshift.bulk.load.max.rows.before.flush

When the max rows is reached, the flat file is sent to S3 and loaded into the database. The default is 100,000 rows.

redshift.bulk.load.max.bytes.before.flush

When the max bytes is reached, the flat file is sent to S3 and loaded into the database. The default is 1,000,000,000 bytes.

redshift.bulk.load.s3.bucket

The S3 bucket name where files are uploaded. This bucket should be created from the AWS console ahead of time.

redshift.bulk.load.s3.access.key

The AWS access key ID to use as credentials for uploading to S3 and loading from S3.

redshift.bulk.load.s3.secret.key

The AWS secret key to use as credentials for uploading to S3 and loading from S3.

redshift.bulk.load.s3.endpoint

The AWS endpoint used for uploading to S3. This is optional. You might need to specify if you get warnings about retrying during the S3 upload.

To clean and organize tables after bulk changes, it is recommended to run a "vacuum" against individual tables or the entire database so that consistent query performance is maintained. Deletes and updates mark rows for delete that are not automatically reclaimed. New rows are stored in a separate unsorted region, forcing queries to sort on demand. Consider running a "vacuum" periodically during a maintenance window when there is minimal query activity that will be affected. If large batches are continually loaded from SymmetricDS, the "vacuum" command can be run after committing a batch by using a load filter (see Section 3.9) for the "batch commit" event, like this:

for (String tablename : context.getParsedTables().keySet()) {
    engine.getSqlTemplate().update("vacuum " + tablename, new Object[] { } );
}

C.28. SingleStore

SingleStore is only available in the professional version of SymmetricDS. Supports bi-directional replication with SingleStore using log based data capture from the audit logs.

C.28.1. Requirements

User provided to connect SymmetricDS to the SingleStore database must be different than the user account used by applications writing to SingleStore. This is required for configurations where SingleStore is used as a source.

Several database variables need to to be setup to support audit logging when SingleStore is a source node. These variables can be set using the sdb-admin tool.

Setting variables
sdb-admin update-config --key auditlog_level --value ALL-QUERIES-PLAINTEXT --all
sdb-admin update-config --key auditlog_disk_sync --value ON --all
sdb-admin update-config --key auditlog_retention_period --value 7 --all

C.29. Snowflake

Use symadmin module install snowflake to install driver files, or copy your own files into the lib sub-directory.

C.29.1. Setup

Snowflake supports running as a load only node in SymmetricDS. See Load Only Node for details on setting up a load only node in SymmetricDS. Snowflake can also be set up to capture changes using log-mining.

Setup the Snowflake node by using the Connect Database wizard and selecting Snowflake as the type.

snowflake database settings

After hitting next you can setup advanced options for your Snowflake node.

C.29.2. Permissions

Log Miner Permissions

If the tables that are being tracked for change data capture using log mining are already created, each table needs to be altered to add change data tracking to it. An example DDL statement to turn this on is shown:

ALTER TABLE TABLE1 SET CHANGE_TRACKING=TRUE;

Permissions must also be added to the database user ID that is used to access the Snowflake database. Assuming that you wanted to use a role called SYMMETRIC and a user called SYMMETRIC, here are the set of permissions that must be set in order for SymmetricDS to successfully run the log mining (called Streams in Snowflake) functionality (replace <warehouse>, <database>, and <schema> with real values):

CREATE ROLE SYMMETRIC;
GRANT USAGE ON WAREHOUSE <warehouse> TO ROLE SYMMETRIC;
GRANT USAGE ON DATABASE <database> TO ROLE SYMMETRIC;
GRANT USAGE ON SCHEMA <database>.<schema> TO ROLE SYMMETRIC;
GRANT CREATE STREAM ON SCHEMA <database>.<schema> TO ROLE SYMMETRIC;
GRANT SELECT ON ALL TABLES IN DATABASE <database> TO ROLE SYMMETRIC;
GRANT CREATE TABLE ON SCHEMA <database>.<schema> TO ROLE SYMMETRIC;
GRANT CREATE FILE FORMAT ON SCHEMA SYMMETRIC_TEST.SYMMETRIC_TEST TO ROLE SYMMETRIC;
CREATE USER SYMMETRIC PASSWORD = 'symmetric' DEFAULT_ROLE = 'SYMMETRIC';
GRANT ROLE SYMMETRIC TO USER SYMMETRIC;

C.29.3. Bulk Loading

If you have SymmetricDS create the tables to synchronize into Snowflake, SymmetricDS creates the tables with a cluster column specification based on the primary key of the table. If you do not want SymmetricDS to create the clustered columns, set the following property in the symmetric-server.properties file in the conf directory and restart the instance:

snowflake.cluster.primary.keys=false

If you do not have SymmetricDS create the tables, and you want clustered columns to be specified, you will have to define the cluster specification yourself. Here is an example statement of a clustered column specification for a primary key column called ID on table TABLE1:

ALTER TABLE TABLE1 CLUSTER BY (ID);

When defining the triggers, make sure to specify the source catalog (which would be the Snowflake database value) and the source schema for each table.

Choose a bulk load storage option

SymmetricDS will create and send CSV files to the a desired storage location (see below) as part of the load. Once the CSV files have been uploaded to a selected storage area Snowflake’s COPY INTO command will be used to load the data into Snowflake. Once the COPY INTO has completed SymmetricDS will also remove the CSV file from the storage container.

There are currently 3 supported storage options to stage the CSV files prior to loading into Snowflake
  • Snowflake Managed (internal storage)

  • AWS: S3

  • Azure: Storage Account

    SNOWFLAKE MANAGED

    Use a Snowflake managed internal stage.

    snowflake advanced settings snowflake managed
    AWS S3

    Use an existing AWS S3 cloud storage.

    snowflake advanced settings aws s3
    AZURE Storage Account

    Use an existing Azure Storage Account.

    snowflake advanced settings azure

C.30. SQLite

Use symadmin module install sqlite to install driver files, or copy your own files into the lib sub-directory.

For SQLite, the implementation of sync-on-incoming back and the population of a source node if in the sym data rows relies on use of a context table (by default, called sym_context) to hold a boolean and node id in place of the more common methods of using temp tables (which are inaccessible from triggers) or functions (which are not available). The context table assumes there’s a single thread updating the database at any one time. If that is not the case in the future, the current implementation of sync on incoming batch will be unreliable.

Nodes using SQLite should have the jobs.synchronized.enable parameter set to true. This parameter causes the jobs and push/pull threads to all run in a synchronized fashion, which is needed in the case of SQLite.

The SQLite dialect has the following limitations:

  • There is no transaction ID recorded on data captured. Either specify a tx_id_expression on the TRIGGER table, or set a max_batch_size on the CHANNEL table that will accommodate your transactional data.

  • Due to the single threaded access to SQLite, the following parameter should be set to true: jobs.synchronized.enable. The sync.triggers.thread.count.per.server parameter should be set to 1.

Table 97. Suported Data Types
Data Type Supported?

Text

Yes

Numeric

Yes

Integer

Yes

Real

Yes

Blob

Yes

C.30.1. Loading Data

SymmetricDS has bulk loading capability available for SQLite, when registering a SQLite database with SymmetricDS, bulk loading will be checked by default. SymmetricDS specifies data loader types on a channel by channel basis. Once you have your database registered, to utilize bulk loading versus straight JDBC insert, specify the Bulk Loader ("bulk") in the data_loader_type column of sym_channel table.

C.31. SQL Server

Microsoft SQL Server was tested using the jTDS JDBC driver.

SQL Server allows the update of primary key fields via the SQL update statement. If your application allows updating of the primary key field(s) for a table, and you want those updates synchronized, you will need to set the "Handle Key Updates" field on the trigger record for that specific table. The default for Handle Key Updates is false.

SymmetricDS expects a row count to be returned for data manipulation statements, which is the default setting for most server. However, if the NOCOUNT option is ON for SQL-Server, SymmetricDS will not behave correctly. The NOCOUNT setting can be checked with "select case when (512 & @@OPTIONS) = 512 then 'on' else 'off' end". If you’re unable to change NOCOUNT for the server, the "db.init.sql" parameter can be set to "SET NOCOUNT OFF" in the engine properties file.

Connections are pooled and expected to be in the database context like a new connection, so avoid using the "USE database" Transact-SQL statement in extension code.

If SQL Server is configured with a default collation that does NOT support unicode then we have experienced bad performance for update and delete statements when a table has character based primary keys. This is because statements are prepared for a unicode type and as a result the indexes are not used. You can turn this functionality off in JTDS by appending the following to your db.url: ;sendStringParametersAsUnicode=false

C.31.1. Permissions

SymmetricDS can be configured to work with a variety SQL Server configurations. It is possible to install SymmetricDS in the same database and schema with the same user account your application runs. However you can also set it up to run in its own database with a designated user. Whichever configuration you choose below are the permissions required for SymmetricDS to run.

Symmetric User

Application User

Symmetric Database

CREATE TABLE, CREATE FUNCTION, REFERENCES

INSERT, EXECUTE

Symmetric Schema

ALTER, SELECT, INSERT, UPDATE, DELETE

Application Database

SELECT, ALTER (and INSERT, UPDATE, DELETE if the node is a target)

Example 25. Example Script for a designated SymmetricDS database and user account

Replace the following variables with your desired values.

SYM_USER

The SymmetricDS user

SYM_DATABASE

The database the SymmetricDS runtime tables will be installed in

APP_DATABASE

The application database where sync data resides

APP_USER

The application database user account that the application uses when making changes to the data.

-- SymmetricDS User

CREATE LOGIN SYM_USER
WITH PASSWORD = 'SYM_USER';
GO
use SYM_DATABASE;
GO
CREATE USER SYM_USER FOR LOGIN SYM_USER;
GO
GRANT CREATE TABLE ON DATABASE::SYM_DATABASE to SYM_USER;
GRANT CREATE FUNCTION ON DATABASE::SYM_DATABASE to SYM_USER;
GRANT REFERENCES ON DATABASE::SYM_DATABASE to SYM_USER;
GRANT ALTER, SELECT, INSERT, DELETE, UPDATE ON SCHEMA::dbo TO SYM_USER;
GO

use APP_DATABASE;
CREATE USER SYM_USER FOR LOGIN SYM_USER
GRANT SELECT, ALTER ON DATABASE::APP_DATABASE to SYM_USER;

-- Grant the below permissions to the SymmetricDS User if the node is a target

GRANT INSERT, DELETE, UPDATE ON DATABASE::APP_DATABASE to SYM_USER;

-- Application User

CREATE LOGIN APP_USER
WITH PASSWORD = 'APP_USER';
GO
use APP_DATABASE;
GO
CREATE USER APP_USER FOR LOGIN APP_USER
GO
GRANT SELECT, INSERT, DELETE, UPDATE ON SCHEMA::dbo TO APP_USER;
GO
use SYM_DATABASE;
CREATE USER APP_USER FOR LOGIN APP_USER
GRANT INSERT, EXECUTE ON DATABASE::SYM_DATABASE to APP_USER;
Log Miner Permissions

When using Log Miner instead of triggers, SymmetricDS will alter the database to enable change tracking, if it’s not already enabled, which requires ALTER ANY DATABASE permission.

alter database mydb set change_tracking = on (change_retention = 2 days, auto_cleanup = on);

Snapshot isolation must be enabled for the database to ensure all changes are included in replication.

alter database mydb set allow_snapshot_isolation on;

For each table configured for replication, SymmetricDS will alter the table to enable change tracking, if it’s not already enabled, which requires ALTER TABLE permission.

alter table mytable enable change_tracking;
Table 98. Supported Data Types
Data Type Supported?

BigInt, Int, SmallInt, TinyInt

Yes

Decimal, Numeric

Yes

Bit

Yes

Money, SmallMoney

Yes

Float, Real

Yes

Date, DateTime, Datetime2, SmallDatetime, Time

Yes

Datetimeoffset

Yes

Char, Varchar, Text, Nchar, Nvarchar, Ntext

Yes

Binary, Varbinary

Yes

Image

Yes

Spatial Data Types

No

C.31.2. Bulk Loading

When registering a SQL Server database with SymmetricDS, bulk loading will be be checked by default. In order to make use of bulk loading, the user must provide a path to a bcp executable (installation for this executable is linked in the Connect Database Wizard when asked about bulk loading).

C.32. Sybase ASE

Use symadmin module install sybase to install driver files, or copy your own files into the lib sub-directory.

Sybase Adaptive Server Enterprise (ASE) was tested using the jConnect 7 JDBC driver. The jConnect 7 JDBC driver should be placed in the "lib" folder.

driver class : com.sybase.jdbc4.jdbc.SybDriver

SymmetricDS requires the "select into" database option be turned on for Sybase ASE. Run the following command with the sa account on the master database while replacing the database value with your database name.

sp_dboption 'YOUR_DATABASE', 'select into', true

SymmetricDS requires that the meta data information be installed on each database that will be used in replication. Sybase provides these metadata store procedures in a script that is packaged with the installation.

Without the metadata stored procedures installed above the following error will be produced by SymmetricDS

Error while reading the database metadata: JZ0SJ: Metadata accessor information was not found on this database. Please install the required tables as mentioned in the jConnect documentation.
Set the classpath to contain the jconnect jar and the classes directory
export CLASSPATH=/opt/sybase/jConnect-7_0/classes/jconn4.jar:/opt/sybase/jConnect-7_0/classes
Install appropriate script from the jconnect driver folder in your Sybase installation under the /sp directory
java  IsqlApp -U sa -P {SA_PASSWORD} -S jdbc:sybase:Tds:{HOSTNAME}:{PORT}/{DATABASE} -I sql_server{SYBASE VERSION}.sql -c go

Columns of type DATETIME are accurate to 1/300th of a second, which means that the last digit of the milliseconds portion will end with 0, 3, or 6. An incoming DATETIME synced from another database will also have its millisconds rounded to one of these digits (0 and 1 become 0; 2, 3, and 4 become 3; 5, 6, 7, and 8 become 6; 9 becomes 10). If DATETIME is used as the primary key or as one of the columns to detect a conflict, then conflict resolution could fail unless the milliseconds are rounded in the same fashion on the source system.

On ASE, each new trigger in a table for the same operation (insert, update, or delete) overwrites the previous one. No warning message displays before the overwrite occurs. When SymmetricDS is installed and configured to synchronize a table, it will install triggers that could overwrite already existing triggers on the database. New triggers created after SymmetricDS is installed will overwrite the SymmetricDS triggers. Custom trigger text can be added to the SymmetricDS triggers by modifying CUSTOM_ON_INSERT_TEXT, CUSTOM_ON_UPDATE_TEXT, and CUSTOM_ON_DELETE_TEXT on the TRIGGER table.

SymmetricDS expects a row count to be returned for data manipulation statements, which is the default setting for most server. However, if the NOCOUNT option is ON or the "send doneinproc tokens" setting is 0 for Sybase, SymmetricDS will not behave correctly. The NOCOUNT setting can be checked with "select case when (512 & @@OPTIONS) = 512 then 'on' else 'off' end". The DONEINPROC setting can be checked with "sp_configure 'send doneinproc tokens'". The commands "sp_configure 'send doneinproc tokens', 1" and "SET NOCOUNT OFF" will enable row counts. If you’re unable to change NOCOUNT for the server, the "db.init.sql" parameter can be set to "SET NOCOUNT OFF" in the engine properties file.

Connections are pooled and expected to be in the database context like a new connection, so avoid using the "USE database" Transact-SQL statement in extension code.

Table 99. Supported Data Types
Data Type Supported?

BigInt, Int, SmallInt, TinyInt

Yes

Decimal, Numeric

Yes

Bit

Yes

Money, SmallMoney

No

Float, Real

Yes

NChar, NVarchar

No

UNIChar, UNIVarchar, UNIText

No

Date, DateTime, SmallDatetime, Time

Yes

Char, Varchar, Text

Yes

Binary, Varbinary

Yes

Image

Yes

Blob

Yes

C.32.1. Permissions

The SymmetricDS database user generally needs privileges for connecting and creating tables (including indexes), triggers, sequences, and procedures (including packages and functions). In Sybase ASE, only the System Administrator can create or drop triggers on tables it does not own. The table owner can create or drop triggers freely on their tables. For change data capture, it is recommended that the application user who controls Sybase and SymmetricDS be the same, so that there will be no permissions errors with dropping or creating triggers. This means that all of the Sybase tables pertaining to and being used by SymmetricDS are owned by the same user, and that is the user who is assigned to SymmetricDS. If this is not possible, the SA user is the only other option, as users cannot be granted the permission to create or delete triggers on tables they do not own. Any outside users that will need to Insert, Update, Delete or read from a Sybase table that they do not own will need the following permissions for any tables they need access to:

grant ALL on <catalog.schema.tablename> to <user_name>
go

ALL is referring to the combination of INSERT, UPDATE, DELETE, SELECT, ALTER and REFERENCES. These can be done individually or as the ALL command demonstrated above.

C.32.2. Bulk Loading

SymmetricDS has bulk loading capability available for Sybase ASE, when registering a Sybase ASE database with SymmetricDS, bulk loading will be checked by default. SymmetricDS specifies data loader types on a channel by channel basis. Once you have your database registered, to utilize bulk loading versus straight JDBC insert, specify the Bulk Loader ("bulk") in the data_loader_type column of sym_channel table.

C.33. SQL Anywhere

Use symadmin module install sybase to install driver files, or copy your own files into the lib sub-directory.

SQL Anywhere and Sybase Adaptive Server Anywhere (ASA) were tested using the jConnect JDBC driver. The jConnect JDBC driver should be placed in the "lib" folder.

C.34. Amazon S3

Use symadmin module install s3 to install driver files, or copy your own files into the lib sub-directory.

Send changes from your relational database to S3 as a CSV file. An S3 node can be setup as a Load Only Node to receive changes from another node that is capturing changes.

Setup the S3 node by using the Connect Database wizard and selecting Amazon S3 as the type. User and password are not needed (or used).

After hitting next you can setup the bucket in S3 that the files go to as well as the directory path desired within that bucket. This will also be where you specify the user credentials that have access to your Amazon S3 instance.

C.35. Teradata

Use symadmin module install teradata to install driver files, or copy your own files into the lib sub-directory.

Teradata is only supported as a Load Only Node. It can only be used as a target to load data into from other dialects.

C.35.1. Drivers

All jdbc driver jar files need to be placed in the SymmetricDS installation /lib directory or on the classpath of the app server for a deployment through a war file.

The following jars were required to load data into 15.10
  • terajdbc4.jar

  • tdgssconfig.jar

C.35.2. Permissions

The user provided at the setup of the node but have select access to the DBC database.

GRANT SELECT ON "DBC" TO {USER} WITH GRANT OPTION;

C.35.3. Bulk Loading

Turn off the initial load extract in background job

initial.load.use.extract.job.enabled=false

There is a bulk loader provided for Teradata that can be used by setting the Channels data loader type to bulk for any reload channels used. This will use the FASTLOADCSV process provided by Teradata.

To use the FASTLOADCSV the Teradata table must be empty.

C.36. Tibero

Use symadmin module install tibero to install driver files, or copy your own files into the lib sub-directory.

This section describes Tibero specific SymmetricDS details.

C.36.1. Permissions

The SymmetricDS database user generally needs privileges for connecting and creating tables (including indexes), triggers, sequences, and procedures (including packages and functions). The following is an example of the needed grant statements:

GRANT CONNECT TO SYMMETRIC;
GRANT RESOURCE TO SYMMETRIC;
GRANT CREATE ANY TRIGGER TO SYMMETRIC;
GRANT EXECUTE ON UTL_RAW TO SYMMETRIC;

Appendix D: Data Format

The SymmetricDS Data Format is used to stream data from one node to another. The data format reader and writer are pluggable with an initial implementation using a format based on Comma Separated Values (CSV). Each line in the stream is a record with fields separated by commas. String fields are surrounded with double quotes. Double quotes and backslashes used in a string field are escaped with a backslash. Binary values are represented as a string with hex values in "\0xab" format. The absence of any value in the field indicates a null value. Extra spacing is ignored and lines starting with a hash are ignored.

The first field of each line gives the directive for the line. The following directives are used:

nodeid, {node_id}

Identifies which node the data is coming from. Occurs once in CSV file.

binary, {BASE64|NONE|HEX}

Identifies the type of decoding the loader needs to use to decode binary data in the pay load. This varies depending on what database is the source of the data.

channel, {channel_id}

Identifies which channel a batch belongs to. The SymmetricDS data loader expects the channel to be specified before the batch.

batch, {batch_id}

Uniquely identifies a batch. Used to track whether a batch has been loaded before. A batch of -9999 is considered a virtual batch and will be loaded, but will not be recorded in incoming_batch.

basetime, {unix_timestamp}

Base create time (as a unix timestamp integer) of first row (insert, update, delete) in batch.

ts, {unix_timestamp}

Offset create time (as a unix timestamp integer) for the current row (insert, update, delete). This is a small offset integer that is added to the base time.

schema, {schema name}

The name of the schema that is being targeted.

catalog, {catalog name}

The name of the catalog that is being targeted.

table, {table name}

The name of the table that is being targeted.

keys, {column name…​}

Lists the column names that are used as the primary key for the table. Only needs to occur after the first occurrence of the table.

columns, {column name…​}

Lists all the column names (including key columns) of the table. Only needs to occur after the first occurrence of the table.

insert, {column value…​}

Insert into the table with the values that correspond with the columns.

update, {new column value…​},{old key value…​}

Update the table using the old key values to set the new column values.

old, {old column value…​}

Represent all the old values of the data. This data can be used for conflict resolution.

delete, {old key value…​}

Delete from the table using the old key values.

sql, {sql statement}

Optional notation that instructs the data loader to run the accompanying SQL statement.

bsh, {bsh script}

Optional notation that instructs the data loader to run the accompanying BeanShell snippet.

create, {xml}

Optional notation that instructs the data loader to run the accompanying DdlUtils XML table definition in order to create a database table.

commit, {batch_id}

An indicator that the batch has been transmitted and the data can be committed to the database.

Example 26. Data Format Stream
nodeid, 1001
channel, pricing
binary, BASE64
batch, 100
schema,
catalog,
table, item_selling_price
keys, price_id
columns, price_id, price, cost
insert, 55, 0.65, 0.55
schema,
catalog,
table, item
keys, item_id
columns, item_id, price_id, name
insert, 110000055, 55, "Soft Drink"
delete, 110000001
schema,
catalog,
table, item_selling_price
update, 55, 0.75, 0.65, 55
commit, 100

Appendix E: Version Numbering

The software is released with a version number based on Semantic Versioning guidelines. In summary, the version is denoted as three integers in the format of MAJOR.MINOR.PATCH. Major versions are incompatible at the API level, and they can include any kind of change. Minor versions are compatible with older versions at the API and binary level, and they can introduce new functions or remove old ones. Patch versions are perfectly compatible, and they are released to fix defects.

Appendix F: Upgrading

The SymmetricDS Pro setup program is an executable JAR file that can upgrade an existing installation to a newer version while preserving settings. Download the setup program from SymmetricDS Pro Downloads. Run the setup program the same way as Standalone Installation. It is recommended to run the upgrade in a lower environment first to become familiar with the procedure and identify any issues. For deployments with a large number of nodes, upgrade the registration server first, then roll out upgrades to client nodes. For a clustered deployment, stop all clustered nodes, upgrade one node and start it before proceeding to the other nodes.

Stop SymmetricDS and close the SymmetricDS Pro Control Center before performing an upgrade.
install1

The first screen shows the SymmetricDS Pro software version. The setup program will ask a series of questions before writing files to disk.

To begin selecting options, click Next.

install2

Carefully read the SymmetricDS Pro License Agreement.

If you accept, select I accept the terms of this license agreement and click Next.

upgrade 1

Specify Upgrade existing software to upgrade an existing version of SymmetricDS to a newer version.

For installing a new copy of SymmetricDS for the first time, see Standalone Installation.

Click Next to continue.

upgrade 2

Provide the directory of the existing installation. The directory will be verified as a valid installation of SymmetricDS before advancing.

Choose the directory path, then click Next to continue.

install5

Select the packages you want to upgrade and verify disk space requirements are met. By default, all packages are selected. Drivers for popular databases are included, but they can be unselected if you don’t plan to use them.

After selecting packages, click Next.

install10

Confirm your upgrade settings look correct.

Click Next to begin upgrading and installing files.

install11

The packages you selected are upgraded or installed to disk. For some upgrades, it may add or adjust settings in the "conf" subdirectory to accommodate the new version.

After it finishes, click Next.

upgrade 3

During the finish step, it will upgrade the module system (optional packages that were installed outside of the setup program) and perform any cleanup.

After it finishes, click Next.

install13

The upgrade is now complete. Choose if you want to open the SymmetricDS Pro Control Center where you can view the server status and open a web console.

Click Done to exit the setup program.

install14

From the SymmetricDS Pro Control Center, you can start the server again. On the first startup after an upgrade, the server will make any required alterations to its runtime tables in the database.

Verify that the server is running, and then click Open Web Console.

Appendix G: Server Migration

The SymmetricDS installation can be moved from one server to another. If the hostname and/or port is changing, then update the sync_url recorded in NODE and let it sync to other nodes before migration, by using the following steps:

  • Stop the SymmetricDS service.

  • Edit the engine properties file and update the sync.url property to use the new hostname and/or port

  • Start the SymmetricDS service.

  • Verify that a batch on the "config" channel is synced to other nodes.

To migrate to a new server, use the following steps:

  • Stop the SymmetricDS service.

  • Copy the SymmetricDS installation directory to the new server.

    • The contents of the "tmp" sub-directory do NOT need copied, which can speed up the migration.

  • In the database, clear the NODE_HOST and LOCK tables to avoid startup checks for the old host.

  • Install the service on the new server (See Running as a Service.)

  • Start the SymmetricDS service.

If any files are excluded from the migration, be careful to at least copy the "security/keystore", "conf/sym_service.conf", and "bin/setenv" files that contain the encryption key and keystore password.

If the registration server is moved, new nodes will need to use a registration.url property that references the sync.url of the new registration server. Optionally, the existing nodes can also be updated with the new registration.url, but they will continue to sync with nodes based on the sync_url recorded in NODE.

Appendix H: Utilities

H.1. DBCompare

The DbCompare utility provides the ability to compare two SQL-based data sources and output a report of the differences, and optionally generate SQL to bring the target into sync with the source.

The dbcompare wizard is accessible for multi-homed nodes from the explore tab of the pro console. The pop-up window asks for source and target engines for the comparison as well as the the tables to be compared.

dbcompare tab

The next window allows specific tables to be selected for comparison.

dbcompare table select

The results of the compare can be presented in four ways:

  1. Show Report

  2. Write Differences to SQL Editor

  3. Write Differences to File

  4. Apply Differences to Target Tables

The Show Report option will open a new tab in the SQL editor showing the results of the compare.

dbcompare results table

The Write Differences to SQL Editor option will write all the required SQL to get the tables in sync and copy it to the current instance of the SQL editor.

dbcompare write sql

The Write Differences to File option will write the required SQL to get the tables in sync to the file system.

The Write Differences to Target Tables option will automatically execute the SQL necessary to bring the tables into sync.

H.2. DBImport

The DbImport utility has the ability to import the structure and data of a tables(s) from a file to database tables.

The DbImport wizard is available from the explore tab of the pro console. The popup window asks for the file format, catalog, schema, and tables to import to the database. It also asks for the number of rows to commit, as well as multiple more specific options for the import.

dbimport tab

Input files can be in SQL, XML, CSV, CSV_DQUOTE, or SYM_XML file formats.

The following is an example of the SQL format used for DbImport.

CREATE TABLE ITEM(
    ITEM_ID INTEGER NOT NULL,
    NAME VARCHAR(100) NULL,
    PRIMARY KEY (ITEM_ID)
);
insert into ITEM (ITEM_ID, NAME) values (1805,'lVlmyxlADKEMZFIfWiipfiBqoeMHsSjsbgsoqBuEAWKhmJfaNA');
insert into ITEM (ITEM_ID, NAME) values (2649,'qjyWrDHXnsfdSDBqUzHMJPkRFQmwmWPWxBPPwaQmgpoQrqLcQC');
insert into ITEM (ITEM_ID, NAME) values (3334,'sCTJrMPFLauMvNrmYVyuLwcrinAVDVNmuHLULCWGYYmHRSmbsc');
insert into ITEM (ITEM_ID, NAME) values (3613,'xCNfevpgVBfegbrXHQOsChxCYPcwbjkpHXZpLFkLxMACaFWJnF');
insert into ITEM (ITEM_ID, NAME) values (3622,'UPTWPHujLPjFvusxAqzKSZCSFXXyhqfkbmholwDvuNdhLfIBGY');

The following is an example of the CSV format used for DbImport.

"ITEM_ID","NAME"
"12","zSYwAyLGsbvsLhYFLBqmeprkfISVTlRnfHwsHFZcmZUpKQMXkT"
"471","RFHEaWcFtPDOkqEtbomEomuEWDQoAuyvJVnjPEsPpaqLxEuWpj"
"113","SDFSDf\"SD\"ggdD"

The following is an example of the CSV_DQUOTE format used for DbImport. CSV_DQUOTE escapes quote characters with double quotes if there are quotes in the column values so it can be correctly parsed by programs like Excel. For example, "SPbSDf""SD""gNdD" parses to SPbSDf"SD"gNdD.

"ITEM_ID","NAME"
"12","zSYwAyLGsbvsLhYFLBqmeprkfISVTlRnfHwsHFZcmZUpKQMXkT"
"471","RFHEaWcFtPDOkqEtbomEomuEWDQoAuyvJVnjPEsPpaqLxEuWpj"
"113","SPbSDf""SD""gNdD"

The following is an example of the XML format used for DbImport.

<database xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="dbimport">
        <table name="ITEM">
                <column name="ITEM_ID" primaryKey="true" required="true" type="INTEGER" size="10">
                        <platform-column name="h2" type="INTEGER" size="10"/>
                </column>
                <column name="NAME" type="VARCHAR" size="100">
                        <platform-column name="h2" type="VARCHAR" size="100"/>
                </column>
        </table>
<table_data name="ITEM">
        <row>
                <field name="ITEM_ID">1805</field>
                <field name="NAME">lVlmyxlADKEMZFIfWiipfiBqoeMHsSjsbgsoqBuEAWKhmJfaNA</field>
        </row>
        <row>
                <field name="ITEM_ID">2649</field>
                <field name="NAME">qjyWrDHXnsfdSDBqUzHMJPkRFQmwmWPWxBPPwaQmgpoQrqLcQC</field>
        </row>
        <row>
                <field name="ITEM_ID">3334</field>
                <field name="NAME">sCTJrMPFLauMvNrmYVyuLwcrinAVDVNmuHLULCWGYYmHRSmbsc</field>
        </row>
        <row>
                <field name="ITEM_ID">3613</field>
                <field name="NAME">xCNfevpgVBfegbrXHQOsChxCYPcwbjkpHXZpLFkLxMACaFWJnF</field>
        </row>
        <row>
                <field name="ITEM_ID">3622</field>
                <field name="NAME">UPTWPHujLPjFvusxAqzKSZCSFXXyhqfkbmholwDvuNdhLfIBGY</field>
        </row>
<table_data>

H.3. DBExport

The DbExport utility has the ability to export the structure of data from database tables to file.

The DbExport wizard is available from the explore tab of the pro console. The popup window asks for the catalog, schema, and specific table(s) to export to a file.

dbexport database export tab

The next tab allows for specific choices for the format and content of the exported file.

dbexport format screen

The format option allows for choices between SQL, XML, CSV, CSV_DQUOTE, and SYM_XML. The Compatibility tab allows for a choice between any of the supported SQL dialects. The list of check boxes allows for more specific options for the content of the file including insert, drop or create scripts. A where clause can also be added to make the SQL more specific.

The last option allows for the scripts to be exported as either a file or directly to the SQL editor within the pro console.

H.4. DBFill

The DbFill utility provides the ability to generate random data to populate any tables in Symmetric.

DBFill allows for any of a nodes tables to be filled with random tables. The popup window is accessible from the explore tab of the pro console.

dbfill tab

The window asks for the Catalog, Schema, and selection of tables to fill.

dbfill settings

The next window allows for a count of rows as well as the interval between execution. Insert, update, and, delete statements can also be specified along with options to ignore errors, and fill dependent tables.

Appendix I: Replication Modes

I.1. Trigger-based

In a trigger-based replication mode application tables are monitered for changes by creating insert, update, and delete triggers on all configured tables. As changes occur on these tables the triggers will capture the row data and old data for the change which will be used as the starting point in the SymmetricDS replication workflow.

Overview

  • Supported on a wide variety of ANSI SQL databases

  • Transactional guarantee that data as changes are committed triggers capture.

  • Changes will be captured even if SymmetricDS is not running (outage, maintenance etc).

  • Less threat of data loss as long as the database is scaled to grow as more data is collected.

  • Old data can also be captured and used in the replication workflow.

  • Triggers are maintained by SymmetricDS and requires no user interaction.

  • Nightly job to verify triggers are in place and rebuilt if table structure changes (SyncTriggers job, scheduling can be configured).

  • Will work with existing triggers on tables as long as database supports multiple triggers on a table. SymmetricDS triggers are designed to fire last after all other triggers to ensure the final state of the data is captured.

More to Consider

  • Requires alter permission to replicated tables so triggers can be installed.

  • Requires SymmetricDS runtime tables to present in the same installation and available for the triggers to write changes.

  • Incurs an additional insert into the SYM_DATA capture table with each DML. === Log-based

In a log-based replication mode application the transactional logs are used to read changes to start the SymmetricDS replication workflow. The implementation varies from vendor to vendor on how the logs are setup and how they are read.

Additional log mining details can be found in Appendix C under each database that supports it.

Overview

  • Potentially low or no overhead on the existing database.

  • No need for alterations to an existing database structure

More to Consider

  • Requires that the transactional logging is turned on and available to the SymmetricDS user.

  • Old data is not available and would not be available to drive business decisions in the replication workflow.

  • If logs roll over before SymmetricDS can read from them there is a potential loss of data === Time-based

In a time-based replication mode setup the changes are ready in by selecting data out of the source based on a schedule. This setup is generally used when read only access is all that is available from the source and a trigger-based or log-based approach is not an option.

Overview

  • Read only access is much easier to achieve on a data source.

  • Less performance impact on application users of the source as data is only selected out of the source system.

More to Consider

  • Tables to configure as sources must have one of two features.

    • A timestamp column to be used as a last update column that will be used to select data with each job execution.

    • A single primary key column that is numeric to be used to select newly inserted data with each job execution.

  • Deletes will not be captured in this approach.

  • Old data will not be captured in this approach.

  • For tables setup based on a numeric primary key only inserts will be captured not updates.

I.1.1. Setup

  • Use the Add Table wizard to select tables for replication.

  • After tables are selected on a time-based node the following screen will be available to configure the appropriate columns to use as part of the replication.

mode time based tables
  • Once tables are selected the changes will be detected as part of the log miner job and the schedule for this job can be configured on the configure jobs screen to run as frequently as needed.

Columns selected for time-based can be either a Timestamp column or a numeric column that represents the single PK column on the table.

I.2. Load Only

A load only node is unable to capture any changes and instead is only used as an end point (target) for other nodes to deliver changes.

Overview

  • User has the choice to install the SymmetricDS runtime tables in a different database than the targeted database.

  • Many more end points are available including NoSQL (Redshift, Mongo), non relational (Cassandra) and even message queues (Kafka).

More to Consider

  • Changes can not be captured for this database while in a load only mode.