6.1.5. Deployment with Provisioning

Version Support: 2.2.1 and later

You can setup the extractor from Oracle to automatically read and
provision the slave database by using the
Section 8.11, “Using the Parallel Extractor”. The parallel extractor
reads information from the source database schema in chunks and then feeds
this information into the THL data stream as row-based INSERT operations.
When the slave connects, these are applied to the slave database as with a
normal INSERT operations. The parallel extractor is particularly useful in
heterogeneous environments such as Oracle to MySQL where the slave data
does not already exist on the slave.

The basic provisioning process operates in two stages:

Provisioning data is extracted and inserted into the THL. One event is
used to contain all of the data from a single table. If the table is
too large to be contained in a single event, the data will be
distributed over multiple events.

Once provisioning has finished, data is extracted from the CDC as
normal and added to the THL

Important

The parallel extractor is not restart safe, and the process should not
be interrupted.

This allows existing data to be extracted and processed through the
replicator path, including filters within the applier. Once the initial
data has been extracted, the change data to be applied.

To use the parallel extractor to provision data into the slave, the
configuration must be performed as part of the installation process when
configuring the master replicator.

Alternatively, the system change number (SCN) identified when CDC
capture is first enabled through setupCDC.sh
can be used to provide a point-in-time provisioning. This can be
useful if the data has previously been loaded and then CDC started
by enabling provisioning from the start point. To use this method,
identify the start position indicated by
setupCDC.sh: