*** <tt>HashAttribute</tt> – specify if a hash should be created (true or false).

+

**** <tt>Name</tt> (required) – attributes name.

−

*** <tt>KeyAttribute</tt> – creates a key for this object, for example for record id (true or false).

+

**** <tt>HashAttribute</tt> – specify if the attribute is used for the hash used for delta indexing (''true'' or ''false''). Must be true for at least one attribute which must always have a value. Usually the attribute containing the ''LastModifiedDate'' will be a good candidate to set this to ''true'' for.

−

*** <tt>Attachment</tt> – specify if the attribute return the data as attachment of record.

+

**** <tt>KeyAttribute</tt> – specify if the attribute is used for creating the record ID (''true'' or ''false''). Must be true for at least one attribute. All key attributes must identify the file uniquely, so usually you will set it ''true'' for the attribute containing ''Path'' FileAttribute.

−

+

**** <tt>Attachment</tt> – specify if the attribute return the data as attachment of record.

+

*** sub elements:

+

**** <tt>FileAttributes</tt> - specify the file attribute to write into the target attribute. The content of the element must be one of

+

***** ''Name'': name of file, without the directory path

+

***** ''Path'': complete path including file name.

+

***** ''Size'': size in bytes.

+

***** ''LastModifiedDate'': Date of last modification

+

***** ''Content'': Content of file. Unconverted binary if written to an attachment. Else the crawler tries to detect the encoding and converts the content to a string (with fallbacks to UTF-8 or default encoding of the operating system).

+

***** ''FileExtension'': The part of the filename after the last "." character (without the dot). An empty string if the filename does not contain a dot.

* <tt>Process</tt> – contains parameters for gathering data.

* <tt>Process</tt> – contains parameters for gathering data.

** <tt>BaseDir</tt> – the directory the crawling process begin (if is null, cannot be found/access or is not a directory a CrawlerCriticalException will be thrown).

** <tt>BaseDir</tt> – the directory the crawling process begin (if is null, cannot be found/access or is not a directory a CrawlerCriticalException will be thrown).

Crawling configuration explanation

The root element of crawling configuration is DataSourceConnectionConfig and contains the following sub elements:

DataSourceID – the identification of a data source

SchemaID – specifies the schema for a crawler job

DataConnectionID – describes which agent crawler should be used

Crawler – implementation class of a Crawler

Agent – implementation class of an Agent

CompoundHandling – specify if packed data (like a ZIP containing files) should be unpack and files within should be crawled (YES or NO).

Attributes – list all attributes which describe a file.

Attribute

attributes:

Type (required) – the data type (String, Integer or Date).

Name (required) – attributes name.

HashAttribute – specify if the attribute is used for the hash used for delta indexing (true or false). Must be true for at least one attribute which must always have a value. Usually the attribute containing the LastModifiedDate will be a good candidate to set this to true for.

KeyAttribute – specify if the attribute is used for creating the record ID (true or false). Must be true for at least one attribute. All key attributes must identify the file uniquely, so usually you will set it true for the attribute containing Path FileAttribute.

Attachment – specify if the attribute return the data as attachment of record.

sub elements:

FileAttributes - specify the file attribute to write into the target attribute. The content of the element must be one of

Name: name of file, without the directory path

Path: complete path including file name.

Size: size in bytes.

LastModifiedDate: Date of last modification

Content: Content of file. Unconverted binary if written to an attachment. Else the crawler tries to detect the encoding and converts the content to a string (with fallbacks to UTF-8 or default encoding of the operating system).

FileExtension: The part of the filename after the last "." character (without the dot). An empty string if the filename does not contain a dot.

Process – contains parameters for gathering data.

BaseDir – the directory the crawling process begin (if is null, cannot be found/access or is not a directory a CrawlerCriticalException will be thrown).

Filter – select file type and crawling mode.

Recursive – (true or false).

CaseSensitive – true or false

Include – file to crawl.

Name - String e.g. "*.txt" (crawl all text files). Everything that is not included is excluded automatically. You could use a star * as wildcard.

Exclude – files to leave out while crawling.

Name – String e.g. "*test*" (leave out all text files which have test in the filename).