A File Index job must be set up and run in order to search archived files—see Configure File Archiving.The File Index Job will also allow you to perform policy based file archiving in the indexed locations.
In NetGovern, the File Index job is also called a crawler job. You must create a crawler job for each location whose files you want to index.
To configure a crawler job for the location, you must firs register the relevant application(s) with Office 365, create the required location in the NetGovern Admin UI, configure the connection to the external data source, and perform User Mapping—start with Configuring Locations.
- Log into the NetGovern Admin UI.
- Select Archiving > Agents > File Indexing > name of your Index Location. This is the name of the previously configured Live Location or Storage Location.
- The Job Settings tab opens.
- In the Description field, enter a description for the job. This is helpful if there are multiple jobs of the same type.
- In the Job Priority section, set a priority, choosing from Low, Normal, and High. Normal is the default priority.
This setting is useful in scenarios where multiple jobs are running concurrently and you want to control which job takes priority with respect to thread allocation. The job priority you select determines which job in the queue is selected next by the JobManager process, which is responsible for allocating queued job threads to available thread slots on the archive nodes. Prioritization is categorized according to user account. For example, you may want to assign higher priority to crawling files created by VIP users. If a normal priority job is already running and using all available job threads, setting the priority to high and executing it will direct any freed threads to be used on the new high-priority job. This feature works in conjunction with load balancing in order to control crawling job distribution.
- In the Mode section, select the mode in which the index will operate:
- Continuous: Indexes operates on a continual basis without interruption. The index keeps up to date with all changes in the location.
- Pause: Stops indexing. Once the initial index is built, there is no need to continue running it.
- Pause and delete indexes: Stops indexing and deletes the existing index.
Then, it re-indexes once, and pauses.
- Select the Criteria tab.
- In the Data Source section, select the location you want the crawler to crawl.
- In the Indexer section, select one of the following:
- lncremental: Indexes only new data since the last index update. This is an efficient option if your users need the latest data or if there is a constant flow of new data being created in the location.
- Recreate: Rereads the entire contents of the location and indexes it. This will require some time depending on the size of the location.
- In the Options section, you can specify the (sub)folders to be indexed. You can do this through the process of inclusion or exclusion. Select one of the following:
- Extract files from all folders: Does not restrict the scope of the crawling job to specific (sub)folders.
- Include only the following folders: Restricts the scope of the crawling job by including only the specified (sub)folders. Enable Recursive to apply the inclusion policy to the subfolders of a specified folder.
- Process all folders except the following: Restricts the scope of the crawling job by excluding the specified (sub)folders. Enable Recursive to apply the exclusion policy to the subfolders of a specified folder.
- (Optional) For troubleshooting purposes, configure the log settings—see Configuring Logging.
- (Optional) For troubleshooting purposes, configure the report settings—see Viewing Reports.
- Click Save at the bottom-left of the screen.