(Email Archive ⇒ Settings ⇒ Storages button;
File Archive ⇒ Settings ⇒ Storages button;
SharePoint Archive ⇒ Settings ⇒ Retentions button;
Custom plugins ⇒ General ⇒ Storages button)
For storage configurations, open the Storages page (navigate to Storages button on your ribbon). The storages configured on this page can be selected as a destination for the processed binaries when configuring a certain contentACCESS job. contentACCESS supports Disk storage (most frequently used type), H&S Hybrid Store, Perceptive, Datengut storage etc. The table of storages is initially empty.
To configure a new storage, click on + new on the Storages page. The Storage repository window will open. Type in the Store name and select a Store type from the list. The required storage settings depend on the storage type that you have selected.
Configurations of the most frequently used storage types will be detailed in the following sections of this chapter:
✓ Disk store type
This store type is used if the user would like to store the binaries on a single local or remote disk. This is the most frequently used store type from the types listed above. After this store type has been selected, fill in the Path (the target destination for the binaries) and enter credentials if required.
The user also has the possibility to decide if he wants to use the Compression function. With checking this checkbox, all files larger than 4 kilobytes will be compressed, except of already compressed file formats such as JPG, MP3 etc. This feature may slow down the store functionality, but on the other hand it will spare storage space.
It is also possible to enable SnapLock for disk store. To do this, check the Enable SnapLock checkbox – thanks to this the storage system can prevent deletion of the files until the specified retention is expired. This feature can be used for storage systems which support this function, for example NetApp.
If you wish to keep only one copy of the physical file in the store, check the Use single instancing checkbox.
To enhance the storage security, it is possible to store files in encrypted form in the disk storage by checking the Use file encryption checkbox. This way the files will be encrypted (using the AES encryption algorithm) and stored in one go. The encryption key must be stored in case of database failure, without it the stored files will be unreadable. Download it by clicking on the Download link in this section. If the user tries to save storage configuration without downloading the encryption key, system will warn about this.
Under Database settings select a Database connection. This configuration will play an important role when using the Storage replication plugin. It is also possible to run a test control via Storage replication plugin. We recommend NOT to change the already configured database that the disk storage uses (however, it can be changed using the “Change” button if it is needed).
We also advise to verify the connection using the Test button. The test is checking if the right credentials have been specified, and if the connection with the database has been established.
When using Disk store type, please consider requirements for data deduplication. Data deduplication finds and removes duplication within data on a volume while ensuring that the data remains correct and complete. This makes it possible to store more file data in less space on the volume. Deduplication is not supported on:
✓ System or boot volumes;
✓ Remote mapped or remote mounted drives;
✓ Cluster shared volume file system (CSVFS) for non-VDI workloads or any workloads on Windows Server 2012;
✓ Files approaching or larger than 1 TB in size;
✓ Volumes approaching or larger than 64 TB in size.
What to consider before a previously configured disk storage is changed?
In some cases, it may happen that the administrator needs to move the already configured storage to a new location (e.g. if “C” disk got full and the storage should be moved to disk “D”) In this case, the following steps must be executed to ensure an access to the already archived data and to continue with archiving the new data:
- Stop all running jobs!!!
- Move your old storage location manually into the new storage location (e.g. the folder of “C” disk into a new location on “D” disk);
- Open the Storages page on the contentACCESS Central Administration ribbon, locate your old storage in the list and double click on it to open the Storage repository window for that storage;
- In the Storage repository window specify the new storage path (where you moved your old storage location);
These settings will ensure, that the archive and restore jobs will automatically use this new storage path. The administrator will not be required to change the storage settings on the jobs’ configuration page, too. The new value will be automatically used by the jobs, where the “old” storage was already configured.
contentACCESS supports the connection with the Hybrid Store. This connection allows contentACCESS to connect to any third-party storage that is supported by this store type. If you want to store the binaries in the Hybrid Store, select it from the Store type dropdown list and specify the required connection settings. The following storage settings are required:
Store name: optional name for the Hybrid Store that contentACCESS will use
Store type: HybridStore
Server name: the server where the Hybrid Store is installed
Binding: http – universal protocol
net.tcp –can be used only in case that the Hybrid Store and contentACCESS are in the same domain
net.pipe – the fastest and recommended protocol; can be used only in case that the Hybrid Store and contentACCESS are installed on the same machine
Use secure connection: Check this option to allow a secure connection with the Hybrid Store. The communication will be secured by a Windows authentication (the contentACCESS service user will be used).
Hybrid store tenant: Click on “Load” to load the list of available tenants based on your Hybrid Store configurations and select the one that should be applied. There are some requirements to load the tenants:
- NET.TCP or NET.PIPE connection must be used (on HTTP the loading is not supported)
- the contentACCESS service user must be a local system administrator on the HybridStore machine
If the tenants are not loaded, then the user needs to enter the HybridStore tenant ID (GUID) manually.
Scheme: Click on “Load” to load the available Hybrid Store schemes and select the one that should be applied. If Hybrid Store uses secure connection and the option is not checked in the dialog, the schemes will not be loaded.
Database settings: Select here an already created database that contentACCESS will use to store the necessary data. It is NOT recommended to change the already configured database that the Hybrid store uses. However, it can be changed using the “Change” button if it is really necessary.
It is recommended to verify the database connection using the “Test” button at the bottom of the dialog.
This storage type is currently used when synchronizing emails in multiple mailboxes. For more information please refer to Email synchronizer plugin.
After this store type has been selected in Storage settings, name your storage. contentACCESS will use this name to display the storage on the Storages page. Further configure the following sections in the dialog:
Connection configuration: necessary settings to establish a connection with the Datengut storage
- Endpoint URL: Datengut service URL
- Api key: optional setting, any value can be entered here
- Email archive default folder id: during the archiving process Datengut storage saves the data into a specific storage folder; the ID of this folder must be set here
User credentials: set the Username and Password that can be used to connect to the storage
Plugin configuration: The email synchronizer job(s) that are already created in Custom plugins ⇒ General ⇒ Jobs will be listed in the dropdown list. Select from the dropdown list the Email synchronizer job that will collect the metadata of the archived emails in a queue. This job is used to synchronize the email message categories in multiple mailboxes based on the metadata that are saved into its queue.
Database settings: select an already created database that Datengut storage will use to store the necessary data
After this store type has been selected, name your storage. contentACCESS will use this name for the given storage on the Storages page. The user is further required to specify the following connection parameters:
- Server name: the server name where the storage is installed
- CC library path: path on which the Classic Connector jar files can be reached, e.g. C:\Program Files\SAPERION\Application\scr\scr-classicconnector\lib
- JAVA_HOME: JAVA home directory must be set depending on the application bitness, x64 or x86, e.g.: C:\Program Files\Java\jre1.8.0_66
- Tenant: in case of non-multitenant system this fields is blank, otherwise the Perceptive tenant should be specified
- Login type: “INDEX”, “ADMIN” or “ELM” can be used, the recommended type is “INDEX”
- Authentication: choose the applicable authentication from the dropdown list; the recommended is “UserName”
- User, Password: specify the applicable user and credentials to be applied for the connection with the storage
- DDC: the DDC name where the files will be packed
✓ Kendox storage
Kendox storage can be used for various operations like email and file archiving, SharePoint archiving and publishing data into SharePoint libraries and custom lists as well. In the Storage repository window, select the Kendox store type, name your storage and configure the connection as follows:
Connection configuration section:
- Endpoint URL: Kendox service URL
- Culture: The names of folders, properties will be shown in the language specified here. The language can be a country-code (e.g. de) or both the language and country-code (e.g. de-CH).
- Username and password: Specify the Username and Password that can be applied to connect to the storage.
Use empty password: Use empty password. If this is checked in, a password empty string is saved. Otherwise – if the username was not changed – the old password is saved.
- Default folder path: During the archiving or publishing process the data will be saved/published from the folder, which is specified in the path.
- Mobile Web Client URL: This configuration option needs to be set only in case that the storage is used in a SharePoint Publishing job, and the job is configured to create custom lists with links pointing to the source documents. The URL must be the Kendox Mobile Web Client’s URL, where the published documents can be accessed. The URL is used to generate the retrieve URL of the published document.
Import template configuration:
In the archiving process, import template is used to map the metadata between the source document and the newly created document in the Kendox store. Click the “Get available import templates” button to get the available import templates. This import template is only used if the target folder doesn’t have any default import template, otherwise it is ignored.
From the given storage’s context menu, on the Storages page, you can modify (Edit/Delete/Set default) the storage settings by clicking on ellipsis (…). The configurations set in the Storage repository window can be viewed in the grid.
The “manage access” option allows to grant access permissions on the selected storage for a second user. This “manage access” option is available for the logged on administrator, if his role assignment contains the Edit repository items – All allowed permission on the tenant. Read more in Managing access to contentACCESS objects.
✓ Azure storage
Azure Storage is a cloud storage solution. It also provides the storage foundation for Azure Virtual Machines and is accessible from anywhere in the world, from any type of application and any type of device. Azure storage is using blob store to store its metadata.
In case of the user having/using a German cloud (dedicated and isolated Microsoft Azure version for Germany), the Azure storage account name must contain the .core.cloudapi.de suffix. This is displayed in Azure configuration, when the user creates the storage.