Migrating data from an Octopus 2.6 backup file into an Octopus 2018.10 LTS instance can take a significant time to run (hours or even days in some cases).
We strongly recommend taking the following actions to minimize the migration duration.
Remove Unnecessary Data From Your 2.6 Instance
Our strongest recommendation is to use retention-policies in your Octopus 2.6 instance to remove unnecessary data.
The goal is for the document count in the 2.6 RavenDB to be as low as possible. You can find the document count by viewing the RavenDB studio through the Octopus Manager. The document count is in the footer of the RavenDB studio. Less than 150k documents is a rough guide, though obviously some customers will simply have more required data than this.
The original complete backup can always be retained if it is required for audit purposes.
Limit Historical Data
By default we migrate everything from your backup including all historical data. You can use the
maxage= argument when executing the migrator via the command-line to limit the number of days to keep. For example:
maxage=90 will keep 90 days of historical data ignoring anything older.
RAM, RAM and More RAM
The migrator is a memory-hungry process. Allocate the machine which will execute the migrator process as much memory as possible. The more memory is available, the faster the process will run.
As a rule-of-thumb:
- If your .octobak file is >500MB, allow at least 16GB of RAM
- If your .octobak file is >1GB, allow at least 32GB of RAM
This RAM is only required for the migration, and can be deallocated once it is complete.
To minimize the initial migration time, you can skip migrating the server-task log files.
This option is available as a check-box in the Octopus Manager, or can be supplied as a
--nologs option if running via the command-line.
The logs can always be imported later using the
--onlylogs option if required