- Why Transactions?
- Terminology
- Application Structure
- Opening the Environment
- Opening the Databases
- Recoverability and Deadlock Avoidance
- Atomicity
- Repeatable Reads
- Transactional Cursors
- Nested Transactions
- Environment Infrastructure
- Deadlock Detection
- Performing Checkpoints
- Database and Log File Archival Procedures
- Log File Removal
- Recovery Procedures
- Recovery and Filesystem Operations
- Berkeley DB Recoverability
- Transaction Throughput
Database and Log File Archival Procedures
The third component of the administrative infrastructure, archival for catastrophic recovery, concerns the recoverability of the database in the face of catastrophic failure. Recovery after catastrophic failure is intended to minimize data loss when physical hardware has been destroyedfor example, loss of a disk that contains databases or log files. Although the application may still experience data loss in this case, it is possible to minimize it.
First, you may want to periodically create snapshots (that is, backups) of your databases to make it possible to recover from catastrophic failure. These snapshots are either a standard backup, which creates a consistent picture of the databases as of a single instant in time; or an online backup (also known as a hot backup), which creates a consistent picture of the databases as of an unspecified instant during the period of time when the snapshot was made. The advantage of a hot backup is that applications may continue to read and write the databases while the snapshot is being taken. The disadvantage of a hot backup is that more information must be archived, and recovery based on a hot backup is to an unspecified time between the start of the backup and when the backup is completed.
Second, after taking a snapshot, you should periodically archive the log files being created in the environment. It is often helpful to think of database archival in terms of full and incremental filesystem backups. A snapshot is a full backup, whereas the periodic archival of the current log files is an incremental backup. For example, it might be reasonable to take a full snapshot of a database environment weekly or monthly, and archive additional log files daily. Using both the snapshot and the log files, a catastrophic crash at any time can be recovered to the time of the most recent log archival; a time long after the original snapshot.
To create a standard backup of your database that can be used to recover from catastrophic failure, take the following steps:
- Commit or abort all ongoing transactions.
- Force an environment checkpoint (see db_checkpoint for more information).
- Stop writing your databases until the backup has completed. Read-only operations are permitted, but no write operations and no filesystem operations may be performed (for example, the DBENV[Right Arrow]remove and DB[Right Arrow]open functions may not be called).
- Run db_archive -l to identify all the log files, and copy the last one (that is, the one with the highest number) to a backup device such as a CD-ROM, alternate disk, or tape.
- Run db_archive -s to identify all the database data files, and copy them to a backup device such as a CD-ROM, alternate disk, or tape.
If the database files are stored in a separate directory from the other Berkeley DB files, it may be simpler to archive the directory itself instead of the individual files (see DBENV[Right Arrow]set_data_dir for additional information). Note: if any of the database files did not have an open DB handle during the lifetime of the current log files, db_archive will not list them in its output! This is another reason it may be simpler to use a separate database file directory and archive the entire directory, instead of archiving only the files listed by db_archive.
To create a hot backup of your database that can be used to recover from catastrophic failure, take the following steps:
Archive your databases, as described in the previous step 4. You do not have to halt ongoing transactions or force a checkpoint. In the case of a hot backup, the utility you use to copy the databases must read database pages atomically (as described in "Berkeley DB Recoverability").
When performing a hot backup, you must additionally archive the active log files. Note that the order of these two operations is required, and the database files must be archived before the log files. This means that if the database files and log files are in the same directory, you cannot simply archive the directory; you must make sure that the correct order of archival is maintained.
To archive your log files, run the db_archive utility using the -l option to identify all the database log files, and copy them to your backup media. If the database log files are stored in a separate directory from the other database files, it may be simpler to archive the directory itself instead of the individual files (see the DBENV[Right Arrow]set_lg_dir function for more information).
Once these steps are completed, your database can be recovered from catastrophic failure (see "Recovery Procedures" for more information).
To update your snapshot so that recovery from catastrophic failure is possible up to a new point in time, repeat step 2 under the hot backup instructionscopying all existing log files to a backup device. This is applicable to both standard and hot backups; that is, you can update snapshots made either way. Each time both the database and log files are copied to backup media, you may discard all previous database snapshots and saved log files. Archiving additional log files does not allow you to discard either previous database snapshots or log files.
The time to restore from catastrophic failure is a function of the number of log records that have been written since the snapshot was originally created. Perhaps more importantly, the more separate pieces of backup media you use, the more likely it is that you will have a problem reading from one of them. For these reasons, it is often best to make snapshots on a regular basis.
Obviously, the reliability of your archive media will affect the safety of your data. For archival safety, ensure that you have multiple copies of your database backups, verify that your archival media is error-free and readable, and that copies of your backups are stored offsite!
The functionality provided by the db_archive utility is also available directly from the Berkeley DB library. The following code fragment prints out a list of log and database files that need to be archived:
void log_archlist(DB_ENV *dbenv) { int ret; char **begin, **list; /* Get the list of database files. */ if ((ret = log_archive(dbenv, &list, DB_ARCH_ABS | DB_ARCH_DATA, NULL)) != 0) { dbenv_err(dbenv, ret, "log_archive: DB_ARCH_DATA"); exit (1); } if (list != NULL) { for (begin = list; *list != NULL; ++list) printf("database file: %s\n", *list); free (begin); } /* Get the list of log files. */ if ((ret = log_archive(dbenv, &list, DB_ARCH_ABS | DB_ARCH_LOG, NULL)) != 0) { dbenv_err(dbenv, ret, "log_archive: DB_ARCH_LOG"); exit (1); } if (list != NULL) { for (begin = list; *list != NULL; ++list) printf("log file: %s\n", *list); free (begin); } }