The events of September 11 have caused many IT decision makers to rethink their organization’s disaster recovery plans. IT executives’ spending priorities have recently shown a shift towards disaster recovery, security and videoconferencing. In one recent survey, when asked which specific areas will receive more money in 2002, 45 percent of its 598 IT respondents answered with ‘disaster recovery’.
Now that the extreme importance of having a complete disaster recovery solution has been established, let’s move onto what IT executives need to consider when deciding what they need to protect their organization’s most critical information, their data.
A crucial component within a backup and disaster recovery plan that’s often overlooked is the importance of backing up files that are open and in use. One of the biggest problems associated with getting complete and accurate copies of a server’s data is ineffectively backing up open files.
These days, 24/7 access to mission-critical applications such as web sites, databases, and messaging systems is essential. With that in mind, as organizations and the amount of data continue to grow, trying to schedule backups during periods of non-activity has become nearly impossible for IT administrators.
Because a lot of these mission-critical applications are in use 24/7, consequently, they don’t get properly backed up. To lose such essential data can be disastrous for any organization.
Many backup software packages will skip open files initially, and then go back and try to access them again at the end of the backup. In many cases, however, the files are still open and therefore not backed up. Even if the backup software is able to access a file in a subsequent attempt, that file will not be synchronized with other related files, which can ultimately lead to corrupt data on tape.
In another scenario, the backup package could force the backup of open files. If any changes occur to a file during the backup, parts of the changes might be captured by the backup software, while other parts may be in areas of the file that have already been read for backup. This will also lead to corrupt data on tape. Some backup packages offer an open file solution through the use of application-specific or generic agents.
Dedicated application agents are available for a handful of database and email applications. They are designed to work with specific versions of the applications and only with a particular backup software program. Using application-specific agents can be quite costly, as agent upgrades may be necessary when deploying updated versions of applications and/or backup software.
Generic agents are intended to work with all applications, but are generally only compatible with a single backup package. Changing versions of backup software often requires customers to obtain updates to their generic agents from the vendor in order to continue using those agents in the new environment. Additionally, generic agents are often limited in their synchronization capabilities, providing file-by-file or volume-by volume synchronization . This allows for compromised relational integrity of databases that span several volumes, leading to corrupt data on tape. In some cases, administrators must manually identify groups of related files that need to be handled in a synchronized manner.
The solution to protecting open and in use files during the backup process is finding a complementary utility that helps the backup software successfully capture open files, without locking users out of applications. If that solution was compatible with all backup software and hardware, there would be no need for costly application-specific agents.
John Jones is the president and CEO of St. Bernard Software.
St. Bernard Software is maker of Open File Manager that provides the functionality described above. For more information or a free trial of the program, visit www.stbernard.com.