For content digitized before it comes to us, see Incoming Digital Content.
In general, digitized content is:
- created in Digital Services,
- temporarily stored on the Share drive (a Windows server) for:
- quality control verification and
- batch completion,
- and then uploaded to the libcontent server in two forms:
- derivatives, for web delivery, and
- the archival files, for archiving.
Thus, it makes sense to divide our documentation by what happens on the Share drive, what happens on the Linux Server (Server-Side Work), and the workflow that ties them together:
Some work, however, crosses boundaries. We test our content for validity and well-formedness during the quality control process, prior to uploads. The information generated via this testing is used in creating our standardized preservation metadata, some of which is then stored in a database for easy search and retrieval.
Once in the archive, content needs to be managed for long-term access.
- tracking and monitoring different formats so we know where to find content in need of migration
- tracking what is in LOCKSS and when it was added, and pushing for complete content in LOCKSS
- checksum verifications weekly
- extraction and storing of appropriate technical metadata, including some in the database for easy access
- documenting anomalies that will need to be managed long-term
- ensuring we have adequate and viable backups
- updating metadata, transcripts, and tags
- documenting the storage system, content and structure, including software needed for access and reconstruction
- preservation planning
- business continuity planning
- disaster recovery preparation
For information on incoming digital content, see Born_Digital.