Before and After: Digital Media Accessioning Service

by Alice Prael

Before started in 2016 with a plan for centralizing born digital media accessioning.  There’s already a blog post about before, so I’ll skip ahead. “After” started in 2022 with a weird post-it map of born digital archiving workflows. Many thanks to my colleagues for bearing with those colorful semi-organized thoughts – at least there were fun illustrations! 

Blue, red, green, yellow and pink post-its with small writing and drawings related to born digital archives( e..g. "Born Digital on Media in Rare Books" with an image of a book open to a CD sleeve or "Email" with the letter icon and a computer screen open

A sample of the colorful post-its with illustrations from the Born Digital Archiving Mind Map

Some of these thoughts remained on a post-it, some were written into long term goals, and a few turned into work plans, projects, and documentation. One such post-it (okay, a few post-its) turned into a set of recommendations for the Digital Media Accessioning Service.  

In 2023 I worked with Mike Rush in Beinecke Operations and Richard Lynch in Library IT to develop SharePoint infrastructure to manage the service and track submissions of media.  The Born Digital Archives Working Group (now Advisory Group) and submitters to the Service provided feedback and recommendations. And now I’m working with my colleagues in Digital Preservation Services to identify areas of collaboration and capacity building for the service. 

A lot has changed, so here’s the before-and-after overview for the Digital Media Accessioning Service 

  • Before (2016): A Digital Accessioning Support Service for backlogs of digital media. 
  • After (2024): A Digital Media Accessioning Service that includes submitters to the service as collaborators and addresses bottlenecks to increase our overall capacity.
  • Before: By default, the service creates a disk image for all submitted media.  Submitters can request file transfers. 
  • After: By default, the service does a logical file transfer for most submitted media, exceptions for disk imaging floppy disks (due to Kryoflux capture process) and Macintosh computers manufactured prior to 2000 (due to known complications with the file system).  Submitters may still request a disk image.   
    • Impact: Logical file transfers do not include system files or deleted files (which we cannot provide access to and do not intend to preserve but cannot be excluded from a disk image capture), resulting in less digital storage for files that should not be acquired. 
    • Impact: Logical file transfers can more easily be characterized and managed in Preservica. 
    • Impact: In some unique cases, logical file transfers might break software dependencies required for long-term access via emulation. The Software Preservation and Emulation team works closely with Digital Preservation Services to connect with staff who might benefit from requesting disk imaging.
  • Before: Media label must be transcribed to the Title field in the metadata spreadsheet. 
  • After: If Title field is left blank the service will transcribe the media label.  If staff are reviewing media prior to Aspace/Preservica ingest, they will have the opportunity to review and change Titles.
    • Impact: lower barrier to submission, additional service labor 
  • Before: Required each piece of tangible media to be labeled with the Component Unique ID (CUID) 
  • After: The service will label media with the CUID 
    • Considerations: the service confirms the number of media in a box and labels each piece of media with CUID that matches the media type.  If the service is transcribing titles and labeling media with CUIDs, the submitter cannot control the order of CUIDs within a box.  If this is important to the submitter, they must either transcribe media labels to the Title field in the spreadsheet or label each piece of media with the CUID. 
    • Impact: Lower barrier to submission, additional service labor
  • Before: All media is photographed and scanned for Personally Identifiable Information (PII) unless requested otherwise.  
  • After: Photographing media and scanning for PII is now a choice in the submission survey.  This decision is applied to all media in that submission. 
    • Considerations: How will the PII scan and photograph be used?  Will the collection receive further processing, including more collection-specific searching?
    • Considerations: Can the file formats present be scanned for PII?  For example, PII scans will not identify personal information in audio files.
    • Impact: Service saves time and digital storage by not creating unnecessary files.
  • Before: Required the ArchivesSpace URL in submitted metadata spreadsheet (either parent archival object URL for the service to create child item records OR existing archival object to update) 
  • After: ArchivesSpace URL is not required until the spreadsheet is ready for Aspace import.  If submitter does not plan to review data before import, then Aspace URLs are required in the submitted spreadsheet. 
    • Impact: Staff can review files before deciding on descriptive structure, resulting in more integrated description of born digital materials.
  • Before: Submitter uses online survey to send the spreadsheet, then a working copy is saved in network storage until the data is imported to Aspace. 
  • After: Spreadsheet is submitted via SharePoint survey and is immediately stored in a Document Library, where it is available to both the BDS staff and the submitter.  Submitters can also add other library staff as “watchers”, which provides edit access to the spreadsheet.
    • Impact: Increased transparency and collaboration as data can be shared and updated collectively.  Also removes confusion of multiple copies held by each department. The spreadsheet in the SharePoint Document Library is the working copy until the service exports data as CSV and imports to ArchivesSpace.   
  • Before: Submitter does not have access to files until after ArchivesSpace description is updated and files are ingested to Preservica – unless special arrangements were requested in the submission survey or via email. 
  • After: When the submitter is notified that the submission is ready for their review, they can either 1) respond to the notification email and request download access to the entire submission or a specific set of CUIDs to be delivered via OneDrive or 2) Schedule a lab appointment to review materials on a machine in C131.
    • Consideration: This workflow assumes that material is usually deaccessioned at the media level.  If file-level appraisal becomes more common, we will need to automate this process. 
    • Impact: More opportunity for appraisal of born digital material before ingest to Preservica. 
  • Before: Submissions were tracked in a Box spreadsheet, where I manually created a new row for each box submitted to the service, including CollectionName_BoxNumber, SpreadsheetName, number of media and the date of submission.  I updated this Box Tracking Spreadsheet with dates for: files capture, media photographed, scanned for PII, spreadsheet imported to Aspace, files packaged for Preservica ingest, and Preservica ingest complete.  
  • After: Each submitted spreadsheet in the SharePoint Document Library has a Status that can be updated by any staff with access to that submission.  When a status is updated, SharePoint triggers a notification email to the submitter and DPS. 
    • Impact: The Box Tracking Spreadsheet provided an easy way to calculate the number of media and track how quickly materials were accessioned by the service.  The new model relies on Preservica and ArchivesSpace for this type of reporting.  
    • Impact: Submitter receives automatic notifications when a major workflow step is complete (received, in progress, ready for submitter review, ready for ArchivesSpace import, ready for Preservica packaging and ingest). Improved transparency and communication. 
  • Before *as of 2018: My position, Digital Archivist for Yale Special Collection, was in the Manuscript Unit, Beinecke Library Technical Services.  I spent 90% of my time on born digital materials in Beinecke Library, 10% on digital media accessioning for other repositories. 
  • After: As of September 2023, my position is Born Digital Specialist in Digital Preservation Services.  We are still identifying capacity limits and determining the best way for my position as Born Digital Specialist to support Yale Libraries.   
    • Impact: Digital Preservation Services is oriented towards staff training and system maintenance, which aligns with the Born Digital Specialist’s role.  
    • Impact: Better coverage for born digital support.  When the Born Digital Specialist is unavailable, staff can request support from digitalpreservation@yale.edu.   
    • Impact: Lauren Work serves as DPS Liaison for Media Accessioning Service, packaging and ingesting to Preservica for backlog and new acquisitions. 

 

Bonus Round: No Before, Just After 

These are new processes that the service can support but are not part of the standard workflow.

  • Instead of 1 row = 1 piece of media, do you want 1 row = 1 folder or 1 file?  We can use the Notes column in the spreadsheet to document the path to file or folder. 
    • Consideration: This is possible right now but may require development to become a standard workflow option.   
    • Deep-ish dive, feel free to skip: I’m using Excel concatenate function to create Teracopy command line scripts, using the file/folder list from the Notes column.  The Teracopy command line doesn’t include reporting function, so a log file is only created if an error or failed transfer occurs.  So far, the requests have been simple enough to manually separate the File Manifest (created by running Siegfried over the files on original media).  For more complex separation, I could run Siegfried on the post-transfer files in the Content folders instead. 

 

  • If staff need access to born digital content that can’t be accurately rendered using standard tools for legacy files (e.g. FTK Imager, FTK, QuickViewPlus) – Digital Preservation Services has support from the Software Preservation and Emulation team to emulate the computing environment needed to view the content.  The Emulation-as-a-Service Infrastructure (EaaSI) system can be used to provide staff with access to born digital files in an emulated computing environment with software dependencies installed.   
    • Consideration: Requesting emulation will often require additional research and trial-and-error testing to identify the correct combination of OS and software to support the content. 
    • Consideration: Although the EaaSI system is user-friendly, the original software presented in emulation may not be.  For example, if you request access to database files created in MS-DOS, plan for a learning curve to working with MS-DOS database software. 

To Image or Copy -The Compact Disc Digital Audio Dilemma

Anyone with a pre-MP3 music collection has seen this icon before.Compact Disc Digital Audio Icon

It’s possible you didn’t notice it, lightly printed in small text, hidden next to the band name or track listing, but it’s there. In looking through my personal music collection I was surprised to find it on nearly every album from the Velvet Underground to Britney Spears.

This icon identifies optical media as Compact Disc Digital Audio, or CD-DA, a format created specifically for storing digital audio. In 1987 Philips and Sony released IEC 60908, which outlined the technical details for a CD-DA. Most mass produced music in the ’90s was written to a CD-DA because this format allows for higher storage capacity, meaning more musical tracks can fit on a single disc.

Why does this matter to our born digital archives, and specifically for our born digital accessioning service?

Because manufacturers of CD-DAs exchanged more data per sector (i.e. more music on a disc) for a higher error rate than your standard CD-ROM. According to “An Introduction to Optical Media Preservation” by AVPreserve, standard commercial hardware is only 95% accurate at the track level. Meaning that up to 5% of content could be lost if we approached CD-DAs with the same disk imaging workflow used for other digital media.

Our original workflow for optical media required that the service creates an iso image using Forensic ToolKit. However given the high error rate on this type of disc, it’s imperative that we change our workflow for CD-DAs. Disk images are created by reading the data on the disc once, which would create an image including this high error rate. In order to avoid losing up to 5% of the relevant data, we’ve decided to change our workflow from creating a disk image to copying the audio files using a tool made specifically for working with CD-DAs. Following the suggested workflow proposed by AVPreserve, we’ve adopted the use of Exact Audio Copy (EAC), a software tool created specifically for extracting high quality audio tracks. The tool reads each sector of the disc multiple times to detect any errors. If an error is detected EAC continues to reread the disc 16 times and only considers the sector error-free if at least 8 of those 16 attempts retrieved the same data. If fewer than 8 reads match, EAC notifies the user and provides the time position of that error.

But how do we know for the sure if the disc in front of us is a CD-DA or a CD-ROM or a DVD? Although the icon discussed above is a strong indicator of a CD-DA, it isn’t always accurate. Some CDs are formatted as a CD-DA, but lack the printed icon. The tool IsoBuster allows us to review the properties of a disc and determine if it is a CD-DA.

Now that we’ve identified the correct tools for identifying and capturing files from CD-DAs, we need to consider which configurations will work best for our purposes. Here are a few of the areas we’re considering in testing our configurations for EAC. The majority of our research on configurations came from the EAC wiki provided by The Hydrogen Audio Knowledgebase, and testing in our lab.

Secure Mode:

The main benefit of Exact Audio Copy for archives is the secure mode for copying audio media. This is the mode that ensures each disc is read a minimum of two times, and up to 16 times if an error is detected in the first two reads. Secure mode can be enabled through the drive options, but there are a few settings you need to select about your optical media drive. If you don’t know if your drive caches audio data or is capable of retrieving C2 error information, that’s okay! EAC is able to determine what features are available on your drive. By entering a disc into the drive and clicking the “Detect Read Features…” you can automatically apply the correct settings.

Screenshot of EAC detecting features of CD-ROM drive

EAC detecting read features for secure mode of copying

Silent Blocks between Tracks:

Since the gaps are part of how the record creator worked with the material we have decided to preserve these silent blocks. In order to preserve these gaps, we detect them using the “Action” menu. You can confirm that these silent blocks were preserved, by reading the log file after you have completed copying the disc.

Copying Options

EAC has three options for copying media: Copy Selected Tracks Uncompressed, Copy Selected Track Compressed, and Copy Image and Cue Sheet. We decided not to use the compression options because it’s unclear from the tool’s documentation how much loss will be incurred through compression.

Copy Selected Tracks Uncompressed results in a WAV file for each selected track on the disc. This allows the archivist to make appraisal decisions and only capture selected tracks, rather than the entire disc. If a single track has an associated name in EAC, either embedded in the file by the creator, entered into EAC by the archivist, or identified from a connected database, that will be used as the file name. Otherwise the WAV files will be named 01 Track01.wav, 02 Track02.wav, etc. This option does not result in the creation of a cue sheet, but a log file can be created after copying, which contains all of the same information except the amount of silent time between tracks. Instead of recording these timestamps in the cue file, this option appends the silent block to the preceding track.

Both copying options allow the user to create a log file after the copying process has completed. This log file provides information on both the disc and the process used for copying. This includes the selected read mode, which settings were utilized, if silent blocks on the disc were deleted, the table of contents extracted from the CD, and other technical details of the copying process. This log file also provides information about any errors that may have occurred.

Sample log and cue files from test media

Sample log and cue files from test media

Copy Image and Cue Sheet results in a single WAV file for the entire disc and a .cue text file. That file will be named from the information entered in the metadata fields by the archivist following the convention of “CD Artist – CD Title”. If no information is listed, the file will be named the default of “Unknown Artist – Unknown Title.” Since the file naming convention for all other output from the digital accessioning service is to use the media number, this option would require the archivist to enter the media number in the Title field and delete the artist information from the filename. Since this option creates a single WAV file for the entire disc, embedded track file names are no longer associated with the WAV file, but instead are listed in the log file and cue sheet.

The .cue file provides the absolute position of of all track markers, including the times of silent gaps between tracks. Since the tracks are combined into a single WAV file during the copying process, it is important to preserve this track information. The cue file also indicates where and how long the gaps between tracks are. Each track listed include the track name, the composer and performer information as entered in the metadata fields, and the index where the track began. Each track listing includes the Index 00, indicating where the gap began, and Index 01, indicating where the track began. If Index 00 and Index 01 are listed at the same time, there was no gap before the track.

We decided to proceed with the Copy Image and Cue Sheet option.  This option results in a single WAV file, rather than a file for each track, which is the best practice for preservation and follows our practices for preservation copies of digitized audio. The audio file may be split into track files later, according to the information in the cue sheet.  We anticipate that the split track files will be used for access copies and possibly for archivists managing arrangement and description

Metadata Fields:

When working with material from a manuscript repository, the standard music metadata schema provided in EAC is rarely applicable. EAC provides options for “Artist”, “Performer”, “Composer”, and “Genre”, most of which will not usable in our context. However, this information is preserved in the log file, so it may be worthwhile to consider crosswalking our existing metadata schema to this music based schema.

For the moment, this metadata crosswalking is only an idea. The service is only working with the “CD Title” and “CD Artist” fields in EAC, as these fields appear as the filename for the log file. If left blank the log file will be named “Unknown Artist – Unknown Title.log”. For the service we are listing the CD Title as the media number and deleting “Unknown Artist” from the log filename post-creation. This is consistent with the file naming convention for log files created through other capture processes used by the service.

Hybrid CDs: The Next Problem

We’ve recently become aware of a new potential problem with optical media: hybrid CDs. These discs contain both digital audio and other data types and are formatted partially as a digital audio disc and partially as a data disc. One commercial instance of this might be a CD-DA that also contains a bonus music video. The service has not yet come across this type of hybrid disc in our collections, but we’re currently researching how to address it so we’ll be ready.

A Rose by Any Other Naming Convention

 

In working on the Digital Accessioning Service workflow, we approached an interesting question: How do you label the physical media?  A seemingly simple issue, but it becomes more complicated when you consider the range of media types and the risks related to applying labels directly to media. The service requires participating units to label media prior to submission for accessioning both to confirm that we are associating the content with the correct description and to ensure that the physical media can be linked back to the content in the future.  It’s important that these labels be as permanent as the life of the media.  

Although the service requires that all media be labeled prior to submission, we do not prescribe a specific method for labelling.  The final decision on how to apply labels is left to the special collection unit that owns the media.  We tested the following labeling methods in order to provide guidance on how units may label media in way that ensures permanence and minimizes risk of damaging the media.

The labeling question is two fold – how do we apply the label and what should the naming convention be?  Each media piece requires a unique identifier, preferably on the media itself, rather than on a case or other container.  The service uses this unique identifier to confirm metadata prior to imaging.  The identifier is also used as the filename for the disk image and related photographs.  

 

First Question: What is the best naming convention?

First idea: Accession Number + Sequential Number (AccessionNumber-001, AccessionNumber-002, etc)

This plan would rely on existing accession numbers.  In this case half of the work is done for us since many disks have already been assigned an accession number when the larger collection was accessioned.  Using this number for labeling disks also applies semantic meaning to the label.  A disk with an accession number based label could quickly be connected to the larger accession record.  An accession number identifier also allows for more flexibility between units, which may use different accession numbering techniques.

One disadvantage to accession number labeling is that legacy naming conventions have already been written on some processed materials, some of which include a legacy accession number+sequential number convention.  If there are multiple identifiers on a single disk (both legacy naming conventions, and labels assigned by the record creator) there is a risk that the service will use the incorrect label information.  

Another potential drawback to this naming convention is that accession numbers are unique to a given collecting library or museum, but not necessarily unique across all YUL/M collections. Since we accept media from a number of libraries and museums there is no guarantee that the existing accession number will be unique to the service.  Ideally, all identifiers used by the service would be unique to avoid a potential mix-up of material from two different repositories.

Second idea: Barcodes

Barcodes have the advantage that most units already own a barcode scanner and those that don’t can easily purchase one that connects through a standard USB port. The barcode is unique and would remove the risk of someone incorrectly entering a unique ID as it goes through the service.  Any time a long number is rekeyed multiple times there is the risk of mistakes.  The barcode is also small enough that there is little risk of covering existing label information.  The only exception is very small flash drives or memory cards, which would already require external housing for labeling purposes.  Another advantage is that barcodes would obviously answer the question: “Which of these numbers is the correct one to enter?”  As media has been collected for over twenty years here at Yale University Libraries and Museums, the naming conventions have changed and legacy naming conventions still linger on older media.  A barcode workflow would remove the risk of mistakenly entering a legacy identifier.

One major disadvantage of the barcode is that it has no semantic meaning, and is instead a completely random number. Another disadvantage is that our system for descriptive records (ArchivesSpace) already has a field titled “barcode” that refers to the barcoding system applied to containers for managing storage locations of physical items.  The final drawback to applying barcodes is that it requires an adhesive label, which could fall off the media and potentially damage the media and the drive used for reading the media.

 

Decision Time: Accession Number + Sequential Number Wins!

Ultimately we decided that the semantic value and human readability of the accession number label outweighed the ease of scanning a barcode.  The service will most likely be copying the identifier directly from a metadata spreadsheet, which minimizes the risk of entering an incorrect identifier.

 

Second Question: How do we apply our decided-upon naming convention to the media?

First idea: Archival adhesive labels

This plan would require the archivist to write the naming information on an archival label (often used to label folders) and apply it directly to the physical media.  The largest drawback of adhesive labels is their fixed size.  It is unlikely that we would be able to find room for this type of label on a 3.5 inch floppy disk without covering existing label information written by the creator or other archivists.  This is especially problematic if the media may be displayed in an exhibit in the future. Another issue with this kind of label is the risk that it could fall off the media, requiring an archivist to re-image the disk to determine which collection it belongs to.  It’s also possible for the label to fall off while the disk is in use, potentially damaging the disk and the drive used to read it.

Second idea: Pencil

This plan would require the archivist to write the naming information directly on the media in pencil. Most of our plans for labelling require permanent markings on the media, which could be problematic for exhibiting material or if the naming convention were to change over time (or if someone makes a mistake in the original labeling process). Marking the media with pencils would remove the risk of making a mistake permanent on special collection material. However the impermanence of pencils is as much of a drawback as an advantage. Pencil markings smudge and can become illegible, which could require re-imaging to determine the contents of the media. Pencils also create dust, which could make its way through the plastic casing and damage the internal disk as well as the drive mechanism used to read the media.  

Bentley Historical Library (BHL) provides guidance on the risks related to labeling in their IMLS funded report on Removable Media and the Use of Digital Forensics. They recommend against “writing on floppy disks with pencils or ballpoint pens, touching the tape, or affixing labels” because of the internal disk’s fragility.

Third idea: Standard Permanent Marker

Photograph of 3.5 Inch Floppy Disk with fully written label

Some record creators use every inch of their labels, leaving little room for archivists to apply their own naming conventions

 

This plan would require the archivist to write the naming information directly on the media with a permanent marker.  The standard Sharpie marker is already in use for labeling

optical media in many units, so this would be an easy transition.  However, a black marker isn’t visible on all media types and may require the archivist to find space on an already packed label.

 

 

Fourth idea: Silver/Black Paint Pen

This plan would require the archivist to write the naming information directly on the media with an oil-based paint pen.  This would require both a silver and a black pen so that markings would be visible on media of all colors.  These markings would be permanent and once dried would not risk smudging.  The main drawback to this plan would be regarding optical media.  The BHL report recommends against writing on optical media with solvent based markers because they “can penetrate the hard layer and deform, discolor or corrode the disc, causing permanent reading problems for the laser.”  

 

Decision Time: Paint Marker Wins! (with a caveat)

Ultimately we decided there should be slightly different rules for different media.  Solvent based markers are the best solution for media in hard plastic casings (such as floppy disks and hard drives), but could damage optical media.  The Council on Library and Information Resources recommends felt-tip water-based markers for optical media, because these are the least likely to have a damaging chemical reaction.  We decided to use a permanent marker and only apply labels to the inner circle which does not contain data, which will limit the risk of corroding the media.

 

In Conclusion

The Digital Accessioning Service is making recommendations for how a unit may apply labels to their media, however the final decision is left to the unit. We hope that standardized naming conventions and label applications will aid in quality assurance and long term intellectual control of physical media.  By using the naming convention as a label for physical media and the filenames for disk images, packaged transferred files, and associated photographs of media, we can ensure that the digital files will be associated with the media for the long term.

 

The Final Label

Labels written on test media

CD with label written in Sharpie on inner circle

3.5 and 5.25 inch floppy disks with labels written in oil paint pens

Introducing the Digital Accessioning Service

Anyone who has worked with a computer for the past decade or longer probably has a few files saved on floppy disks, zip disks, CDs, and other assorted storage devices. Modern media like CDs and flash drives may still be accessible, but older disks are often unreadable by modern machines, left to languish in attics and filing cabinets. 

The libraries and museums at Yale University are no different. Much of the digital media we have acquired over the last twenty years is trapped on legacy media requiring special hardware and software to access it. Despite the growing presence of born-digital archival material in Yale’s special collections, until now we have not created a system-wide approach to processing these holdings. Repositories with adequate funding and expertise are already providing access to born-digital, but many of these activities have been ad-hoc and the procedures differ based on available technology and expertise. Some repositories have postponed acquiring and processing born-digital media, deciding to wait for the libraries to create a holistic approach.

In January 2015 the Born Digital Working Group was formed to address this need and determine how the different libraries and museums can pool resources and expertise to find a path forward for born-digital archival materials. Our vision is to provide the same level of stewardship for born-digital holdings as is devoted to our physical collections. One of the priority goals is to establish a centralized Digital Accessioning Service for Yale special collections to capture files in a way that maintains their archival integrity and package them for ingest into the Digital Preservation System. This service is still in the beginning stages as we test software and hardware, draft documentation, and ensure that we are ready to begin accessioning archival material. The service will be housed in our new Di Bonaventura Family Digital Archaeology and Preservation Lab, allowing us to provide accessioning services for born-digital media from across Yale University Libraries and Museums.View of the Digital Archaeology and Preservation Lab

Last April the Beinecke’s Technical Services department and the Preservation Department moved to a new facility that includes the shared Digital Archaeology and Preservation Lab. The space currently hosts disk imaging for special collections as well as disk imaging for the general collections managed by Euan Cochrane, Yale’s Digital Preservation Manager. The new lab has more room for collaborations and sharing expertise with staff and visitors. There is additional space and shelving for storage of supplies, tools, and media awaiting accessioning. The lab houses two workstations devoted to disk imaging for special collections, two workstations devoted to imaging disks from the general collection and additional write blockers.

The digital preservation team uses the lab to house legacy computers that can be used to view files in their original environment, interact with digital content that requires original hardware, and test and validate digital preservation approaches such as emulation and migration against the original content executing on contemporaneous legacy hardware. The Digital Preservation team is also creating disk images of the legacy computers which will be attached to emulated versions of the original hardware environments. Using images of the original hard drives with the emulated hardware will help to enable accurate validation of emulated hardware by removing one source of difference between the two environments. Mac Classic II attached to digital forensic station for disk imaging

We use BitCurator, Forensic Toolkit (FTK), and Kryoflux to create disk images in the lab. Each offers different advantages for working with different media types. The extra room also means we can avoid the pesky traffic jams that came up in our former lab space in Sterling Memorial Library.

Now that we have the technology in place, we are working on documentation. That includes reference guides for write blockers, how-to manuals for disk imaging, and workflows to explain how media will make their way from collections through the accessioning process. All of this will inform the service, led by myself as the Digital Accessioning Archivist with the guidance of the Beinecke’s Digital Archivist and the Born Digital Working Group. I’ll be the new go-to person for born-digital accessioning here at Yale University’s Libraries and Museums. For the next month I’ll be focused on getting all our docs in a row so that we can begin accessioning digital media from archival collections across campus. I will also be visiting the various special collections and archives across campus to familiarize myself with the collections and processes in place, so we can make born-digital accessioning a seamless part of archival processing here at Yale. As the born-digital program continues to develop, we hope to invite the Yale community into the lab to learn more about our work with digital preservation, emulation tools, and disk imaging.