Building Community for Archivematica

By Shira Peltzman, Nick Krabbenhoeft and Max Eckard

In March of 2018, the Archivematica User Forum held the first in an ongoing series of bi-monthly calls for active Archivematica users or stakeholders. Archivematica users (40 total!) from the United States, Canada, and the United Kingdom came together to share project updates and ongoing challenges and begin to work with their peers to identify and define community solutions.


The Archivematica user community is large (and growing!), but formal communication channels between Archivematica users are limited. While the Archivematica Google Group is extremely valuable, it has some drawbacks. Artefactual prioritizes their paid support and training services there, and posts seem to focus primarily on announcing new releases or resolving errors. This sets an expectation that communication flows there from Artefactual to Archivematica users, rather than between Archivematica users. Likewise, Archivematica Camps are an exciting development, but at the moment these occur relatively infrequently and require participants to travel. As a result, it can be hard for Archivematica users to find partners and share work.

Enter the Archivematica User Forum. We hope these calls will fill this peer-to-peer communication void! Our goal is to create a space for discussion that will enable practitioners to connect with one another and identify common denominators, issues, and roadblocks that affect users across different organizations. In short, we are hoping that these calls will provide a broader and more dynamic forum for user engagement and support, and ultimately foster a more cohesive and robust user community.


The User Forum is not the first group created to connect Archivematica users. Several regional groups already exist; the Texas Archivematica Users Groups and UK Archivematica Users Group (blog of their latest meeting) are amazing communities that meet regularly. But sometimes, the people trying to adapt, customize, and improve Archivematica the same way you are live in a different time zone.

That situation inspired the creation of this group. After realizing how often relationships would form because someone knew someone who knew someone doing something similar, creating a national forum where everyone had the chance to meet everyone else seemed like the natural choice.


It takes a lot to build a new community, so we have tried to keep the commitment light. To start with, the forum meets every two months. Second, it’s open to anyone using Archivematica that can make the call, 9AM on the West Coast, 12PM on the East Coast. That includes archivists, technologists, developers and any other experts actively using or experimenting with Archivematica.

Third, we have some in-scope and out-of-scope topics. In-scope includes anything that helps us continue to improve our usage of Archivematica: project announcements, bug tracking/diagnosis, desired features, recurring problems or concerns, documentation, checking-in on Archivematica implementations, and identifying other users that make use of the same features. Out-of-scope includes topics about getting started with digital preservation or Archivematica. Those are incredibly important topics, but an over commitment for this group.

Finally, we don’t have any official relationship with Artefactual Systems. We want to develop a user-led community that can identify areas for improvements and contribute to the long-term development of Archivematica. Part of the development is finding our voice as a community.

Current Activity

As of this blog post, the Archivematica Users Forum is two calls in. We’ve discussed project announcements, bug tracking/diagnosis, recurring problems or concerns, desired features (including this Features Request spreadsheet), local customizations and identifying other users that make use of the same features.

We spent a good deal of time during our first meeting on March 1, 2018 gathering and ranking topics that participants wanted to discuss during these calls, and intend to cover them in future calls. These topics, in order of interest, include:

Topic Number of Up-votes
Processing large AIPs (size and number of files) 12
Discussing reporting features, workflows, and code 10
How ingest is being tracked and QA’ed, both within and without Archivematica 9
Automation tools – how are people using them, issues folks are running into, etc. 7
How to manage multi-user installations and pipelines 7
Types of pipelines/workflows 7
Having more granularity in turning micro-services on and off 6
Troubleshooting the AIC functionality 3
What other types of systems people are using with Archivematica – DPN, etc. 3
Are people doing development work outside of Artefactual contracts? 2
How to add new micro-services 2
How to customize the FPR, how to manage and migrate customizations 2
How system architectures impact the throughput of Archivematica (large files, large numbers of files, backup schedules) 1

As you can see, there’s no shortage of potential topics! During that meeting, participants shared a number of development announcements:

  • dataverse Integration as a data source (Scholars Portal);
  • DIP creator for software/complex digital objects via Automation Tools (CCA);
  • reporting – development project to report on file format info via API queries (UCLA/NYPL);
  • turning off indexing to increase pipeline speed (Columbia);
  • micro-service added to post identifier to ArchivesSpace (UH); and
  • micro-service added to write README file to AIP (Denver Art Museum).

During our second meeting on May 3, 2018, we discussed types of pipelines/workflows as well as well as how folks decided to adopt another pipeline versus having multiple processing configurations or Storage Service locations. We heard from a number of institutions:

  • NYPL: Uses multiple pipelines – one is for disk images exclusively (they save all disk images even if they don’t end up in the finding aid) and the other is for packages of files associated to finding aid components. They are considering a third pipeline for born-digital video material. Their decision point on adopting a new pipeline is whether different workflows might require different format policies, and therefore different FPRs.
  • RAC: Uses multiple pipelines for digitization, AV, and born-digital archival transfers. Their decision point is based on amount of processing power required for different types of material.
  • Bentley: Uses one pipeline where processing archivists arrange and describe. They are considering a new pipeline with a more streamlined approach to packaging, and are curious when multiple configurations in a single pipeline is warranted versus creating multiple pipelines.
  • Kansas State: Uses two pipelines – one for digitization (images and text) and a second pipeline for special collections material (requires processing).
  • University of Houston: Uses two pipelines – one pipeline for digitization and a second pipeline for born-digital special collections.
  • UT San Antonio: Uses multiple configurations instead of multiple pipeline.

During that call, we also began to discuss the topic of how people deal with large transfers (size or number of files).

Next Call and Future Plans!

We hope you will consider joining us during our next call on July 5, 2018 at 12pm EDT / 9am PDT or at future bi-monthly calls, which are held on the first Thursday of every other month. Call in details are below!

Join from PC, Mac, Linux, iOS or Android:

  • iPhone one-tap (US): +16699006833,854186191# or +16465588656,854186191#
  • Telephone (US): +1 669 900 6833 or +1 646 558 8656
  • Meeting ID: 854 186 191

International numbers available:

Shira Peltzman is the Digital Archivist at the University of California, Los Angeles Library.

Nick Krabbenhoeft is the Head of Digital Preservation at the New York Public Library.

Max Eckard is the Lead Archivist for Digital Initiatives at the Bentley Historical Library.



A Day in Review: Personal Digital Archiving Conference Recap

by Valencia Johnson

The Personal Digital Archiving conference, which took place April 23-25, 2018, was hosted by the University of Houston. The conference was a mixture of archival professionals, librarians, entrepreneurs, and self-taught memory workers.  The recurrent theme this year, from the perspective of a newcomer at least, was personal digitization. Each demographic offered battle-tested advice for digitization and digital preservation. From these personal testimonies several questions occurred to me and other conference attendees. How is the digital world transforming memory and identity? How can the archival community improve the accessibility of tools and knowledge necessary to create and manage digital cultural heritage? What does it look like when we work with people instead of working for people? If these questions trigger a post-modernism bell in your mind, then you are on the right path.

Each presentation touched upon the need within communities to preserve their history for one reason or another. The residents of Houston are in some ways still recovering from Hurricane Harvey; institutions and homes were flooded, and pictures and home videos were lost to the gulf. Yet, through this disaster the Houston community is finding ways to rebuild and recapture a small piece of what was lost. Lisa Spiro opened the first day of the conference with her presentation “Creating a Community-Driven Digital Archive: The Harvey Memories Project.” This archive aims to document the experience of the survivors of Harvey and offer an additional personal narrative to the official record of the disaster. Expected to launch in August 2018, the first anniversary of  Hurricane Harvey, the project is built by community members and something to keep an eye out for.

The following session was comprised of multiple presenters diving into community archives. Presentations covered how researchers Ana Roeschley’s and Jeonghun (Annie) Kim’s project about a memory roadshow in Massachusetts is uncovering the complex nature of human memory and attachment; Sandra Wang’s quest to preserve her family history by travelling to China and interviewing her grandfather about topics like shame and self-doubt; and Lisa Cruces’s work with Houston Archives Collective, an organization that educates and supports efforts of the community to preserve their history for themselves. Finally, all the way from Alaska, Siri Tuttle and Susan Paskuan discussed the Eliza Jones’ Collection, a true collaboration between an institution and a community to preserve and use material vital to interior Alaskan native communities.

This is a slide from Scott Carlson’s presentation “Megaton Goes Boom: Archiving and Preserving My Father’s First Comic Book,” 25 April 2018.

Later that day were lightning talks about tools useful in the digital age. For example, did you know you can save voicemails? I did not, but thanks to Lucy Rosenbloom’s presentation, I know iPhone users are able to save the voicemails by clicking the square box with the up arrow and emailing the message as a .mp4. Here is a link to a useful article about saving voicemail. Rosenbloom converts her .mp4s into .mp3s and she also uses an auto transcription tool to create transcripts of her messages. The day winded down with personal tales of archiving family history solo and on a budget from Leslie Wagner and Scott Carlson respectively. For more information about the tools and projects discussed at the conference, please visit the program.

Valencia L. Johnson is the Project Archivist for Student Life for the Seeley G. Mudd Manuscript Library at Princeton University. She is a certified archivist with an MA in Museum Studies from Baylor University.

Diving into Computational Archival Science

by Jane Kelly

In December 2017, the IEEE Big Data conference came to Boston, and with it came the second annual computational archival science workshop! Workshop participants were generous enough to come share their work with the local library and archives community during a one-day public unconference held at the Harvard Law School. After some sessions from Harvard librarians that touched on how they use computational methods to explore archival collections, the unconference continued with lightning talks from CAS workshop participants and discussions about what participants need to learn to engage with computational archival science in the future.

So, what is computational archival science? It is defined by CAS scholars as:

“An interdisciplinary field concerned with the application of computational methods and resources to large-scale records/archives processing, analysis, storage, long-term preservation, and access, with aim of improving efficiency, productivity and precision in support of appraisal, arrangement and description, preservation and access decisions, and engaging and undertaking research with archival material.”

Lightning round (and they really did strike like a dozen 90-second bolts of lightning, I promise!) talks from CAS workshop participants ranged from computational curation of digitized records to blockchain to topic modeling for born-digital collections. Following a voting session, participants broke into two rounds of large group discussions to dig deeper into lightning round topics. These discussions considered natural language processing, computational curation of cultural heritage archives, blockchain, and computational finding aids. Slides from lightning round presenters and community notes can be found on the CAS Unconference website.

Lightning round talks. (Image credit)


What did we learn? (What questions do we have now?)

Beyond learning a bit about specific projects that leverage computational methods to explore archival material, we discussed some of the challenges that archivists may bump up against when they want to engage with this work. More questions were raised than answered, but the questions can help us build a solid foundation for future study.

First, and for some of us in attendance perhaps the most important point, is the need to familiarize ourselves with computational methods. Do we have the specific technical knowledge to understand what it really means to say we want to use topic modeling to describe digital records? If not, how can we build our skills with community support? Are our electronic records suitable for computational processes? How might these issues change the way we need to conceptualize or approach appraisal, processing, and access to electronic records?

Many conversations repeatedly turned to issues of bias, privacy, and ethical issues. How do our biases shape the tools we build and use? What skills do we need to develop in order to recognize and dismantle biases in technology?

Word cloud from the unconference created by event co-organizer Ceilyn Boyd.


What do we need?

The unconference was intended to provide a space to bring more voices into conversations about computational methods in archives and, more specifically, to connect those currently engaged in CAS with other library and archives practitioners. At the end of the day, we worked together to compile a list of things that we felt many of us would need to learn in order to engage with CAS.

These needs include lists of methodologies and existing tools, canonical data and/or open datasets to use in testing such tools, a robust community of practice, postmortem analysis of current/existing projects, and much more. Building a community of practice and skill development for folks without strong programming skills was identified as both particularly important and especially challenging.

Be sure to check out some of the lightning round slides and community notes to learn more about CAS as a field as well as specific projects!

Interested in connecting with the CAS community? Join the CAS Google Group at:!

The Harvard CAS unconference was planned and administered by Ceilyn Boyd, Jane Kelly, and Jessica Farrell of Harvard Library, with help from Richard Marciano and Bill Underwood from the Digital Curation Innovation Center (DCIC) at the University of Maryland’s iSchool. Many thanks to all the organizers, presenters, and participants!

Jane Kelly is the Historical & Special Collections Assistant at the Harvard Law School Library. She will complete her MSLIS from the iSchool at the University of Illinois, Urbana-Champaign in December 2018.

Improving Descriptive Practices for Born-Digital Material in an Archival Context

by Annie Tummino

In 2014/15 I worked at the New York Metropolitan Library Council (METRO) as the Project Manager for the National Digital Stewardship Residency (NDSR) program in New York City, providing administrative support for five fantastic digital stewardship projects. While I gained a great deal of theoretical knowledge during that time, my hands-on work with born digital materials has been fairly minimal. When I saw that METRO was offering a workshop on “Improving Descriptive Practices for Born-Digital Material in An Archival Context” with Shira Peltzman, former NDSR-NY resident (and currently Digital Archivist for UCLA Library Special Collections), I jumped at the opportunity to sign-up.

For the last two years I have served as the archivist at SUNY Maritime College, working as a “lone arranger” on a small library staff. My emphasis has been on modernizing the technical infrastructure for our special collections and archives. We’ve implemented ArchivesSpace for collections management and are in the process of launching a digital collections platform. However, most of my work with born-digital materials has been of a very controlled type; for example, oral histories and student photographs that we’ve collected as part of documentation projects. While we don’t routinely accession born-digital materials, I know the reckoning will occur eventually. A workshop on descriptive practices seemed like a good place to start.

Shira emphasized that a great deal of technical and administrative metadata is captured during processing of born-digital materials, but not all of this information should be recorded in finding aids. Part of the archivist’s job is figuring out which data is meaningful for researchers and important for access. Quoting Bertram Lyons, she also emphasized that possessing a basic understanding of the underlying “chemistry” of digital files will help archivists become better stewards of born-digital materials. To that end, she started the day with a “digital deep dive” outlining some of this underlying digital chemistry, including bits and bytes, character encoding, and plain text versus binary data. This was followed by an activity where we worked in small groups to analyze the interdependencies involved in finding, retrieving, and rendering the contents of files in given scenarios. The activity definitely succeeded in demonstrating the complexities involved in processing digital media, and provided an important foundation for our subsequent discussion of descriptive practice.

The following bullets, pulled directly from Shira’s presentation, succinctly summarize the unique issues archivists face when processing born digital materials:

  • Processing digital material often requires us to (literally) transform the files we’re working with
  • As stewards of this material, we must be prepared to record, account for, and explain these changes to researchers
  • Having guidelines that helps us do this consistently and uniformly is essential

The need for transparency and standardization were themes that came up again and again throughout the day.

To deal with some of the special challenges inherent in describing born-digital materials, a working group under the aegis of the UC Born-Digital Content Common Knowledge Group (CKG) has developed a UC-wide descriptive standard for born-digital archival material. The elements map to existing descriptive standards (DACS, EAD, and MARC) while offering additional guidance for born-digital materials where gaps exist. The most up-to-date version is on GitHub, where users can make pull requests to specific descriptive elements of the guidelines if they’d like to propose revisions. They have also been deposited in eScholarship, the institutional repository for the UC community.

Working in small groups, workshop participants took a closer look at the UC guidelines, examining particular elements, such as Processing Information; Scope and Content; Physical Description; and Extent. Drawing from our experience, we investigated differences and similarities in using these elements for born-digital materials in comparison to traditional materials. We also discussed the potential impacts of skipping these elements on the research process. We agreed that lack of standardization and transparency sows confusion, as researchers often don’t understand how born-digital media can be accessed, how much of it there is, or how it relates to the collection as a whole.

For our final activity, each group reviewed a published finding aid and identified at least five ways that the description of born-digital materials could be improved in the finding aid. The collections represented were all hybrid, containing born-digital materials as well as papers and other analog formats. It was common for the digital materials to be under-described, with unclear access statuses and procedures. The UC guidelines were extremely helpful in terms of generating ideas for improvements. However, the exercise also led to real talk about resource limitations and implementation. How do born-digital materials fit into an MPLP context? What do the guidelines mean for description in terms of tiered or efficient processing? No solid answers here, but great food for thought.

On the whole, the workshop was a great mix of presentation, discussion, and activities. I left with some immediate ideas to apply in my own institution. I hope more archivists will have opportunities to take workshops like this one and will check out the UC Guidelines.


Tummino Picture

Annie Tummino is the Archivist & Scholarly Communications Librarian at SUNY Maritime College, where she immerses herself in maritime special collections and advocates for Open Access while working in a historic fort by the sea. She received her Masters in Library and Information Studies and Archives Certificate from Queens College-CUNY in December, 2010.

Embedded Archives at the Institute for Social Research

by Kelly Chatain

This is the fourth post in the BloggERS Embedded Series.

As any archivist will tell you, the closer you can work with creators of digital content, the better. I work for the Institute for Social Research (ISR) at the University of Michigan. To be more precise, I am a part of the Survey Research Center (SRC), one of five centers that comprise the Institute and the largest academic social science research center in the United States. But really, I was hired by the Survey Research Operations (SRO) group, the operational arm of SRC, that conducts surveys all over the world collecting vast and varied amounts of data. In short, I am very close to the content creators. They move fast, they produce an extraordinary amount of content, and they needed help.

Being an ‘embedded’ archivist in this context is not just about the end of the line; it’s about understanding and supporting the entire lifecycle. It’s archives, records management, knowledge management, and more, all rolled into one big job description. I’m a functional group of one interacting with every other functional group within SRO to help manage research records in an increasingly fragmented and prolific digital world. I help to build good practices, relationships, and infrastructure among ourselves and other institutions working towards common scientific goals.

Lofty. Let’s break it down a bit.

Find it, back it up, secure it

When I arrived in 2012, SRO had a physical archive of master study files that had been tended to by survey research staff over the years. These records provide important reference points for sampling and contacting respondents, designing questionnaires, training interviewers, monitoring data collection activities, coding data, and more. After the advent of the digital age, a few building moves, and some server upgrades, they also had an extensive shared drive network and an untold number of removable media containing the history of more recent SRO work. My first task was to centralize the older network files, locate and back up the removable media, and make sure sensitive data was out of reach. Treesize Professional is a great tool for this type of work because it creates detailed reports and clear visualizations of disk space usage. This process also produced SRO’s first retention schedule and an updated collection policy for the archive.

Charts produced by Treesize Professional used for the initial records survey and collection.
A small selection of removable media containing earlier digital content.

Welcome, GSuite

Despite its academic home, SRO operates more like a business. It serves University of Michigan researchers as well as external researchers (national and international), meeting the unique requirements for increasingly complex studies. It maintains a national field staff of interviewers as well as a centralized telephone call center. The University of Michigan moved to Google Apps for Education (now GSuite) shortly after I arrived, which brought new challenges, particularly in security and organization. GSuite is not the only documentation environment in which SRO operates, but training in the Googleverse coincided nicely with establishing guidance on best practices for email, file management, and organization in general. For instance, we try to label important emails by project (increasingly decisions are documented only in email) which can then be archived with the other documentation at the end of the study (IMAP to Thunderbird and export to pdf; or Google export to .mbox, then into Thunderbird). Google Drive files are downloaded to our main projects file server in .zip format at the end of the study.

Metadata, metadata, metadata

A marvelous video on YouTube perfectly captures the struggle of data sharing and reuse when documentation isn’t available. The survey data that SRO collects is delivered to the principal investigator, but SRO also collects and requires documentation for data about the survey process to use for our own analysis and design purposes. Think study-level descriptions, methodologies, statistics, and more. I’m still working on finding that delicate balance of collecting enough metadata to facilitate discovery and understanding while not putting undue burden on study staff. The answer (in progress) is a SQL database that will extract targeted structured data from as many of our administrative and survey systems as possible, which can then be augmented with manually entered descriptive metadata as needed. In addition, I’m looking to the Data Documentation Initiative, a robust metadata standard for documenting a wide variety of data types and formats, to promote sharing and reuse in the future.

DDI is an international standard for describing data.

Preserve it

The original plan for digital preservation was to implement and maintain our own repository using an existing open-source or proprietary system. Then I found my new family in the International Association for Social Science Information Services & Technology (IASSIST) and realized I don’t have to do this alone. In fact, just across the hall from SRO is the Inter-University Consortium for Political and Social Research (ICPSR), who recently launched a new platform called Archonnex for their data archive(s). Out of the box, Archonnex already delivers much of the basic functionality SRO is looking for, including support for the ever-evolving  preservation needs of digital content, but it can also be customized to serve the particular needs of a university, journal, research center, or individual department like SRO.

Searching for data in OpenICPSR, built on the new Archonnex platform.


The embedded archivist incorporates a big picture perspective with the specific daily challenges of managing records in ways that not many positions allow. And you never know what you might be working on next…

Kelly Chatain is Associate Archivist at the Institute for Social Research, University of Michigan in Ann Arbor. She holds an MS from Pratt Institute in Brooklyn, NY.

Thoughts from a Newcomer: Code4Lib 2018 Recap

by Erica Titkemeyer

After prioritizing audiovisual preservation conferences for so long, this year I chose to attend my first Code4Lib annual conference in Washington, D.C. I looked forward to the talks most directly related to my work, but also knew that there would be many relatable discussions to participate in, as we are all users/creators of library technology tools and we all seek solutions to similar data management challenges.

The conference started with Chris Bourg’s keynote, detailing research on why marginalized individuals are compelled to leave their positions in tech jobs because of undue discrimination. Calling for increased diversity in the workplace, less mansplaining/whitesplaining, more vouching for and amplification of marginalized colleagues, Dr. Bourg set out to “equip the choir”. She also pointed out that Junot Diaz said it best at ALA’s midwinter conference, when he called for a reckoning, explaining that “A profession that is 88% white means 5000% agony for people of color, no matter how liberal and enlightened you think you are.”

I appreciated her decision to use this opportunity to point out our own shortcomings, contradictions and need to do better. I will also say, if you ever needed proof that there is an equity problem in the tech world, you can:

  1. Listen to Dr. Bourg’s talk and
  2. Read the online trolling and harassment that she has since been subjected to because of it.

Since the backlash, Code4Lib has released a Community Statement in support of her remarks.

Following the keynote, the first round of talks further assured me that I had chosen the right conference to attend. In Andreas Orphanides’ talk: “Systems thinking: a practical field guide”, he cleverly pointed out system failures and hacks we all experience in our daily lives, and how they are analogous to the software we might build and where there might be areas for improvement. I also appreciated Julie Swierczek’s talk “For Beginners – No Experience Necessary”, in which she made the case for improving how we teach to true beginners in workshops. She also argued that instructors should not assume everyone is on the same level-playing field just because the title includes “for beginners” as it is not likely that attendees will know how to self-select workshops, especially if they are truly beginners to the technology being taught.

As a fan of Arduino (an open source hardware and software platform that supports DIY electronic prototying), I was curious to hear Monica Maceli’s “Low-cost preservation Environment Monitoring” talk, where she described her experience developing an environmental datalogger using the Raspberry Pi (similar microcontroller concept to Arduino) comparing the results and associated costs with a commercial datalogger, the PEM2. While it would require staff with appropriate expertise, it seemed to be a worthwhile endeavor for anyone wishing to spend a quarter of the price.

With the sunsetting of Flash, I was eager to hear how Jacob Zaborowski’s talk “Save Homestar Runner!: Preserving Flash on the Web” would address the preservation concerns surrounding Homestar Runner, an online cartoon series that began in 2000 using flash animation. Knowing that tools such as Webrecorder and Archive-it would capture, but not aid in preserving the SWF files comprising the animations, Zaborowski sought out free and/or open source tools for transcoding the files to a more accessible and preservation-minded format. Like many audiovisual-based digital formats, tools for transcoding the SWF files were not entirely reliable or capable of migrating all of the unique attributes to a new container with different encodings. At the time of his talk, the folks at Homestar Runner were in the midst of a site redesign to hopefully resolve some of these issues.

While I don’t have the space to summarize all of the talks I found relatable or enlightening during my time at Code4Lib, I think these few that I’ve mentioned show how varied the topics can be, while still managing to complement the information management work we are all charged with doing.

TitkemeyerErica.jpgErica Titkemeyer is the Audiovisual Conservator for the Southern Folklife Collection (SFC) at the University of North Carolina at Chapel Hill. She is the Project Director on the Andrew W. Mellon Foundation grant Extending the Reach of Southern Audiovisual Sources, and overseas the digitization, preservation and access of audiovisual recordings for the SFC.

Embedded Memory

by Michelle Ganz

This is the third post in the BloggERS Embedded Series.

As an archivist in a small corporate repository for an architecture, design, and consulting firm I have a unique set of challenges and advantages. By being embedded with the creators as they are creating, I have the opportunity to ensure that archival standards are applied at the point of creation rather than after the collection has been transferred to a repository.

My environment is especially unique in the types of digital files I’m collecting. From the architecture side, William McDonough + Partners, I acquire architectural project files (CAD files), sketches, renderings, photographs and video of project phases, presentations, press, media files surrounding the project, and other associated materials. Many of these files are created on specialized, and expensive, software.

From McDonough Innovation, William McDonough’s sustainable development advising firm, I collect the world-changing ideas that Mr. McDonough is developing as they evolve. Mr. McDonough is a global thought leader who provides targeted ideas, product concepts, and solutions to a wide range of sustainable growth issues faced by corporate officers, senior executives, product designers, and project managers. He often works with CEOs to set the vision, and then with the management team to set goals and execute projects. His values-driven approach helps companies to embed sustainable growth principles into their corporate culture and to advance progress toward their positive vision. Archiving an idea is a multi-faceted endeavor. Materials can take the form of audio notes, sketches in a variety of programs, call or meeting recordings, and physical whiteboards. Since my role is embedded within the heart of Mr. McDonough’s enterprises, I ensure that all the right events are captured the right way, as they happen. I gather all the important contextual information and metadata about the event and the file. I can obtain permissions at the point of creation and coordinate directly with the people doing the original capture to ensure I get archival quality files.

Challenges in this environment are very different than what my academic counterparts face. In the academic world there was a chain of leadership that I could advocate to as needed. In my small corporate world there is no one to appeal to once my boss makes up their mind. Corporate interests are all focused on ROI (return on investment), and an archival department is a financial black hole; money is invested, but they will never see a financial return. This means that every new project must show ROI in more creative ways. I focus on how a project will free up other people to do more specialized tasks. But even this is often not enough, and I find myself advocating for things like file standards and server space. Many of the archival records are videos of speeches, events, meetings, or other activities and take up a huge amount of server space. A single month’s worth of archival files can be as large as 169 GB. In the academic setting where the archives is often a part of the library, the IT department is more prepared for the huge amounts of data that come with modern libraries; folding the archival storage needs into this existing digital preservation framework is often just a matter of resource allocation or funds.

Also, nearly every archival function that interacts with outside entities requires permissions these firms are not used to giving. Meetings can include people from 3 or 4 companies in 4 or 5 countries with a variety of NDAs in place with some, but not all, of the parties. In order to record a meeting I must obtain permission from every participant; this can be rather complicated and can create a lot of legal and privacy issues. A procedure was put in place to request permission to record when meetings are set up, as well as when meetings are confirmed. A spreadsheet was created to track all of the responses. For regular meeting participants annual permissions are obtained. This procedure, while effective, is time-consuming. For many meeting participants they are unfamiliar with what an archive is. There are many questions about how the information will be used, stored, disseminated, and accessed.  There are also a lot of questions around the final destination of the archive and what that means for their permissions. To help answer these questions I created fact sheets that explain what the archives are, how archival records are collected and used, deposit timelines, copyright basics, and links to more information. To further reassure participants, we give them the option of asking for a meeting to be deleted after the fact.

This is the server stack for the archives and the two firms. The archive uses 2 blades.
This hub connects the archive to workstations and handles the transfer of TBs of data.

Preservation and access are unique challenges, especially with the architecture files. Many of the project-related files are non-traditional file formats like .dwg, .skb, .indd, .bak, et al., and are created in programs like AutoCAD and SketchUp Pro. I work with the IT department to ensure that the proper backups are completed. We back up to a local server as well as one in the city, but offsite, and a third dark archive in California. I also perform regular checks to confirm the files can open. Due to the fact that projects are often reopened years later, it is impractical to convert the files to a more standardized format. To ensure some level of access without specialized terminals, final elements of the project are saved in a .pdf format. This includes final drawings/renderings and presentations.

Furthermore, I often find myself in the awkward position of arguing with records creators in favor of keeping files that they don’t want but I know have archival value. Without the benefit of patrons, and their potential needs, I am arguing for the future needs of the very people I am arguing with! Without a higher level of administration to appeal to, I am often left with no recourse but to do things that are not in the best interests of the collection. This leads to the unfortunate loss of materials but may not be as bad as it first appears. When considering how traditional archival collections are created and deposited, it is well within reason that these items would never have made it into the collection. I like to think that by being embedded in the creation process, I am able to save far more than would otherwise be deposited if the creators were left to make appraisal decisions on their own.

Michelle Ganz is the Archives Director at McDonough Innovation and the former archivist at Lincoln Memorial University. She received her MILS from the University of Arizona and a BA from the Ohio State University. In addition to her passion for all things archival Michelle loves to cook, read, and watch movies.

Modernization of the A.D. Hopkins collection at the Smithsonian Institution Department of Entomology

by Teresa Boyd

This is the second post in the BloggERS Embedded Series.

The Smithsonian Institution’s Department of Entomology has recently finished phase one of their multiyear project to digitize their portion of the A.D. Hopkins notes and records system, which includes about 100 years of observations, both in the field and in the lab. A.D. Hopkins developed the system in order to collect biological and natural history notes about species, the environment they were in, as well as the collectors and locations of collection. This collection was adopted by the United States Department of Agriculture (USDA) when Hopkins was named Chief of Forest Insect Investigations, though Hopkins is known to have developed and used the system while working at West Virginia University in the late 1800s. The Smithsonian Institution’s Department of Entomology has historically worked very closely with the USDA and therefore obtained the largest portion of the Hopkins card file over the years.


It was important to Hopkins to collect as much information as possible about specimens because he felt it was the quickest way to understand the situation of potential pests and to find solutions to harmfully invasive species. Part of Hopkins’ methodology was to encourage average citizens to send in specimens and observations to the USDA, the Smithsonian, or one of the forest experiment stations that were located throughout the United States, which were then incorporated into the Hopkins note system. Some of these notes are also documentation about lab research such as specimen rearing, specimen transfers, and communications between lab and field. A few of these notes are also cross-referenced, so often a lab note can be traced back to a field note, making it easier for researchers to quickly see the correlation between field and lab (work that was often done by different individuals.) The numbers on each individual card within the A.D. Hopkins system correlates to specimens that are housed in various locations. Traditionally a researcher or scientist would ask for the notes that were associated with a a specimen number. By creating an online repository of the notes, the Smithsonian hopes to further enrich researchers with new tools to expand their work and perhaps find new ways to use the data which has been collected by past researchers and scientists.

I have been working on this project as a lone archivist for the past 5 years, scanning the card file portion of the collection, and am now working on preparing these scans for a website that will be built specifically for this type of collection. The Smithsonian Institution’s Department of Entomology hopes to begin sending the scans of the cards to the Smithsonian Transcription Center soon to crowdsource the transcriptions. This cuts down on the time it takes to transcribe the older material which is all handwritten. I will be adding the transcribed notes to the digitized card on the website so that researchers will be able to go to the website, look up a specific card, and see both the original scan and the transcribed notes, making it easy for anyone to be able to use the information contained in the Hopkins collection. Additionally these scans will be incorporated into the Department of Entomology’s collections database by matching specimens to their unique card numbers;  thereby giving researchers the complete picture.

The Smithsonian Institution’s work to digitize and make their A.D. Hopkins collection publicly available is not the first of its kind; the USDA had previously accomplished this in the 1980s, and has made their documents available on the USDA website, HUSSI. There is hope that in the future other institutions that have their own portions of the A.D. Hopkins notes and records system will also begin to digitize and make them available online, supplementing the Smithsonian and USDA efforts to make this invaluable data available to researchers. 

Teresa Boyd is an archivist for the Department of State and a volunteer archivist for the Smithsonian Institute’s Department of Entomology. She holds a degree in Library and Information Science from the University of Arizona.

Capturing Common Ground

by Leslie Matthaei

This is the first post in the BloggERS Embedded Series.

Every Tuesday I am asked the same question: “T Coast today?” T Coast, or Tortilla Coast, is the preferred lunch location for some of the photographers that occupy the Photography Branch, in the Curator Division, of the federal agency Architect of the Capitol. The agency oversees the maintenance of building and landscapes on Capitol Hill to include the Library of Congress buildings, Supreme Court, United States Botanic Gardens, House and Senate Office Buildings, Capitol Power Plant, and, of course, the United States Capitol. I have joined the professional photographers at T Coast for more than a dozen lunches now. I am here for the Taco Salad and comradery but mostly I am here to listen. And to ask questions. I am an embedded archivist.

The Library of Congress Thomas Jefferson Building
United States Capital Building
United States Botanic Garden – Bartholdi Park

I use my time at the T Coast lunch table to get to know the photographers and for them to get to know me. I discovered very quickly that the photographers and I have a lot in common. For example, the photographers are often assigned to shoot a long-term project (Collection) which may have multiple phases (Series), and for each phase, they go out on specific days to shoot (File Units). They cull excess and/or duplicate photographs. And they generally have a tried workflow for ingesting their born-digital objects to edit in Adobe Lightroom then upload them to an in-house Digital Asset Management system known as PhotoLightbox. Within PhotoLightbox, they are responsible for defining the security status of an individual image or group of images and providing the descriptive metadata. Tapping into parallel duties has allowed me to bridge potential knowledge gaps in explaining what roles and functions I can provide the branch as a whole.

One rather large knowledge gap is descriptive metadata. To be sure, the photographers in our agency are incredibly busy and in high demand. And they are professionally trained photographers. They see the world through aesthetics. It is not necessarily their job to use PhotoLightbox to help a researcher find images of the East Front extension that occurred in the 1950s, for example. That is my role, and when I query PhotoLightbox, the East Front extension project is represented in multiple ways: EFX, East Front (Plaza) Extension, East Extension, Capitol East Front Extension. You may see where this is going: there is no controlled vocabulary. When, in a staff meeting, I pitched the idea of utilizing controlled vocabularies, they immediately understood the need. Following their lead, the conversation turned to having me develop a data entry template for each of their shoots.

An example of the data entry template.

I admit now that my first spin through PhotoLightbox revealed a pressing need for controlled vocabularies, among other concerns the database presented. I am the type of person that when I see a problem, I want to fix it immediately. Yet I knew that if my first professional introduction to the photographers was a critique of how unworkable their data entry was and had been over time, I might turn them off immediately. Instead, I went to lunch. I credit the results produced by this particular staff meeting to the time that I put in getting to know the photographers, getting to understand each of their respective workflows, and understanding a little bit about the historic function and purpose of the office within the agency.

I have another half dozen lunches to go before I begin to talk to the photographers about the need for digital preservation of born-digital images over the long-term and both of our roles in the surrounding concepts and responsibilities. I have a few more lunches after that to get their assistance in codifying the decision we are making together into branch policies. I feel confident, however, that I have their complete buy-in for the work that I have been tasked to do in the branch. Instead of seeing me as another staff member making them do something they do not want, I am seen as someone who can help them gain control of and manage their assets in a way that has yet to be done in the branch. I cannot do it alone, I need their help. And some chips and salsa every once in a while.

Leslie Matthaei

Leslie Matthaei is an Archivist in the Photography Branch, Curator Division, Architect of the Capitol. She holds an MLIS from the University of Arizona, and an MA and BA in Media Arts from the University of Arizona.