Eight UT graduate students were selected to participate in the program cohort. The students–in MA and PhD programs–are studying in the African & African Diaspora Studies, English, History, and Middle Eastern Studies departments, as well as in the UT iSchool. They have a variety of experiences with research in libraries and archives, with digitization, and with publishing scholarship, all of which they bring to their cohort discussions. However, they are united to realize the goals of this symposium program, which include reading about, discussing, and creating approaches for research and collection development in a digitizable environment. The latter can be described with the question: what does it mean to create or select print and electronic content in an environment in which digitization is possible and high quality; in which there is support for the applications of machine-readable text; and in which the materials are stewarded by libraries and used by researchers outside of the materials’ region of origin?
Cohort participants are encouraged to engage with existing writing (scholarly and popular) on these topics in thoughtful and critical ways, with the end goal being to create a sense of belonging to the conversation. What gets digitized and how it gets digitized are decisions that affect everyone, but most of all, marginalized communities that have been historically disadvantaged from participation in scholarship and the building of library collections (even, and especially, collections for which they are the subject). As part of this program, cohort participants are trained in the basics of scanning, OCR, and outputs/applications with a material selection of their choice, so that they have insight into the hands-on processes of digitization and how to use this technology for their goals. The program’s culminating public symposium puts the cohort’s theoretical and practical experiences in conversation with a digital cultural heritage scholar and engagement with the audience in order to realize new approaches to digitized resources.
I developed The Theory & Practice of Digitization Community Symposium Program as the final project for my Mellon Fellowship for Diversity, Inclusion & Cultural Heritage at the Rare Book School. As fellows, we are asked to put together a community symposium at our home institution that advances understanding of cultural heritage, archives, and/or special collections and allows us to promote aspects of our collections to broader publics and communities. With the development of the new Scholars Lab at the Perry-Castañeda Library, and considering my own interests in reparative and restorative practices in librarianship and scholarship, I wanted to create an opportunity for graduate students to expand their researcher skill-sets and build reflective approaches to their future professions. We are incredibly fortunate to have a wide range and depth of expertise at the UT Libraries, and it is from this well of experience and insight that this program has drawn.
Our first session, held at the end of August shortly after the semester began, featured a conversation with Rachel E. Winston (Black Diaspora Archivist at the Benson Latin American Collection) and Beth Dodd (Curator at the Alexander Architectural Archives) on defining terms for our work in this program through their experience with digitization as archivists at UT. Rachel and Beth presented on the process of selecting and adding items to the archives, including when, how, and why they make decisions around digitization. Their experiences with a variety of collections––from donors or vendors; recent or older; created in the U.S. or around the world––gave them insight to respond to students’ questions regarding the ethics of archival digitization and stimulated the students’ engagement with crucial concepts by providing real and tried examples for them to consider.
The program’s second session introduced students to the basic principles of handling cultural heritage materials and digitizing them. My colleagues from the UT Libraries’ Stewardship department, Brittany Centeno (Preservation Librarian) and Kiana Fekette (Head of Digitization) led the students through a review of best practices for handling paper materials such as books, periodicals, and personal archives. The session was held in the new Scan Tech Studio in the Scholars Lab, which functions as a self-service facility for independent researcher digitization, image processing, and text recognition-based scholarship. Brittany and Kiana brought sample materials so that the students could get a sense of what to do for for different preservation situations, such as a book with a broken spine, brittle and flaking paper or leather, bent or misshapen items, and materials that are tightly bound. They also demonstrated how to use a diffuser light set up, which can be particularly useful for items with a difficult-to-capture sheen (such as different types of photographs) or for mobile applications when traveling for research.
In our third session, we met with Allyssa Guzman (Head of Digital Scholarship Services) and Ian Goodale (European Studies Librarian) for a survey of, training with, and discussion of tools that the students might use for their research with digitized materials. Allyssa covered how to get started with digital scholarship, including project planning/management and tool selection. She created an excellent LibGuide for the cohort to refer back to as they move forward with their work. Ian reviewed a number of tools that we recommend and regularly use here at the UT Libraries for transcription/OCR correction and text analysis, including some that he has developed himself.
The cohort’s efforts will culminate in a community symposium on November 9, 2023, 5 – 7 PM in the PCL Scholars Lab Data Lab. This event is free and open to the public: everyone is invited and encouraged to attend. The symposium is an opportunity for the UT, Austin, and greater central Texas communities to learn about the digitization of cultural heritage through the experiences of the student cohort members. It’s also an opportunity to hear from a respected scholar of digital cultural heritage, Dr. Raha Rafii, who will be giving the keynote address. Her lecture, titled, “Navigating the Ethical Landscape of Manuscript Digitization,” will look at recent examples of digitized forms of cultural heritage and the impact on their origin communities in order to think through complex issues of ethics, and to determine the lines between academic researcher priorities and digitization as an extension of colonial and imperialist practices. For more information on the community symposium, please see the UT Libraries’ Events page.
A project to provide digital access to an important collection of geologic cartography from the Walter Geology Library has been completed.
The Deep Time Maps are a collection of paleogeographic maps showing the landscapes and oceans of ancient Earth through hundreds of millions of years of geologic time. These maps are an extraordinary resource for geoscientists, but have been inaccessible to users due to limits on the technology available for allowing access to this large of a collection.
The project to make this resource accessible online through the Libraries’ online presence was an idea that had been sitting around collecting “digital dust” for quite some time due to limits on the technology available for our use.
Senior Content Management Specialist Stacy Ogilvie took lead on the project to provide digital access to views of the Earth’s continents over the course of millions of years through the Libraries’ unified management resource system component Alma Digital. Adding this collection to Alma Digital is a significant step in increasing its accessibility to our users and fulfilling a goal that our late colleague Dennis Trombatore had in purchasing the materials.
“The process also served as our first big test of adding a large collection to Alma Digital and the experience Stacy gained from working on this from scratch will help inform how we work more closely with SRD and add additional large collections to the Alma Digital workflow,” says Head of Content Management Corey Halaychik. “Her work on this front is invaluable to our team.”
Dale J. Correa is the Middle Eastern Studies Librarian and History Coordinator for the UT Libraries, and she regularly teaches on research data/citation management for the humanities at The University of Texas at Austin.
Hannah Chapman Tripp serves as the Biosciences Librarian and has provided research help with a variety of citation management programs at The University of Texas at Austinand previous institutions.
Where Did My Data Go?
In Fall 2020, registered Mendeley users received a message via email titled “Improving Mendeley to Better Support Researchers,” regarding some intended updates to Mendeley’s service model. These changes included the removal of several Mendeley library features, including the Public Groups feature that allowed for large groups to share references and notes openly. These groups were particularly appealing to some scholars as they represented a method to share resources openly, publicly, and free of cost in both invited and open group settings (without a limit on membership to the group). Under the Public Groups umbrella, both the invite-only and the open groups were included in Mendeley’s feature-removal plans. Unfortunately, Mendeley’s email did not explicitly state the intention to delete the Public Groups from individual Mendeley users accounts with the coming update — which went into effect in March 2021, and meant that individual users found their locally-stored files from these groups deleted on their own machines.
Researchers who used this feature were somewhat unlikely to have encountered that email message or have read it through thoroughly. After all, many emails from services utilized by researchers contain information about updates, but much of it goes unread. And, of course, some email systems would automatically detect messages like this one as spam or junk, and so would send them directly to a folder that, unless checked, frequently goes unnoticed and unchecked.
As “announced,” Mendeley went ahead with the plan and began removing certain features, including Mendeley Feed, Mendeley Profiles and Mendeley Funding in December 2020. In March 2021, Mendeley began retiring Public Groups. It does not seem that there was further, specific communication regarding the Public Groups retirement in the lead-up to this change in March.
While we fully acknowledge the need for commercial companies to pivot priorities, continue development of what’s working and in some cases remove features that are less popular and see less return on investment, the awareness campaign for these changes clearly did not reach enough of the affected audience to warrant the deletion of features from an individual user’s Mendeley library. The failure of this important information to reach registered Mendeley users is evidenced by many, many, many reactions on Twitter from the scholarly community. While most scholars understand the need to make changes to a platform and continue to improve the services offered, they are also outraged at the lack of effective communication prior to deleting this feature.
Mendeley has acknowledged that there was not enough time or communication involved in this plan to remove features, and has since re-enabled the invite-only groups, a subset of the Public Groups, for a brief period of time so users can retrieve their data. It is a significant concern of many researchers that all of the content in the Open Groups (which was the other option under the Public Groups umbrella) is not going to be restored and that the data has been lost permanently. For many academics, this is a devastating realization, as years of research and references have been erased with deficient notice. Although Mendeley has apologized for the handling of these changes, the fact remains that some scholars — including those in the more vulnerable categories of PhD student, post-doc and non-tenured faculty — are left without vast quantities of their research.
Lessons Learned, Principles to Practice
While this is an unfortunate situation, we hope that some takeaways can be gained from the experience. For researchers, the importance of backups, knowing your product and an awareness of the fact that changes are quite likely, are a few of the points we hope to address.
Backing up research data is important, regardless of the type of data or original format. A best practice in data retention habits is the 3-2-1 rule, wherein three copies of research data are maintained, in two separate formats locally, and one copy offsite. Some researchers wrongly assumed that with Mendeley’s storage and syncing they were achieving at least a portion of this best practice; however, they learned in practice that when data is deleted from the Mendeley web version, that deletion can be synced down to any local copy of Mendeley connected to the web. In order to have the 3-2-1 rule appropriately in practice with Mendeley data, researchers must back up a copy of their data to an external hard drive location and an online cloud storage solution separate from Mendeley. What makes this situation trickier is that, starting in 2018, Mendeley began encrypting researchers’ local data folders, making it very difficult to access one’s own data when not using the Mendeley interface (although some researchers have identified workarounds to the encryption). What should be backed up, rather, is data exports from Mendeley in open file formats and PDFs, including notes, to ensure that researchers will be able to access, use, and rebuild their reference libraries if their Mendeley data itself becomes corrupt or a change in Mendeley services affects their access.
With RIS (Research Information Systems bibliographic citation file format) files and PDFs backed up to the local machine as well as to a back up option like UT’s Box, researchers would have the option to continue using Mendeley, or move their data to another citation management software such as Zotero or EndNote. For those who are continuing to use Mendeley, incorporating a backup system as described above is the recommended option for ensuring long term access to integral research references, notes, and files (particularly annotated PDFs).
Mendeley — owned by a for-profit company — will continue to optimize the most attractive, state-of-the-art, and revenue-generating features and functionality in their product. This process inevitably means refocusing efforts and making tough decisions about what features to no longer support. However, the realities of software changes and obsolescence are not confined to Mendeley or, for that matter, to for-profit companies. For example, the backups you made decades ago to a floppy disk are likely no longer retrievable due to hardware changes and potential software obsolescence.
So, whether you have lost your data with this change in Mendeley services or you are one of the lucky ones who was not relying so heavily on the free Public Groups features, we strongly recommend that you use a sensible back up system; back up in open formats from which you can easily retrieve your data no matter what system you’re using; and keep an eye on the crucial changes that come with software updates. We are here to assist with data and citation management best practices — please see the Research Organization with Citation Managers LibGuide for more information.
If there’s a single lesson to take away from this year, it’s that libraries are a lot more malleable than their long history may have given them credit for.
We’ve previously covered the Herculean effort by University of Texas Libraries’ staff to pivot from their natural in-person work environs to a distance service, then a subsequent limited return to the former, but a lot of that agility was due in no small measure to underlying efforts that were already underway when the health crisis washed over campus and the country.
Strategically, this institution has been focusing on the idea of the library as a platform: not just a storehouse for books or website of searchable journals, but an active ecosystem where resources, tools, services, spaces, expertise and community intermingle with a constantly variable presence of users to spin off scholarship and innovation back into the world. This idea factory of ever-evolving components works at its best when it creates opportunities for discovery through constant interaction of the various parts.
With the pandemic creating greater physical distance between the parts, though, it’s become essential that we focus on those tools that could best allow us to reach our users where they are, be that in an apartment in West Campus, or on the other side of the globe.
Last year, we announced the launch of a pair of systems designed to organize, preserve and create accessibility for digital iterations of physical materials that otherwise would only be available to people who could visit the Forty Acres. Our Digital Asset Management System (DAMS) was deployed in September, 2019, and in November, we published the Collections Portal on the Libraries’ website. The culmination of these two projects proved to be far more fortuitous than we could’ve imagined.
A couple months later as leadership at the Libraries was fleshing out a new strategic plan that placed special emphasis on the concept of Libraries as platform, the first case of coronavirus was discovered in the Pacific Northwest. Then, in March as the spread of the pandemic began to accelerate, The University of Texas at Austin announced first the delay of spring classes, followed quickly by a directive to move all but the most critical staff to remote work away from campus, and to shift to online learning for the remainder of the semester.
More than ever, the adaptability of the Libraries to changes in user behaviors was the institutional characteristic that needed to be positioned in response to the extraordinary situation that fell so quickly upon us all. And refocusing our collective energies on tools with the greatest potential to serve the largest number of people while considering the long-term goals of the Libraries made these new systems a natural priority for applying institutional resources.
At its most basic, a Digital Asset Management System is a locally-developed digital repository designed to store, describe and manage digital assets of the Libraries. Digital assets are comprised of a primary digital files like scanned images, book pages, audio or video recordings, with varying component parts: metadata, or data about the data that includes information about the origin of the file, specifications and descriptive data used for locating the asset; additional secondary files that can be machine-readable and/or provide additional technical information; and derivatives, such as thumbnail images, other file versions, and PDFs.
The DAMS serves as the central preservation and management hub for Libraries’ digital assets, built by the Libraries Information Technology Support (LITS) team in coordination with staff library professionals, who also manage the operations of the system. The DAMS project began in 2016, and in an effort to prioritize two of our most notable collections, staff at the Benson Latin American Collection and the Alexander Architectural Archive began preparing digital collections for the system.
“The digital asset management system was many years in the making,” says Jennifer Lee, Director of Discovery and Access. “And for many, many years before that it was just an idea, like an item on a collective wish list. Now, it’s become a reality. And over the past seven months in particular, we’ve made excellent progress on adding content.”
The Collections Portal
The Collections Portal serves as an access point on the Libraries’ website allowing users to undertake remote research and study utilizing rich resources that have previously only been available in person or through more time-intensive digitization on demand processes.
Developed in 2018-19 by LITS in close coordination with other Libraries professional staff as a logical progression from the DAMS, the Portal provides students, faculty, researchers and the broader public access to collections that have not been directly available in the past, and the project’s infrastructure creates a framework for a more consistent stream of new digital content in the future. Each item in the portal also contains contextual data – drawn from the DAMS – in order that users may learn underlying information about the material, locate physical counterparts and determine reuse rights for digital files.
The relationship between the DAMS and the Portal can create confusion since both systems deal with the same assets, but it’s useful to think about the interrelationship between the parts. The DAMS is the back-end storage and management environment, where preservation, description and accessibility of the resources are controlled. The Collections Portal draws on the information contained within the DAMS to make some of the content that exists there discoverable and accessible for remote use through a public web interface. The dual structure allows for our staff to determine what is suitable for partial or full public access based on issues like copyright or embargo status.
“These two are separate but closely connected software systems,” explains Mirko Hanke, Digital Asset Management System Coordinator, who has been one of the driving forces behind efforts to refine and build out the systems. “This overall architecture of having two separate systems allows the curators to choose which of the content they’re managing in the DAMS they want to make publicly available.”
Both systems were implemented by LITS staff using open source software components and they built software to bridge the two systems from scratch.
The basic workflow for getting items from the shelves into the systems involves digitization, file management, metadata creation and ingestion.
The Libraries has been digitizing physical materials for decades, including thousands of items that were digitized previous to the development of the DAMS, and those files can be retrieved and processed for inclusion in the new systems. Accessing the digital forms of materials can extend the life of fragile special collections and makes near-immediate global access possible. Physical materials are often reformatted as digital files in their entirety to minimize handling and ensure future access to unrequested sections at a later date. Additional processes in digitization allow for the enhancement of usability of the digital iterations, as well, including optical character recognition, making scanned documents searchable and information contained within more easily findable. The automation of many digitization processes makes pagination and file structuring more manageable and speeds up ingestion and thus accessibility of content.
Requests for digitization are made either through a formal submission or directly to Libraries’ Digitization Services, with special priority given to our two notable special collections – the Benson Latin American Collection and the Alexander Architectural Archive – both of which are heavily used by the public and thus have significant back catalogs of digitized materials, making them fertile resources for populating the DAMS and Collections Portal. Special consideration has also been extended to time-sensitive projects, such as those slated for exhibition loan or items that are being or have been retired from other access points.
Once files have been digitized, they are passed through specialized workflows based on the type of content and its historical origin that add and/or enhance metadata, secondary files and derivatives to create singular digital assets that can then be ingested into the DAMS and potentially projected out to the Collections Portal.
Staff professionals working with LITS professionals have developed scripts and processes that can help to speed up the packaging of digital assets both for newly digitized items, but also from previously digitized materials that exist from earlier Libraries efforts. There is ongoing work to track digitization, management and ingestion processes to create ongoing improvements to the workflows.
Hitting the Gas
Realizing the important potential of the two systems for remote users in response to the health crisis, the Libraries reconfigured workflows and redirected staff to accelerate work already occurring to populate and invigorate the DAMS and by extension, the Collections Portal. The first order of business was to formalize workflows to prioritize the digitization and processing of materials.
Resources at the Benson and Alexander Archive proved to be low-hanging fruit for their outsized use in research and because of existing expertise in digital preservation, so projects originating from those collections received significant attention.
Staff at the Benson Latin American Collection have been working on a project to digitize the Genaro García Collection – the Benson’s massive foundational collection, acquired in Mexico City in 1921 by university representatives on a diplomatic visit. The Libraries will next year be celebrating the 100th anniversary of that acquisition as the establishment of Latin American collections on campus, so the effort to provide online access to this important collection made it a priority for addition to the Collections Portal.
“Because we’ve established some good local practices for collection creation and we have a set of well documented requirements on the DAMS ingest side, it becomes much easier to develop batch processing workflows to prepare scans and metadata for upload into the DAMS without manipulating each collection object, one at a time,” says David Bliss, Digital Processing Archivist at the Benson Latin American Collection.
A team-based approach was coordinated by Latin American Archivist Dylan Joy. Staff Photographer and Library Specialist Robert Esparza spent several months carefully digitizing the Genaro García Imprints and Images collections in their entirety, following a process developed locally at the Benson. Concurrently, GRA Diego Godoy compiled item level metadata based on a template developed by Metadata Librarian Itza Carbajal. Bliss then worked to develop a script for ingesting the scans and accompanying metadata from the collection into the DAMS, bypassing hours of monotonous and error-prone work in favor of a process using existing metadata in a hands-off approach that occurs in minutes instead.
“We didn’t just wake up one day and decide to make our file naming practices more consistent and systematic or suddenly realize that we should be gathering good metadata,” says Bliss. “This kind of scripting work is only possible because significant resources were dedicated to equipment and project staff.”
Benson staff, in coordination with Libraries’ Content Management and Digitization Services teams, have worked prodigiously on the Benson Rare Book Collection, including the high visibility Primeros Libros – the first books published in the Americas prior to 1600; so far, 21 full volumes are published to the Collections Portal, with more in process. Libraries Technology Coordinator Benn Chang worked with Benson Latinx Studies Archivist Carla Alvarez to make newly available several hundred previously digitally-preserved photographs in the George I. Sánchez papers, which are now part of the Collections Portal, as well.
“This work really does take a village and there is no one singular workflow or approach that suits all collections,” says Benson’s Head of Digital Initiatives Theresa Polk.
At the Alexander Architectural Archive, staff have been working to process both newly-digitized and legacy digital assets. “Architectural collections staff have worked closely with Digitization Services to adjust our workflow to include ingesting assets and metadata into the DAMS,” says Archivist for Access and Preservation Stephanie Tiedeken. So far, over 21,000 assets have been ingested into the DAMS from the Alexander Archives and Architecture & Planning Library’s Special Collections, and over 2,000 of those have been published into the Collections Portal, including 270 publications and over 1,800 digitized drawings or photographs.
Archive staff are also working to move legacy assets into the DAMS. The Alexander’s GRA, Alyssa Anderson, recently completed a project to ingest 262 legacy images of scanned drawings and photographs from ten sites, primarily missions, in Texas and Mexico images and create MODS metadata. Now that these items are in the DAMS, they are more usable and visible to researchers.
Head of Architectural Collections Katie Pierce Meyer worked with Mirko Hanke and staff from Digitization Services to develop a process for ingesting legacy digitized photographs from the David Reichard Williams collection, a regionalist and architect who documented vernacular architecture in Texas in the 1920s and 1930s. Colleagues from Libraries’ Branch and Borrow Services transferred data from finding aid, added descriptions of photographs, bringing expertise and fresh eyes to these historic images of buildings and places across the state.
Building on transformation processes and documentation work previously done by David Bliss and Benn Chang, and working closely with Mirko Hanke, Pierce Meyer was able to take the data, map it to DAMS metadata fields in the data editing tool OpenRefine, then export it and create individual metadata files for each image. The image and the metadata files could then ingested and published in large batches.
After materials were ingested from the David Reichard Williams photography collection at the Alexander Archive and became available via the Collection Portal, colleagues in Content Management conducted quality assurance on the ingested data and enhanced the metadata. Finally, Alexander Architectural Archives’ Curator Beth Dodd introduced these published assets to historic preservation professionals and donors to the Alexander Archives, who provided additional information to further describe and enhance information about the buildings in the photographs. Over the course of the project, the crowdsourced assistance of many participants have been instrumental to ingesting assets and enhance the metadata, making for a more robust and discoverable resource for future researchers.
“The Williams project has been a particular example of a collaborative, iterative process to transfer our legacy assets to the DAMS and publish them to the collections portal. It has also been a great learning opportunity and we are taking what we have done here to inform future collaborative work with our collections and metadata transformation” says Katie Pierce Meyer.
Another extremely visible digital collection has also played a significant role in the growth of DAMS and Collections Portal content. The PCL Maps Collection – which is celebrating its 25th anniversary this year – is perhaps the most heavily used of our collection, largely due to the 70,000 items that are available through the Libraries’ legacy website. Visitation to the online maps has accounted for over 50% of all Libraries’ web traffic at points, and has exceeded 5 million views with consistent frequency. The Libraries’ launched a new website in 2018, and have begun to migrate the Maps Collection into the DAMS where it will be available through the Collections Portal. The legacy website remains active largely to maintain access to the collection, so ingesting the digital content from the Maps Collections is another high priority for the overall project.
The migration of the collection into the DAMS is providing the opportunity to greatly improve upon the associated metadata and, in some cases, to provide even higher quality digital scans for use by researchers. “In the DAMS we can store and serve larger format images, which is a great improvement and there are established organization standards, where the legacy site grew organically from its early adoption roots,” says Maps Collection Coordinator Kat Strickland. “Many of the maps in the collection have made their way here without any context. So being able to show somebody the image and describe with more robust metadata is also going to improve discoverability for people.”
“The DAMS is going to benefit users because collections can be organized in a way that will help users find the context of individual maps by linking to a subcollection of related maps.”
When the university shuttered operations in March and physical access to the Maps Collection was halted, only 77 items had been migrated to the DAMS. A short seven months later, there are over 14,000 maps in the system and Libraries’ staff are currently working on metadata for another 11,600 to make those available.
That experience mirrors the shift in focus since remote work has become the prevailing mode of service at the Libraries and online content has become the primary resources for users. In March, there were approximately 2,500 digital assets available through the Collections Portal. Today, there are over 20,000 assets available through the Collections Portal, and those numbers are expanding apace as more resources are committed to the work and staff adapt innovative approaches to their processes.
“There’s been an eightfold increase in content since March, which is just amazing progress and wouldn’t have been possible without the support of many colleagues,” says Mirko Hanke.
The Simple Web Archiver—a straightforward, open source web archiving tool to create personal archives of websites and the files they host—has been published on GitHub under the GNU General Public License, allowing users to use and remix the tool with minimal limitations. The tool, built in Python, provides a GUI interface, and uses BeautifulSoup and wget to parse websites and download files, respectively. I created the tool as part of my work as the European Studies Librarian at the UT Libraries.
Archiving websites is an important practice for anyone interested in preserving digital history. Digital media, and media online, is particularly vulnerable to being lost, as it is often ephemeral in nature and not preserved in an archival format. Saving born-digital materials complements the archiving, curation, and preservation of physical materials, and helps to ensure that internet-based ephemera will be preserved into the future.
Why use this tool?
This tool provides an easy way to create small, personal archives that live offline. While there are many useful web archiving tools available (listed below), this program fills a gap not addressed by existing solutions. Its scope is intentionally small: it aims to create low-memory- use archives for personal use, and to be as easy to use as possible so that users with limited technical knowledge can begin using it immediately, without a complicated setup process or learning curve.
The tool uses a GUI to make the tool very easy to use. Following the directions on the GitHub site allows one to set up the tool and begin using it almost instantly. Another important aspect of this tool is the ease with which it can be modified, by those with some coding experience, to accomplish something else or adapt to the behavior of a certain site. One example of such remixing is this code to capture Omeka sites, specifically, downloading more of the site’s content than the Simple Web Archiver does by default.
Extant web archiving tools tend to accomplish different things than the Simple Web Archiver, and to encompass different scopes. Here is a brief review of other popular software:
Internet Archive – An excellent and easy-to-use tool, but the archives created are hosted online, on the Internet Archive’s servers. Also, not all files will be preserved when crawling a website (PDFs, for example, cannot be archived).
ArchiveIt: also from the Internet Archive, this is great for institutions who want online hosting. It operates on a paid model and, again, not ideal for individual researchers or archivists who want a quick, easy archive of a site and its files.
HTTrack: HTTrack is a free, offline browser utility. Per the tool’s website, it “allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your computer. HTTrack arranges the original site’s relative link-structure.” This is good for those who want a thorough, complete archive of a site, but is not geared toward quick, low-memory-use archives to be stored for personal use.
WarcIt – This tool is entirely programmatic, and only provides WARC (Web ARChive) files. There is no GUI available.
Archivebox – This tool is self-hosted, but programmatic. More robust than the Simple Web Archiver, but its features are not necessarily needed for quick, easy-to-setup or one-off archives. It does not save PDFs, or other files.
Wget – A programmatic tool to download content from the internet. The Simple Web Archiver tool primarily uses wget on the backend to grab online materials.
This tool should work well out of the gate, but there is always the possibility that certain websites, due to their specific architectures, may not be completely archived. The tool’s code was written to be open-ended and adapt to many different types of sites, but for users with specific wants or use cases, it provides a blueprint for the creation of a variation on the tool, or even a completely new piece of software. It is also designed to run relatively quickly, and to grab the main content of a site without unnecessarily consuming CPU power.
The code is simple, and all in one file. The two main functions in the tool download either HTML/CSS/other file types or WARCs, depending on user preference.
The code is also released under the GNU General Public License v3.0. This is a strong copyleft license conditioned on making available complete source code of licensed works and modifications, which include larger works using a licensed work, under the same license. Using this license allows for a wide range of remix and reuse by users and programmers.
I would encourage anyone interested in web archiving to give the tool a try, and to contribute in any way they’d like: by remixing the tool’s code, forking the GitHub repository, or by simply using the tool and providing any feedback they’d like to share.
The Archive of the Indigenous Languages of Latin America (AILLA) is delighted to announce the launch of a free online course called Archiving for the Future: Simple Steps for Archiving Language Documentation Collections, available at https://archivingforthefuture.teachable.com/. The course material is based upon work supported by the National Science Foundation under Grant No. BCS-1653380 (Susan S. Kung and Anthony C. Woodbury, PIs; September 1, 2016, to August 31, 2020). The course is licensed under a Creative Commons Attribution-ShareAlike 4.0 International license.
The course is a resource to aid people of all backgrounds in organizing born-digital and digitized language materials and data for deposit into any digital repository (not just AILLA) for long-term preservation and accessibility. The target audience for this course is anyone who is engaged in creating materials in or about Indigenous, endangered, under-documented, or minority languages as part of language documentation efforts, including language rights, maintenance, and revitalization. It was designed particularly for individuals or groups made up of academic researchers and/or Indigenous or endangered language speakers and community members, though anyone may benefit from it.
The curriculum follows simple steps to guide participants through three phases of work to organize language documentation materials for archiving, and it explains in detail what to do before, during, and after data collection to facilitate the long-term preservation of the data. The course is designed to be informative, engaging, and accessible to anyone, especially to those with no previous experience archiving collections of language materials.
This course was developed by four members of the AILLA staff: Susan Kung, AILLA Manager and grant co-PI; Ryan Sullivant, AILLA Language Data Curator; Alicia Niwabaga, Graduate Research Assistant 2017–2018; and Elena Pojman, Undergraduate Research Assistant 2019–2020. Sullivant and Kung interviewed representatives of various DELAMAN (delaman.org) archives and other digital data repositories in the United States, the United Kingdom, the European Union, Australia, and Cameroon. Niwagaba collaborated with Kung and Sullivant to develop an early version of the course that the AILLA team taught live at the Institute on Collaborative Language Research (CoLang 2018) at the University of Florida in Gainesville during June 18–22, 2018. Niwagaba created the educational animated videos that are embedded in the course to illustrate key aspects of the curriculum. Pojman researched curriculum platforms in which to build the online course. Teachable was selected for a variety of reasons, including its simple yet attractive aesthetic that displays all course modules in the left side bar (see illustration below); its ease of use and progress tracking for enrolled students; its responsiveness to different technology; and the built-in ability to quickly and easily set up the same course in multiple languages. This last feature is especially important since AILLA staff plan to translate the curriculum into Spanish and Portuguese to make it more accessible to AILLA’s Latin American audience. Once the curriculum software was selected, Kung and Sullivant expanded the original 2018 workshop curriculum and wrote the additional content. Pojman wrote the objectives and activities for each step, built the English course in Teachable, and created all of the graphics that are used in the curriculum.
In funding and academic environments where it is becoming increasingly common for researchers to be responsible for archiving their own research data, the AILLA staff saw a need to train language researchers to do this work so that the resulting language collections would be well organized, well described, easy to navigate, and available to reuse for further research and education. While there are some language documentation programs in North America, Europe, Australia, and New Zealand that train language documenters to do these tasks, most do not, and almost no training on how to archive language documentation is available in Latin America. The AILLA team created this course to fill these gaps.
Vea abajo para versión en español / Veja em baixo para versão em português
In honor of World Digital Preservation Day, members of the University of Texas Libraries’ Digital Preservation team have written a series of blog posts to highlight preservation activities at UT Austin, and to explain why the stakes are so high in our ever-changing digital and technological landscape. This post is the final installment in a series of five.Read part one, part two, part three, and part four.
BY ASHLEY ADAIR, Head of Preservation and Digital Stewardship, University of Texas Libraries
The UT Libraries’ Digital Stewardship unit supports digital preservation work across the University of Texas Libraries. When Libraries repositories, such as the Alexander Architectural Archives, LLILAS Benson, or the Archive of Indigenous Languages of Latin America begin new digital projects, the Digital Stewardship unit often helps develop initial processing plans. Unit staff install tools and provide training to recover data from older media such as floppy disks and Zip disks, or for acquiring files produced by partner organizations and depositing researchers. Processing of these materials must be planned and undertaken very carefully since data may be at risk of permanent loss due to obsolete formats and media, or because of political or physical issues in local environments.
Taking a life-cycle approach, the unit also coordinates long-term safekeeping of these valuable and sometimes vulnerable files. Digital Stewardship developed file organizing, naming, and description practices for uniformly storing all of UT Libraries’ diverse preservation data in keeping with international standards. When repository staff complete processing, the Digital Stewardship unit takes in copies of data to be preserved, vaults them to long-term storage, maintains detailed centralized records, and manages off-site backup copies. The unit collaborates with UT Libraries repositories continuously over time to enhance organization-wide digital preservation practices, adapting to new developments and the growing scale of data to be preserved.
Traducido por Jennifer Isasi, PhD (@jenniferisve)
La unidad de Administración Digital de las Bibliotecas de la Universidad de Texas (UT) apoya el trabajo de preservación digital en el conjunto de bibliotecas de la universidad. Cuando repositorios como el Archivo de Arquitectura Alexander, LLILAS Benson o el Archivo de Lenguas Indígenas de Latinoamérica comienzan nuevos proyectos digitales, la unidad de administración digital ayuda a desarrollar planes de procesamiento. El personal de la unidad instala herramientas y provee entrenamiento para recuperar datos de medios antiguos como disquetes o discos Zip, o para la adquisición de archivos producidos por organizaciones colaboradoras e investigadores que depositan sus archivos en los repositorios. El procesado de estos materiales debe ser planeado y realizado con mucho cuidado puesto que los datos pueden estar en peligro de borrado permanente debido a formatos o medios obsoletos, o por cuestiones políticas y de tipo medioambiental.
Con un enfoque de ciclo de vida de los datos, la unidad también coordina la custodia a largo plazo de estos archivos valiosos y a veces vulnerables. La administración digital desarrolló prácticas de organización, denominación y descripción de archivos para almacenar de manera uniforme todos los diversos datos de preservación de las bibliotecas de UT de acuerdo con los estándares internacionales. Cuando el personal del repositorio completa el procesamiento, la unidad de Administración Digital toma copias de los datos para preservarlos, los guarda en un almacenamiento a largo plazo, mantiene registros centralizados detallados y administra copias de seguridad en otras localizaciones. La unidad colabora con los repositorios de las bibliotecas UT continuamente a lo largo del tiempo para mejorar las prácticas de preservación digital de toda la organización, adaptándose a los nuevos desarrollos y la creciente escala de datos a preservar.
Traduzido por Tereza Braga
A unidade de Gestão Digital da UT Libraries apoia o trabalho de preservação digital de todas as bibliotecas do sistema. Quando um dos repositórios das Bibliotecas, seja o Alexander Architectural Archives, a LLILAS Benson ou o Archive of the Indigenous Languages of Latin America, inicia um projeto digital novo, a unidade de Gestão Digital geralmente auxilia a criar os planos iniciais de processamento. Os profissionais da unidade instalam ferramentas e dão treinamento para recuperar dados de mídias mais antigas como floppy disks e discos Zip ou para adquirir arquivos produzidos por organizações parceiras e pesquisadores com trabalhos depositados. O processamento desses materiais deve ser planejado e empreendido com muito cuidado, pois os dados podem estar expostos ao risco de perda permanente causado por formatos e mídia obsoletos ou por problemas políticos ou físicos em ambientes locais.
Utilizando uma abordagem de ciclo de vida, a unidade também coordena a guarda a longo prazo desses arquivos valiosos e às vezes vulneráveis. A Gestão Digital desenvolve práticas para organizar, dar nomes e descrever os arquivos visando a armazenagem uniforme de todos os diversos dados de preservação da UT Libraries em conformidade com as normas internacionais. Quando os funcionários de repositórios concluem seu processamento, a unidade de Gestão Digital providencia cópias dos dados a serem preservados, armazena-os em sistema de armazenagem segura de longo prazo, mantém registros centralizados detalhados e providencia cópias de reserva em local externo. A unidade colabora de modo contínuo com os repositórios da UT Libraries ao longo do tempo para aprimorar as práticas de preservação digital em toda a organização, sempre se adaptando aos novos avanços e ao aumento em escala do universo de dados a serem preservados.
Vea abajo para versión en español / Veja em baixo para versão em português
In honor of World Digital Preservation Day, members of the University of Texas Libraries’ Digital Preservation team have written a series of blog posts to highlight preservation activities at UT Austin, and to explain why the stakes are so high in our ever-changing digital and technological landscape. This post is part four in a series of five.Read part one, part two, and part three.
By DAVID BLISS (@davidallynbliss), Digital Processing Archivist, LLILAS Benson Latin American Studies and Collections @llilasbenson
Over the past decade, LLILAS Benson has undertaken post-custodial archival projects in collaboration with partners throughout Latin America and beyond. Post-custodial archival practice encompasses a range of theory and methodology, built on the premise that digital technologies make it possible for collecting institutions like LLILAS Benson to provide access to archival collections from Latin America without taking physical custody or removing them from their original contexts of creation and use.
Through these post-custodial projects, LLILAS Benson staff and partner repository staff work together closely to identify collections of interest, select appropriate digitization equipment, and build metadata collection strategies. The materials are then digitized and described on-site in Latin America by partner repository staff. The digitized collections are then transferred to LLILAS Benson, where they are processed, preserved, and in most cases published online. Because the original collections are often vulnerable or sensitive, frequently touching on delicate human rights issues, long-term preservation of their digital copies is especially important to LLILAS Benson staff and partners in Latin America.
In recent years, the LLILAS Benson team has integrated file fixity checks in all post-custodial projects. When launching a project at a partner site, LLILAS Benson staff now teach project team members the basic principles of digital preservation and the importance of fixity checks, which verify that files have not been altered or corrupted over time. The project teams are taught to create and verify checksums prior to transferring a batch of files to LLILAS Benson, using free software available in Spanish or Portuguese.
These checksums now accompany all file deliveries from project sites, and help the LLILAS Benson team identify corrupted or missing files immediately. These checksums speed LLILAS Benson’s processing and preservation work, allowing the files to be published online and preserved long-term more easily. The checksum workflow also encourages each partner to include fixity checks in any future digitization projects they undertake, thus contributing to the partners’ own digital preservation capacity.
Equipo poscustodial LLILAS Benson
Traducido por Jennifer Isasi(@jenniferisve)
Durante la última década, LLILAS Benson ha emprendido proyectos de archivo de tipo poscustodial junto con socios a lo largo de América Latina. La práctica de archivo poscustodial abarca una serie de teorías y metodologías basadas en la premisa de que las tecnologías digitales hacen posible que las instituciones colectoras como LLILAS Benson provean acceso a las colecciones de archivos de Latinoamérica sin su custodia física o su eliminación del contexto original de su creación y uso.
A través de estos proyectos poscustodiales el personal de LLILAS Benson y sus colaboradores trabajan en estrecha colaboración para identificar colecciones de interés, seleccionar el equipo de digitalización adecuado y desarrollar estrategias de curaduría de metadatos. Los materiales son digitalizados y descritos en Latinoamérica por parte del personal de cada archivo para luego ser transferidos al equipo LLILAS Benson, quien procesa, preserva y publica los materiales en la mayoría de los casos. Debido a que las colecciones originales son a menudo vulnerables o con contenido delicado, y frecuentemente tocan temas relacionados con derechos humanos, la preservación a largo plazo de sus copias digitales es especialmente importante para el personal y los socios de LLILAS Benson en América Latina.
En años recientes, LLILAS Benson ha añadido verificaciones de permanencia de archivos en los proyectos poscustodiales en curso. Con el inicio de cada proyecto en el archivo de los colaboradores, el personal de LLILAS Benson enseña a cada equipo los principios básicos de preservación digital y la importancia de añadir verificaciones de permanencia, que verifican que los archivos no han sido alterados o dañados con el tiempo. Los equipos de los proyectos aprenden a crear y verificar sumas de verificación usando programas gratuitos en español o portugués antes de transferir un conjunto de archivos a LLILAS Benson.
Estas sumas de verificación ahora acompañan todas las entregas de archivos desde el lugar de los proyectos de digitalización y ayudan al equipo de LLILAS Benson a identificar archivos dañados o faltantes de inmediato. Esto acelera las tareas locales de procesamiento y preservación en LLILAS Benson y anima a cada colaborador a incluir controles de verificación en cualquier otro proyecto que puedan emprender en el futuro. Esto a su vez contribuye a la capacidad de preservación digital propia de los colaboradores.
Equipe pós-custodial da LLILAS Benson
Traduzido por Tereza Braga
Durante a última década, a LLILAS Benson empreendeu alguns projetos arquivísticos pós-custodiais, em colaboração com entidades parceiras espalhadas pela América Latina e outros lugares. A prática arquivística pós-custodial engloba uma gama de teorias e metodologias assentadas na premissa de que as tecnologias digitais possibilitam a instituições recolhedoras de coleções, como a LLILAS Benson, disponibilizar o acesso a coleções arquivísticas latino-americanas sem necessidade de obter custódia física ou a remoção das mesmas de seus contextos originais de criação e de uso.
Por meio desses projetos pós-custodiais, as equipes de profissionais da LLILAS Benson e dos repositórios parceiros trabalham em contato estreito para identificar coleções de interesse, selecionar o equipamento de digitalização adequado e criar estratégias de coleta de metadados. O material é então digitalizado e descrito pela equipe de repositório da entidade parceira em cada local específico da América Latina. Em seguida, as coleções digitalizadas são transferidas para a LLILAS Bensonm onde são processadas, preservadas e, na maioria dos casos, publicadas online. Devido ao fato de muitas coleções originais serem vulneráveis ou sensitivas por causa de referências frequentes a questões delicadas de direitos humanos, a preservação a longo prazo de cópias digitais é especialmente importante para a equipe da LLILAS Benson e entidades parceiras na América Latina.
Em anos recentes, os profissionais da LLILAS Benson vêm integrando verificações de fixidez de arquivos em todos os projetos pós-custodiais. Agora, ao lançar um projeto em local parceiro, a equipe ensina às equipes do projeto os princípios básicos da preservação digital e a importância das verificações de fixidez para constatar se os arquivos não foram alterados ou corrompidos ao longo do tempo. As equipes de projeto aprendem a criar e verificar as checksums (somas de verificação) antes de transferir qualquer lote de arquivos para a LLILAS Benson, usando software gratuito disponível em espanhol e português.
Essas checksums já acompanham todas as entregas de arquivos oriundos de locais de projetos e ajudam a equipe da LLILAS Benson a identificar imediatamente arquivos corrompidos ou faltando. As checksums aceleram o trabalho de processamento e preservação da LLILAS Benson, permitindo publicar os arquivos online e preservá-los a longo prazo com mais facilidade. O fluxograma de checksums também incentiva cada entidade parceira a incluir verificações de fixidez em qualquer projeto de digitalização a ser empreendido no futuro contribuindo, assim, para a própria capacidade de preservação digital de cada entidade.
Vea abajo para versión en español / Veja em baixo para versão em português
In honor of World Digital Preservation Day, members of the University of Texas Libraries’ Digital Preservation team have written a series of blog posts to highlight preservation activities at UT Austin, and to explain why the stakes are so high in our ever-changing digital and technological landscape. This post is part one in a series of five.
Introduction to Digital Preservation
BY DAVID BLISS, Digital Processing Archivist, LLILAS Benson Latin American Studies and Collections; ASHLEY ADAIR, Head of Preservation and Digital Stewardship University of Texas Libraries
In recent decades, the archival field has been transformed by the rise of digital historical records. As computers of all kinds have worked their way into many areas of our professional and personal lives, collections of documents donated to archives in order to preserve individual and institutional histories have come to comprise both traditional paper records and those created using these computers. Digital records can be scans of paper or other objects, born-digital files comparable to paper records, such as Word or text documents, or entirely new kinds of objects, such as video games. Archivists are committed to preserving digital records, just like physical ones, for future generations to use and study. Digital preservation refers to the full range of work involved in ensuring digital files remain accessible and readable in the face of changing hardware and software.
Unlike traditional physical media like paper,
which can typically be kept readable for decades or centuries with proper
housing and ambient conditions, digital files can be lost without periodic,
active intervention on the part of archivists: legacy file formats can become
unreadable on modern computers; hard drives and optical media can break or
degrade over time; and power outages can cause network storage to fail. Digital
archivists take steps to prevent and prepare for these contingencies.
There is no one perfect or even correct solution to the challenge of preserving digital files, so each institution may use different tools, standards, and hardware to carry out the work. Typically, however, digital preservation involves choosing suitable file formats, maintaining storage media and infrastructure, and organizing and describing digital objects in a standardized way that ensures future archivists and users can understand and access what has been preserved.
Digital preservation represents a significant effort that cannot be carried out by a single person or group. At the University of Texas Libraries, dissemination of digital preservation knowledge and skills is a crucial part of digital preservation practice. Training and pedagogy spread digital preservation expertise within the organization and out to researchers and partners, allowing the Libraries to preserve an ever-growing amount of valuable data.
Introducción a la
Para el Día Mundial de la Preservación Digital, los miembros del equipo de Preservación Digital de las Bibliotecas de la Universidad de Texas han escrito una serie de entradas de blog que hacen destacar las actividades de preservación en la universidad, y para enfatizar la importancia de la preservación en un presente de cambio tecnológico constante. Este texto es el primero en una serie de cinco.
Traducido por Jennifer Isasi, Postdoctoral Fellow in Data Curation in Latin American and Latina/o Studies
recientes, el ámbito de los archivo se ha visto transformado con el aumento de
los registros históricos digitales. A medida que las computadoras de todo tipo
han pasado a formar parte de muchas áreas de nuestra vida profesional y
personal, las colecciones de documentos donados a los archivos para preservar
historias individuales e institucionales ahora presentan tanto los registros en
papel tradicionales como los creados con computadoras. Los registros digitales
pueden ser copias escaneadas de papel u otros objetos, archivos digitales
nativos similares a los registros en papel, como documentos de Word o texto, o
tipos de objetos completamente nuevos, como los videojuegos. Los archivistas
están comprometidos a preservar los registros digitales, al igual que los
físicos, para que las generaciones futuras los utilicen y estudien. Así, la
preservación digital se refiere a la gama completa de trabajo involucrado en
garantizar que los archivos digitales permanezcan accesibles y legibles ante el
cambio de hardware y software.
diferencia de los medios físicos tradicionales como el papel, que por lo
general pueden ser preservados por décadas o siglos en condiciones de guardado
adecuadas, los archivos digitales pueden perderse sin la intervención periódica
y activa por parte de los archivistas: las computadoras modernas no pueden leer
algunos de los formatos de archivo más antiguos, los discos duros o los medios
ópticos se pueden romper o degradar con el tiempo y los cortes de luz pueden
causar fallos en el almacenamiento en la red. Los archivistas digitales toman
medidas para prevenir o prepararse para este tipo de imprevistos.
No hay una
solución perfecta ni correcta para el desafío de preservar archivos digitales,
por lo que cada institución puede utilizar diferentes herramientas, estándares
y equipos para este trabajo. Por lo general, no obstante, la preservación
digital implica elegir formatos de archivo adecuados, mantener medios de
almacenaje y su infraestructura así como asegurar la organización y la
descripción de los objetos digitales de una manera estandarizada que garantice
que los futuros archivistas y usuarios puedan comprender y acceder al material
El trabajo y esfuerzo necesarios para la preservación digital no puede ser realizado por una sola persona o grupo. En el conjunto de bibliotecas de la Universidad de Texas, la difusión del conocimiento sobre preservación digital es una parte crucial de la práctica de preservación. Mediante esfuerzos de capacitación y pedagógicos tanto dentro de la organización como entre investigadores y colaboradores, estas bibliotecas están logrando preservar una cantidad cada vez mayor de datos relevantes.
Introdução à preservação digital
Traduzido por Tereza Braga
Para o Dia Mundial da Preservação Digital, os membros do equipe de Preservação Digital das Bibliotecas da Universidade de Texas escreveram uma serie de entradas de blog que enfatizam as atividades de preservação na nossa universidad, para explicar a importancia da preservação no contexto de um presente de tecnología em fluxo constante. Este texto é o primeiro numa série de cinco.
O advento dos registros históricos digitais causou uma completa transformação do setor arquivístico nas últimas décadas. Computadores de todos os tipos estão cada vez mais presentes em cada vez mais aspectos da vida profissional e pessoal. Essa mudança também afeta as coleções de documentos que são doadas a instituições arquivísticas com o intuito de preservar histórias individuais e institucionais. Hoje em dia, uma coleção pode reunir tanto registros tradicionais em papel quanto registros criados por esses diversos computadores. O que chamamos de registro digital pode ser uma simples página ou objeto que tenha sido escaneado ou qualquer arquivo que já tenha nascido em forma digital e que seja comparável com um registro em papel como, por exemplo, um texto regidido em Word. Registro digital pode também significar uma coisa inteiramente nova como um videogame, por exemplo. Arquivistas são profissionais que se dedicam a preservar registros digitais para utilização e estudo por futuras gerações, como já é feito com os registros físicos. A preservação digital pode incluir uma ampla variedade de tarefas, todas com o objetivo comum de fazer com que um arquivo digital se mantenha acessível e legível mesmo com as frequentes mudanças na área de hardware e software.
Um arquivo digital é diferente do arquivo em papel ou outros meios físicos tradicionais, que geralmente pode ser mantido legível por muitas décadas ou mesmo séculos, se armazenado em invólucro adequado e sob as devidas condições ambientais. Um arquivo digital pode se perder para sempre se não houver uma intervenção periódica e ativa por parte de um arquivista. Certos arquivos em formatos mais antigos podem se tornar ilegíveis em computadores modernos. Discos rígidos e mídia ótica podem quebrar ou estragar com o tempo. Cortes de energia podem causar panes em sistemas de armazenagem em rede. O arquivista digital é o profissional que sabe tomar medidas tanto de prevenção quanto de preparação para essas e outras contingências.
Não existe solução perfeita, ou sequer correta, para o desafio que é preservar um arquivo digital. Diferentes instituições utilizam diferentes ferramentas, normas e hardware. De maneira geral, no entanto, as seguintes tarefas devem ser realizadas: escolher o formato de arquivo adequado; providenciar e manter uma mídia e infra-estrutura de armazenagem; e organizar e descrever os objetos digitais de uma maneira que seja padronizada e que permita a futuros arquivistas e usuários entender e acessar o que foi preservado.
A preservação digital é um empreendimento importante que não pode ser executado por apenas um indivíduo ou grupo. Na UT Libraries, a disseminação de conhecimentos e competências de preservação digital é uma parte essencial dessa prática. Temos cursos de capacitação e pedagogia para disseminar essa especialização em preservação digital para toda a organização e também para pesquisadores e parceiros externos. É esse trabalho que capacita a Libraries a preservar um grande volume de dados valiosos que não pára de crescer.
Durante el verano, LLILAS Benson y el Museo de la Palabra y la Imagen (MUPI) en El Salvador agregaron otra iniciativa digital a su portfolio de colaboración. Desde 2012, las dos instituciones han trabajado juntos para digitalizar archivos relacionados a la Guerra Civil Salvadoreña (1980–1992), gracias al generoso apoyo de la Fundación Andrew W. Mellon. Continuando estos esfuerzos, esta nueva iniciativa también exploró el potencial de las humanidades digitales para destacar una de las colecciones más impresionantes de MUPI: los bordados testimoniales de refugiados salvadoreños.
Los testimonios sobre la violación de derechos humanos se presentan en diferentes formas, y el fundador y actual director de MUPI, Carlos “Santiago” Henríquez Consalvi, ha procurado preservar la diversidad. Poco después de la firma de los Acuerdos de Paz de Chapultepec en 1992 que pusieron fin a la Guerra Civil Salvadoreña, Santiago dirigió una campaña para rescatar el patrimonio cultural creado antes, durante y después del conflicto armado. Esto ha incluido propaganda política, publicaciones y las grabaciones de la estación de Radio Venceremos. Desde su fundación formal en 1999, MUPI ha continuado esta preservación y ha expandido su enfoque para incluir varios temas sobre la cultura e historia salvadoreña.
La colección que ha crecido más recientemente, y el enfoque de esta nueva iniciativa, consiste de bordados testimoniales creados por campesinas salvadoreñas refugiadas en Honduras durante la guerra civil. Estas piezas fueron creadas para comunicar al mundo las experiencias vividas de los refugiados, y muchos de los textiles se enviaron a grupos y organizaciones de solidaridad en Europa y Canadá para ello. Gracias a una campaña internacional reciente, más de veinte obras han sido repatriadas y enviadas a MUPI. A través de talleres en las comunidades rurales de El Salvador, MUPI ha renovado el aprecio por esta tradición cultural, promoviendo el arte y los esfuerzos de repatriación a través de una exposición titulada Bordadoras de Memoria en la capital.
Ahora que los bordados están volviendo a casa, MUPI está utilizando tecnologías digitales para continuar el trabajo de abogar por los derechos humanos que estas mujeres comenzaron en la década de los 1980s. Para alcanzar y educar a un público más amplio e internacional, específicamente jóvenes descendientes de salvadoreños en los Estados Unidos, el Museo trabajó con el personal de Estudios Digitales en LLILAS Benson (LBDS) para recrear Bordadoras de Memoria en línea. En junio, el equipo de LBDS viajó a San Salvador y capacitó al diseñador gráfico de MUPI, Pedro Durán, en el uso de la plataforma Omeka para que pudiera reconcebir la exhibición digitalmente, utilizando fotografías preliminares de los bordados. El equipo también aprovechó la oportunidad para hablar sobre otras herramientas de código abierto que el personal de MUPI puede usar en su trabajo con jóvenes locales.
La visita también lanzó otro proyecto archivístico pos-custodial para ambas instituciones. Dado el tamaño de algunas obras (la pieza que se muestra arriba es más de 2.5 metros de largo), el proyecto requirió un flujo de trabajo completamente diferente en la digitalización y entrenamiento en nuevos equipos. Capacitados por el personal de archivos pos-custodiales (PC) de la Colección Latinoamericana Benson, el equipo de LBDS trabajó con el personal de MUPI para iniciar la digitalización y la descripción archivística de los bordados. El equipo de PC espera incorporar la colección al portal Latin American Digital Initiatives a finales de este año, así que estense atentos.