2013 TCDL Abstracts

 Monday, May 6

Pre-Conference Event: DigiCamp 1:00 – 4:00 PM

Radisson Hotel & Suites (Old Pecan Street Room)

Tuesday, May 7

Opening Keynote Address (Austin Room)

Curation and Preservation Services: Adapting Frameworks and Tools to Enable Sustainable Programs
Nancy McGovern, Head of Curation and Preservation Services, MIT Libraries

Abstract
Since 1996, the digital preservation community has been developing and refining standards and practice. For more than a decade, organizations responsible for managing digital content over time have been adapting and adopting the results of those community efforts.  This paper looks at organizational examples of using community documents such as Trusted Digital Repositories: Attributes and ResponsibilitiesOpen Archival Information System (OAIS) Reference Model, and Trustworthy Repositories Audit and Certification(TRAC), and other community documents as frameworks for developing and sustaining digital curation and preservation programs. The backdrop for this work is the organizational developmental model developed by Kenney and McGovern for the Digital Preservation Management workshop series that has been running since 2003.

Keywords
preservation


Session 1A (Austin Room):

Online Metadata Generator for Research Data
Michele Reilly, Manual de la Cruz Gutierrez, University of Houston

Abstract
NSF, NIH and other funding agencies are now requiring Data Management Plans that include the expectation of sharing research data. To facilitate this expectation, librarians from the University of Houston are developing an automated online system to document data sets at the various stages of their life cycle. Our process will consist of determining key Dublin Core elements and developing a discipline-specific crosswalk. (This crosswalk translates discipline-specific terminology to that used by Dublin Core.) The discipline-specific terminology and the key element set will be developed in collaboration with faculty researchers. This set will be minimal in order to reduce the burden to the researcher in documenting their data yet provide enough information for the management of the data over its life cycle. Ultimately, the resulting sets will be shared across each discipline community to foster a shared set of standards. This crosswalk will result in a discipline-specific metadata generator. It is envisioned that this discipline-specific metadata generator will be an online system that allows researchers and their assistants to develop metadata on the fly. The discipline-specific metadata generator will crosswalk to Dublin Core ultimately ensuring sharing, discoverability and interoperability of data seamlessly while minimizing researcher effort.

The presenters will discuss how they envision the system will work, what steps they are taking in determining key discipline-specific elements – including how those elements crosswalk to Dublin Core – , the progress to date, and the importance of inter-departmental collaboration necessary to make this system a reality.

Keywords
research data; metadata; Dublin Core

Crowdsourcing + Machine Learning: Building an Application to Convert Scanned Documents to Texas
Nicholas Joel Woodward

Abstract
Widespread digitization initiatives and the concomitant explosion in digital corpora have redefined the roles of academic libraries in recent years. Many current efforts in the academic community focus on making digital content accessible and legible to mass audiences for a variety of purposes, and the transcription of scanned documents is one integral component. Difficulties inherent to the process of optical character recognition mean that most digital artifacts containing text are converted to scanned images that lack full-text search capabilities. In many cases researchers must resort to either manually entered metadata (generally unfeasible with large-scale data) or crowdsourced input from users (only applicable on a per-item basis).

To these ends, I have developed an application and workflow for the large-scale transcription of scanned artifacts by combining limited user input (crowdsourcing) with machine learning on a high performance computing cluster to recognize patterns of matching words across artifacts and mechanically transcribe them. The application consists of a collection of tightly integrated Java libraries that 1) mechanically segment scanned images into individual words, 2) use one of several pattern recognition algorithms available as open source code to match similar images (i.e. words) across the entire corpus, and 3) transcribe these words using either open source OCR software or an online crowdsourcing tool written in PHP.

The Digital Archive of the Guatemalan National Police Historical Archive serves as the test case for the application. The collection has several characteristics that make it ideal for measuring the performance of the transcription process. Its large size (12+ million pages), range of document formats (forms for births, marriages and deaths, typed letters and handwritten journals), and variance in scanning quality all present unique challenges to manual transcription.

The presentation will describe in detail the process of developing the application and workflow, the challenges presented by the test collection and how they were met and, finally, the preliminary results from the transcription process. It will focus on the following:

– Methods for segmenting and matching words

– Challenges to the proposed approach

– Mutually beneficial relationship between user input and machine learning

– Opportunities for future improvements and applications

Keywords
digital archives; scanned documents; ocr; crowdsourcing

Session 1B (Travis I/II Room):

Preserving the Image of Fandom: The Sandy Herald Digitized Media Fanzine Collection at Texas A&M University
Jeremy Brett, Texas A&M University

Abstract
Media fandom – the cultural practice of active interest in various movies or television shows – is a widespread and vibrant part of American popular culture. Fans create all sorts of artifacts related to the objects of their affection, including fanzines. Media fanzines are amateur publications usually (though not always) containing works of fan fiction. Fanzines have been important aspects of fandom for decades – many were created as ephemeral, impermanent print objects, while others were born digital. In either case, whole generations of media fanzines are disappearing and with them, the creative record of this colorful phenomenon. Texas A&M University is involved in creating a unique digital repository consisting of thousands of scanned and archived fanzines dating from the 1960s to the present.

Preserving and providing access to fannish materials presents several unique challenges beyond the merely technical. There is the issue of copyright and appropriate permission: most fanzines were copyrighted by their creator and therefore require permission in order to make them accessible. Older fans have often drifted away from their informal creations and many good-faith efforts to locate them are necessary. In addition, many media fanzines are actually anthologies with multiple authors and artists, requiring an untangling of rights and permissions from a single product. Efforts at graceful diplomacy are often necessary, because fans are often very protective of their fannish identities and activities, and reluctant to have those “exposed” to the outside world. We are obliged to explain our motives and our belief in the strong research potential of fans’ creations.

Metadata creation raises the issue of fannish anonymity. Some fans – those who used pseudonyms or entirely separate cultural identities as fans – do not want their legal names revealed or used, and the desire for continued anonymity requires metadata decisions that reflect this motivation. Media fanzines can be classified under many different subjects and genres – in constructing the metadata structure for the repository we have this ongoing issue of limiting vocabulary to deal with as well.

Finally, we were faced with the issue of how to provide access. Online access would, of course, vastly increase potential audiences, but secured onsite access would assuage privacy concerns of many fannish donors.  The Hereld DMFC operates now under secured onsite access, but as progress on the collection is made this decision may change.

The construction of the Sandy Hereld DMFC provides a number of interesting questions (and some answers) relating to digital archiving and how institutions can successfully build and maintain repositories of digital material, material with incredible potential for institutional and collection promotion and outreach.

Keywords
digital repositories; archives; metadata; access

The Austin Fanzine Project: Phase Two
Jennifer Hecker, UT Libraries; Kevin Powell, UT School of Information

Abstract
A fanzine is a “nonprofessional and nonofficial publication produced by fans of a particular cultural phenomenon (such as a literary or musical genre) for the pleasure of others who share their interest” (Wikipedia). You can think of them as pre-internet blogs. Fanzines are lately gaining popularity in college classrooms, as professors in journalism, anthropology, art, and literature incorporate them into their curricula, and librarians and archivists all over the world are finding ways to increase access to and preservation of these ephemeral, sometimes unique objects of human expression.

Archivist Jennifer Hecker founded Austin Fanzine Project in the summer of 2012 as a way to increase access to the fanzines that document Austin’s 1990s underground music scene by crowd sourcing their digitization, transcription and indexing. The project also functions as an ad-hoc learning lab for issues around digitization, transcription, linked data, digital preservation, privacy and copyright, and archival collaboration and innovation.

Phase One of the project was a self-contained test run focusing on only one document. Phase Two has already begun and will feature more material, more complex issues, and more collaboration. Ms. Hecker will introduce the project and describe progress made to date, while UT iSchool master’s degree candidate, Kevin Powell, will discuss the policy and workflow issues he addressed during his Spring 2013 Capstone work on the project.

Keywords
archives; digitization; preservation; digital humanities; crowdsourcing


Session 2A (Austin Room):

Galileo Knowledge Repository (GKR): A Collaborative Statewide Institutional Repository
Marlee Givens, Georgia Tech Library

Abstract
In 2004, a small group of libraries from the University System of Georgia (USG) began discussing the idea of building a collaborative statewide institutional repository (IR). Georgia Tech was just starting SMARTech, a DSpace repository, and since then a number of other libraries around the system have started their own IR projects. In 2009, Georgia Tech was awarded a grant from IMLS to build the GALILEO Knowledge Repository. Goals for this project include building and hosting four new repositories; developing a central repository of standardized, harvested metadata from participating IRs in the USG; creating the open source collection mapping tool; implementing IR-related services such as digitization and content submission; conducting a statewide faculty survey; and hosting a national symposium and workshop on consortial IRs. Marlee Givens, the GKR’s project manager, will present GKR as a case study in building and managing a collaborative IR service.

Keywords
institutional repository; collaborative institutional repository

The Evolution of the World Digital Library: Challenges and Opportunities
Erin Hawkins, World Digital Library

Abstract
The World Digital Library, a website available in seven languages and focusing on cultural treasures from around the world, was founded in 2009 as a collaborative project of the Library of Congress, UNESCO, and several national libraries. Now encompassing nearly 200 international institutions, our growth has created new challenges and opportunities. What role does translation play in metadata creation? How do we optimize user experience when our site contains obscure scripts and multiple character sets? When we tweet in Arabic, who tweets back? How can we best assist institutions with less digitization experience to produce quality images and metadata? This discussion will touch on these issues, particularly surrounding multilingualism, translation, outreach, and process improvement regarding partner participation and collaboration.

Keywords
metadata

Session 2B (Travis I/II Room):

Panel: Perspectives on Preservation Strategies for Digital Content: Lessons Learned
Shelley Barba, Joy Perrin, Laura Stennett, Chris Starcher, Heidi Winkler, Le Yang, Texas Tech University

Abstract
This panel will discuss Texas Tech University Digital Library’s experiences with preservation in the past eight years including failures and near-misses that caused TTU to seek a better system. This all-encompassing system influences all processes of digitization from scanning, metadata creation, online display, and to long term storage maintenance.

Discussion will include how we created policies and procedures for handling a variety of materials (including scanning resolution and file type) and a file naming system and folder structure. Other topics will include: balancing work on new collections while correcting older collection, managing a growing digital preservation system when few understand what you are doing or why, reaching out to the university community about preservation needs, and staying a step ahead of our current technology to ensure that we make fully informed preservation and migration decisions.

The panel will be comprised of two librarians who were a part of the creation of the TTU Digital Library, two who were hired at the mid-point, and two recent hires; thus, giving varied perspectives on the collections and preservation issues. The panel will be in a “talk-show” format with prepared questions and discussions, and then opened for questions from the audience.

Keywords
digital preservation


Session 3A (Austin Room)

Panel: Challenges and Opportunities: A Library Directors’ Panel Discussion
Donald Dyal, Dean of Libraries, Texas Tech University Libraries; Christine Shupala, Associate VP for Academic Affairs, Texas A&M-Corpus Christi; Douglas Ferrier, Library Director, Texas A&M International University, Killam Library; Joan Heath, Associate VP University Library, Texas State University; Pamela McQuesten, VP and CIO, Southwestern University; Moderator – William Anderson, UT School of Information

Abstract
A panel consisting of several heads of Texas academic libraries will discuss the pressing issues facing university libraries today. The panel discussion will be led by UT i-School faculty member William Anderson and will include leaders from a representative sample of higher education institutions across the state.

Possible topics touched on in this free-flowing discussion may include digital preservation and data curation as well as larger trends in higher education and scholarly communication.

Keywords
academic libraries

Session 3B (Travis I/II Room)

How to Rock the Planning Process for your Next Digital Library Web Project
Rachel Vacek, R. Niccole Westbrook, University of Houston Libraries

Abstract
Does thinking about web projects for digital libraries make you feel as though you’re trapped between a rock and a hard place?  You will need to engage digital library partners and stakeholders; collect and analyze diverse data streams to understand use of your current digital library; define the roles of project team members; develop a comprehensive communication plan that serves a diverse audience; produce and disseminate project plans; and that’s all before web development of the project even begins!  Talk about a Sisyphean task!  Instead of pushing that rock up the mountain only to have it tumble back down again, join us for a presentation that highlights effective strategies for planning digital library web projects.  This presentation will describe the planning process for the University of Houston Digital Library Redesign Project.  We’ll tell you how we rocked it and we’ll point out where things got rocky.  No matter what web project you are hoping to initiate for your digital library, you can be the project planning rock star.

Keywords
project planning; web projects; digital library

Providing Effective Information Services for RCN CE3SAR Project
Hong Xu, Texas A&M University-Corpus Christi

Abstract
The research coordination network (RCN) – Climate, Energy, Environment and Engagement in Semiarid Regions (CE3SAR) is a NSF funded project. The overarching goal of the RCN CE3SAR is to form a robust research, educational and engagement network of regional universities, research centers and institutes. The project has extensive information needs such as project presence, data preservation, digital repositories, scholarly communication, project management, information retrieval, and bibliography service.

Texas Digital Library and the library of Texas A&M University – Corpus Christi have played important roles in providing services to meet the needs of the RCN CE3SAR project. This presentation aims to share the experience of providing information services to RCN CE3SAR project from the following aspects:

  • Outreaching RCN CE3SAR team;
  • Identifying the project users’ needs;
  • Providing information services;
  • Engaging the project users;
  • Discussing challenges and opportunities.

Keywords
institutional repository; project management; website; bibliography


Session 4A (Austin Room)

Panel: How Digital Libraries Can Create a Culture of Open Access on Campus 
Spencer D. C. Keralis, Kris Helge, Laura Waugh, Shannon Stark, Anjum Najmi, University of North Texas

Abstract
As Open Access has flourished into an International movement that is shaping the progressive landscape of scholarly communication, a growing number of institutions are implementing policy changes aimed at the higher institutional levels. Policy implementation, however, is only the one step in creating a culture of Open Access on a campus.

Digital Libraries have led the movement by instituting Institutional Repositories for scholarly works and research data, but it has become increasingly evident that academic institutions must implement strategies for raising the awareness of Open Access and promoting the involvement of their academic scholars and students. It is no longer a question of whether or not to promote the open accessibility of these works among our academic community, but how best to do so.

This roundtable discussion will offer ideas, strategies, and thoughtful conversations on how to equip a campus with the resources it needs to promote and assist researchers in adopting Open Access. This panel will feature faculty; a graduate student; scholarly communications, institutional repository, and strategic projects librarians to provide a balanced perspective of Open Access implementation at one Texas institution.

Keywords
open access; policy; programming; repository services; intellectual property; open data

Session 4B (Travis I/II Room):

Visualizing Metadata
Violeta Ilik, Texas A&M University Libraries

Abstract
We used Viewshare to generate and customize unique, dynamic views of data about various collections. We created interactive maps, timelines, facets, tag clouds, and galleries that will allow users to experience our digital collections on a completely new level. We will report on the issues we faced while testing the Viewshare platform, and our suggestions for improvement that were sent to the creators of this interactive platform.

Keywords
viewshare; linked data; descriptive metadata

Understanding Large Digital Collections and Learning New Tools: The Texas Digital Newspaper Program Visualizations
Mark E. Phillips, Will Hicks, University of North Texas

Abstract
As digital library collections continue to increase in size it becomes necessary to use new tools and techniques to communicate and understand the rich content held in these collections to curators and end users. This presentation discusses the use of the D3 Javascript library to visualize and provide new insight to the Texas Digital Newspaper Program (TDNP) hosted by The Portal to Texas History as a case study. The collection contains over one million pages of Texas newspapers from the 1830’s to modern day covering over one hundred counties, and hundreds of titles. An overview of existing newspaper visualization projects will be presented as well as an explanation of how the presenters prepared the data for these visualizations using the publicly available TDNP OAI-PMH repository and the open source D3 Javascript library.

Keywords
newspapers; visualization; metadata; collections; digital libraries

Determining and Mapping Locations of Study in Scholarly Documents: A Spatial Representation and Visualization Tool for Information Discovery
James Creel, Katherine Weimer, Texas A&M University Libraries

Abstract
Theses and dissertations play a significant role in the scholarly literature and often refer to locations of interest or regions under study. Through geoparsing, which is the identification and disambiguation of place names, we have created a tool to generate interactive maps of the geographic locations referenced in theses and dissertations. Our visualization affords increased awareness of the numerous locations being researched and which departments and majors are studying each location. More broadly, the interface supports multidisciplinary research, student recruitment and faculty collaboration. Using geographic and gazetteer metadata and open source mapping applications, this tool provides researchers with serendipitous geographic and interdisciplinary connections. The beta version consists of several DSpace curation tasks to take a given ETD through each step of the metadata creation and mapping processes. Once the tool has suggested geospatial metadata for an ETD, the DSpace administrative interface allows curators to approve the suggested metadata values.

Our geoparser integrates various open-source tools as well as specialized heuristics to automate the name extraction and disambiguation tasks. We have employed the OpenNLP and Stanford NLP libraries for the name extraction task, and use the Geonames gazetteer as our source for referenced entities. A preliminary evaluation of the tool indicates an accuracy of 84% with regard to the disambiguation of names to specific Geonames IDs. Work toward improving the accuracy is ongoing.

The visualization component of the tool reads geospatial metadata as KML and can render the referenced locations in any of three map visualization options selected by the reader: OpenLayers, OpenStreetMaps and Google Maps. Once a site of interest is located on the map, the reader may select a link to the complete thesis or dissertation stored in the university’s instance of DSpace, our institutional repository. The long-term goal of this project is to extend the content to include all TDL ETDs for a widely used search mechanism.

Keywords
geoparsing; geocoding; toponyms; named entity recognition; name disambiguation; natural language processing


Session 5A (Austin Room)

Automated Archiving of DVD Content
Heriberto Nieto, Texas Advanced Computing Center; Anna Lamphear, UT Austin Libraries

Abstract
A little more than a decade after their introduction, DVDs are becoming obsolete. As a result collecting institutions must find a long term preservation solution for content in DVD format.

To address this challenge the Texas Advanced Computing Center (TACC) developed an automated workflow for transcoding DVDs using a tool built with open source software. Automated and potentially customizable, the tool allows for the creation of a series of files including a preservation quality file for long term storage, an ISO disk image, streaming files for access purposes, as well as technical and preservation metadata. The tool also applies a reference metric that allows for automated quality checks of the video files. The resulting output can be used as a Submission Information Package (SIP) suitable for ingest into a repository.

The University of Texas Libraries is testing the tool on its DVD collections, streamlining workflows for both preservation and access.  Several modifications have been made to the tool to support the production needs of the Libraries.  This presentation will highlight aspects of the development of the tool and will discuss the implementation and testing in reference to the resultant output files and metadata, the differences between the museum and library workflow application, and the overall usefulness of the tool.

Contributors on the project include:  Maria Esteva (TACC, UT Austin); Karla Vega (TACC, UT Austin); Heriberto Nieto (TACC, UT Austin); Bethany Scott (Charlotte Mecklenburg Library); Kertana Kumar (College of Natural Sciences, UT Austin); Vandy Henriksen (UT Libraries); Anna Lamphear (UT Libraries); Jennifer Lee  (UT Libraries); Wendy Martin (UT Libraries).

Session 5B (Travis I/II Room)

Integrating the Classroom and the Digitization Center: An Innovative Approach
Eric Ames, Baylor University Libraries

Abstract
Following several successful years of building increasingly larger, highly impactful digital collections, the Baylor University Digital Projects Group chose the Spring 2013 semester to take the next step in its development by creating a new graduate level course for the Department of Museum Studies called Technology and Outreach in Museums.Taught by Eric Ames, Curator of Digital Collections and a lecturer in the department, the course is an innovative approach to integrating classroom lectures and philosophical exploration of the topic with hands-on, intensive training on current digitization and outreach methods using the resources of the Electronic Library’s Riley Digitization Center. The purpose of the course is to give graduate students the opportunity to select materials from non-digitized archival collections; curate the materials for inclusion in a digital exhibit/collection; digitize the materials using the scanning equipment of the Riley Center; create a digital exhibit; and formulate a marketing plan to promote the exhibit to scholars, faculty and the general public.

This presentation will focus on the development, proceedings and lessons learned over the course of the semester. Ames will also present tips for institutions interested in following a similar path, including how to manage the interests and skill levels of 15 graduate students in a working digitization center, challenges to implementing technological solutions and students’ perspectives on the course. If possible, a graduate student from the course will be invited to attend and lend his/her perspective on the course.

Keywords
graduate education; digitization; outreach; course creation


Closing Keynote Address (Austin Room)

Making the Commitment to Preserve Our Digital Scholarship: Can we and will we? 
Michele Kimpton, CEO of DuraSpace, Project Director for DuraCloud

Abstract
Ensuring digital scholarship produced today is available for future generations will be seen as one of the key accomplishments or failures of our time if we do not get it right. There is an ever increasing amount of content produced natively in digital form, that if actions are not taken today to proactively manage and preserve the content we risk losing it forever. Digital data is being produced in record quantities  in many new formats ,which makes it daunting for any single institution to tackle preservation alone.  There are many projects and collaborations underway to begin to collectively work on finding viable, sustainable solutions across a broad spectrum of use cases. DuraSpace, as well as Texas Digital Library, are players in this pursuit to develop common, scalable solutions for preserving and providing access to scholarship. The focus of the talk will examine the landscape of preservation focused projects and discuss how each of us can and must take part in this process to ensure robust, diversified and sustainable solutions.  We can not leave the job of ensuring long term access of our digital scholarship to others, it is our problem to solve and our legacy to leave.

Keywords
digital preservation


 Session 6A (Austin Room):

Panel: Barn Raising the Digital Humanities
Tanya Clement, Assistant Professor, School of Information, UT Austin; Geneva Henry, Executive Director, Center for Digital Scholarship, Rice University; J. Lawrence Mitchell, Professor, Department of English and Director of Cushing Memorial Library & Archives, Texas A&M University; Ryan Hoover, Assistant Professor, English Writing and Rhetoric, St. Edwards University; Moderators: Sarah Potvin, Texas A&M University Libraries; Matthew Christy, Initiative for Digital Humanities at Texas A&M University

Abstract
Writing in a recent special issue of Journal of Library Administration dedicated to exploring “Digital Humanities in Libraries: New Models for Scholarly Engagement,” Miriam Posner references libraries and archives as formative to Digital Humanities (DH): “what we now call digital humanities grew out of a set of practices, and a community of practitioners, which themselves arose in libraries and archives.” This foundational relationship continues to shift and grow, some have argued, to the mutual benefit of academic DH and cultural heritage institutions (including galleries, libraries, museums, and archives) alike. Jefferson Bailey, writing for dh+lib, notes that “DH tools, methods, and technologies have the potential to help enhance and evolve a wealth of professional practices beyond academic … It is this ability to reinvigorate the work of non-academics, such as librarians, archivists, and collection managers, that has many of us in cultural heritage excited about DH as an emerging idiom within memory institutions.”

With the increasing popularity and growth of Digital Humanities, however, the question of how universities can and will support this emergent area asserts itself. In 2009, Christine Borgman observed a lack of basic infrastructure for DH and recommended: “Much work remains to build the scholarly infrastructure necessary for digital scholarship to become mainstream in the humanities. Humanities scholars must lead the effort, because only they understand the goals and requirements for what should be built. Librarians, archivists, programmers, and computer scientists will be essential collaborators, each bringing complementary skills.” Four years later, what does this infrastructure look like?

With this background in mind, our panel seeks to meet two objectives:

First, and simply, we aim to introduce TCDL attendees to a range of DH projects and initiatives underway in universities across Texas. To this end, we’ve invited a number of distinguished panelists– based in academic departments, libraries, archives, and information schools– engaged with a variety of DH projects and asked them to provide an overview of their work.

Second, we are interested in understanding and examining the structures and supports in place to enable DH collaboration, as well as those being built. In considering this, we’ll pay particular attention to the role of libraries and archives. What has institutional collaboration looked like for those engaged in DH activities from the perspective of libraries, archives, and academic departments? And what has facilitated or impeded this collaboration?

Keywords
interdisciplinary; research infrastructure; libraries; archives

Session 6B (Travis I/II Room)

It’s a Big World After All… The Pitfalls of Digitizing and Describing a Linguistics Project by Non-Linguists
Ramona Holmes, Krystal Schenk, UT Arlington Library

Abstract
A digitized collection of linguistics recordings made by Jimmy G. Harris, a former employee of the United Nations who worked for several decades in many far-flung parts of the world. The recordings are unique, capturing languages and dialects from speakers in currently politically unstable countries where field work is difficult if not impossible. Furthermore, since the recordings were made from the 1960s to the 1980s, they also represent a unique capture of these speakers during a specific time. Adding this digital collection to our institutional repository posed some unique challenges, including reel to reel capture of sound from materials that were decades old and metadata creation without a certain type of specialization.

Keywords
descriptive metadata; linguistics

Extreme Makeover: Digital Library Edition 
Melanie Cofield, Tarleton Law Library, UT Austin

Abstract
Featuring a digital library of early Texas legal documents built more than a decade ago and needing upgrades, this is a tale of renovation efforts requiring a sustained group effort. Librarians, archivists, students, and a programmer have worked together over time to evaluate the resource, identify practical fix-it strategies, and forge a more user-friendly and sustainable web resource for scholars, teachers and students interested in Texas legal history.

Keywords
cultural heritage; digital archives; special collections

Pasteur’s Own Hand: The Creation of a Digital Repository
Mira Greene, UT Medical Branch at Galveston

Abstract
The Truman G. Blocker, Jr. History of Medicine Collections at the Moody Medical Library at the University of Texas Medical Branch, Galveston, Texas, was awarded a NN/LM South Central Region Digital Preservation & Access (DiPA) Award to complete a digital repository of carefully selected material from the private library of Louis Pasteur (1822-1895). This project sought to dramatically improve user awareness and access to the personal and academic thoughts of one of the world’s most celebrated scientists. The Pasteur Collection held at the Blocker Collections contains 154 items including books, offprints, prints, manuscripts and unpublished handwritten letters by Pasteur which he maintained in his personal library. In a recent appraisal, the Pasteur Collection was noted as “easily the most significant in the United States and most likely the entire world outside of the material held by the [l’Institiut Pasteur] in France.”

Our presentation will focus on the award application process and lessons learned, in-house workflows, outsourcing of translation and transcription, creation of audio files, website design, metadata creation, digital repository creation in ContentDM, and exporting of records from ContentDM for batch loading in our TDL repository.

This project has been funded in whole or in part with Federal funds from the National Library of Medicine, National Institutes of Health, under Contract No. HHSN-276-2011-00007-C with the Houston Academy of Medicine-Texas Medical Center Library.

Keywords
digital preservation; descriptive metadata; collection management


 Session 7A (Austin Room)

From Custody to Collaboration: The Post-Custodial Archival Model at the University of Texas Libraries
Kent Norsworthy, T-Kay Sangwand, UT Austin Libraries

Abstract
Despite living in an age of ubiquitous access to digital information, scholars still struggle to access both the physical and digital primary sources needed for research and teaching. This can be due to limited access to physical primary sources (i.e. cultural heritage materials located in another country), lack of resources to make analog primary sources digitally accessible (i.e. limited funds and staff for digitization), or lack of a concerted effort to collect born digital materials (i.e. no clear institutional mandate).  The inaction around preservation and access can result in the loss of the materials themselves, for example to obscurity and obsolescence, as well as an impoverished historical record. In order to confront these threats while simultaneously supporting scholarly research at The University of Texas at Austin, the University of Texas Libraries has adopted a post-custodial archive model to address the collection, preservation, and access challenges for digitized and born digital archival materials. This approach also allows us to secure valuable scholarly resources for The University of Texas at Austin and the global community and to foster deep collaborative relationships with campus faculty and academic units as well as with our partners around the globe, all in close alignment with the strategic priorities of the Libraries and the University. This presentation will explore the background and rationale that underlies the post-custodial archive model and how this model has been implemented through digital projects such as the Guatemalan National Police Historical Archive, Primeros Libros de las Américas, and the Human Rights Documentation Initiative. We will discuss these initiatives as part of broader efforts to redefine the role and identity of the research library as a central component of teaching, scholarship, and resource to 21st century learners, and as an exemplary activity of the University’s commitment to “scholarship and education that advances the social good.”

Keywords
non-custodial archives; digital scholarship

The Digital Preservation Network at UT Austin
Chris Jordan, Texas Advanced Computing Center

Abstract
The University of Texas at Austin is a key participant in the Digital Preservation Network(DPN), an organization that intends to provide a common vehicle for long-term preservation of the most important scholarly products of member institutions. This presentation will provide a brief introduction to the goals of the Digital Preservation Network and its relationship to regional preservation efforts including TDL. We will discuss the developing technical architecture for data replication in DPN and the role of the Texas Advanced Computing Center and TDL in developing and deploying that architecture. We will briefly describe the design principles behind the architecture and the manner in which the replication technology is overlaid on existing ingest and storage mechanisms.

Keywords
digital preservation; data infrastructure; technical architecture

Session 7B (Travis I/II Room)

Vamp it Up and Make it Better: Auditing and Upgrading Metadata in the UH Digital Library
Santi Thompson, Annie Wu, University of Houston Libraries

Abstract

Since 2009 the University of Houston Libraries have expanded access to the rare and special collections through the University of Houston Digital Library (UHDL). Three years and nearly 50,000 digital items later, UHDL has the potential to be a powerful resource for researchers of all kinds. To facilitate this kind of research, users rely on UHDL metadata to be robust, reliable, retrievable and sharable. While the current state of metadata in the digital library allows for browsing and searching, missing and inconsistent metadata restrict both discoverability and interoperability. As a result of these barriers, the Metadata and Bibliographic Services Department is initiating a project to audit and upgrade the existing metadata produced for the UHDL.

This presentation outlines the audit and upgrade process to date. It provides an overview of our methodology, including focus group interviews, data inspection, and benchmarking. It also identifies key recommendations and strategies for upgrading the existing UHDL metadata, including guidelines for future UHDL metadata creation. The presentation is designed to offer suggestions to other institutions interested in conducting a metadata audit. It also encourages audience members to share other tools and techniques to implement an audit.

Keywords
descriptive metadata

From Mishmash to Mesh: Subject Headings Changes for a Collection
Cameron J. Kainerstorfer, Heather Perkins, UT Southwestern Medical Center Library

Abstract
What is the half-life of a subject heading? Transfer between systems, input errors, workflow and staff changes all contribute to decay in subject heading quality. This presentation will cover highlights from a recent project to review and modify subject headings for the electronic theses and dissertations collection within the UT Southwestern institutional repository. Project work was primarily conducted by a UNT graduate school library science student as part of an semester-long internship in 2012. Developed by the repository administrator and metadata manager, this also served as a springboard for future intern projects.

The presentation will also highlight:

  • Targeted quality improvement through a small, manageable project,
  • Cooperation between different library departments,
  • Access issues for repository work in DSpace conducted by non-affiliated personnel (off-campus access; institutional accounts; etc.),
  • Needs of particular interest to a medical institution (MeSH vs. other subject headings; embargoed or limited content).

Keywords
repository management; subject headings; medical subject headings


Workshops

ETD Embargoes: A Comparison of Institutional Policies and Practices
Laura Hammons, Stephanie Larrison, Geneva Henry

2:15 – 3:45 PM | Old Pecan Street Room

Abstract
ETD embargoes policies and practices vary widely among institutions. Although institutional websites often make embargo policies quite clear, the practices that support those policies is less so.  Even more interesting, but less obvious, are the history and rational surrounding the development of embargo policies, as well as how exceptional cases and appeals for extensions, redactions, and permanent holds are handled. The Vireo ETD Submission and Management System is a flexible tool that can accommodate variations in ETD embargo policies and practices to support the needs of the institution. In this birds of a feather session, you will have an opportunity to learn the what, why, and how of embargo policy and management (both inside and outside of Vireo) from three different institutions.

Keywords
electronic theses and dissertations; ETDs

Command-Line Image Manipulation with ImageMagick
Jeremy D. Moore, University of North Texas

2:30 – 4:30 PM | Travis I Room

Abstract
This workshop will introduce command-line image manipulation to its participants using the open source software ImageMagick. While many are familiar with image manipulation software using graphical user interfaces, or GUIs, such as Adobe Photoshop, GIMP, and Pixelmator, there is less awareness of programs built on command-line interfaces, or CLIs. In a CLI, the user issues commands to the program by typing lines of text. This workshop is targeted at first time and novice CLI users and is structured to provide a strong foundation of skills to build upon.

Topics covered will include:

  • History of the command-line interface
  • Navigating folder hierarchies on the command-line
  • The “find” command
  • Installing ImageMagick in Linux, Mac OS X, and Windows XP/Vista/7
  • Creating an image by typing a single line of text
  • Convert versus mogrify
  • Image manipulations possible in ImageMagick
  • Creating web derivatives from master image files
  • The Identify command
  • Applying and removing image compression

Due to the nature of this workshop, participants will need to bring a laptop on which they have Administrative rights.

Keywords
ImageMagick; command-line

New Features in Vireo 2.0
Micah Cooper

3:45 – 5:00 PM | Old Pecan Street Room

Abstract
This presentation will review the results of the development effort led by Texas A&M University with the support of the Texas Digital Library in the Fall of 2013 on the Vireo project. Vireo is an open source turnkey Electronic Thesis & Dissertation (ETD) submission system. Vireo supports a student submitting their thesis after successfully defending it, through the review process, and to publication in a repository. With the 2.0 release there are many new features and abilities that offer greater flexibility on how Vireo can be deployed and customized to the needs of individual institutions.

New Fields

  • Language
    This field allows the student to select from a pre-defined list of languages. The primary motivation for this field is for compatibility with ProQuest as they require the language of an ETD to be identified.
  • Subjects
    The field allows the student to select up to three subjects from a pre- defined list of subjects. As with the previous field the primary motivation for this field for compatibility with ProQuest. The pre-defined list of subjects is defined by ProQuest but may be modified by updating the Excel spreadsheet at /conf/proquest/umi_subjects.xls
  • Published Material
    This field allows for students to describe which sections in their ETD may have been previously published. The field first appears as a checkbox, but when checked a second component is exposed asking the student to identify the material that has been previously published. This enables review staff to identify submissions which may need an additional copyright review before publishing.
  • Defense Date
    This field allows for students to specify when they successfully defended their thesis or dissertation. This date is an important milestone of the process and useful to be recorded with the other metadata about the submission.
  • Program
    This field operates like the College and Department fields have always worked. The Vireo Manager is able to create a list of available programs from which the student chooses. This field allows for different review process for different programs.
  • Dynamic Committee Member Role Types
    Previous version of Vireo had a single check box to specify whether a committee member was the chair or co-chair of a committee. This did not allow for additional role types such as Department Head, or Directory of Research to be specified on the student committee. Further more, some schools do not use the “chair” language instead they follow “Supervisor” or another nomenclature. This new field allows the Vireo Manager to defined role types that are available at each degree level.
  • Administrative and Source files
    Previous versions of Vireo allowed the student to upload a primary PDF manuscript, along with any number of supplementary documents. In addition to these categories Vireo 2.0 introduces two additional types: administrative files and source files. Administrative files are things like signed authorization forms or other items required by local workflow. The administrative files will not leave Vireo. Source files are the original word or LaTex documents used to generate the primary PDF manuscript. These source files are commonly deposited in to the Institutional Repository but not made publicly available. They may be used in the future to enhance the preservation potential of the items being deposited.
  • Reviewer Notes
    This field is not available to students. Instead it is only available to Vireo Reviewers and Managers. The field is visible under the view tab just above the table of action logs. This is a free-form textarea where reviewers can leave private notes about the submission and it’s state.
  • Institutional Identifier This field is available when using Shibboleth or some other protocol to integrate with your central campus IT infrastructure to capture a student’s unique ID number. This number is displayed, although is not editable, within Vireo. Enabling easier integration with local campus procedures.New Features
  • ProQuest Export
    The ProQuest export format has been enhanced supporting the new ProQuest centric fields that have been incorporated into Vireo. We are still awaiting word from ProQuest for them to validate the new format.
  • Remove item individually from filter lists
    This feature helps make the new batch features below more useful. Batch features work based off of filter searches. Crafting a filter search to target a specific population of submissions will sometimes leave one or two that just are not ready to be updated. Using this feature the reviewer can remove individual submissions from the filter, before running a batch operation on them.
  • Batch send email to students
    This allows mass emails to be sent out to students about important information. You may send an individual email or select from among the email templates. All the standard variables are available.
  • Batch assign submissions
    This batch feature allows for a Vireo manager to assign batches of submissions to a particular individuals.
  • Batch status change
    This batch feature allows for batches of submissions to change state when ready. Using this feature the Vireo Reviewer can queue up a set of submissions and move them to the next state all at once. Using this feature allows batch publishing of submissions, and if you move the submissions to the canceled state you may also batch delete submissions.New Settings
  • ProQuest settings
    Several new settings have been added to support the ProQuest export format. The new fields are found under the Application Settings tab and include: ProQuest Institution code, toggle for Open Access publishing, and a toggle for Indexing by Search Engines. This makes configuring the export format much easier by end users.
  • Bulk updates
    Many of the lists under the Configurable Settings tab now have bulk operations. All lists may be sorted or alphabetized, bulk delete is also possible, and some fields have a bulk add feature.
  • Toggle all data field: Disabled, Optional, or Required
    This is a major new feature! All fields (not just the new fields added above) may be configured turned on or off individually. Each field can be set to Disabled, Optional, or Required. Disabled fields are not collected from students, while Optional fields will be displayed to students but they will not be required to provide the data.
  • Customize field labels
    Each field may now have it’s label customized individually. This allows for some institutions to repurpose fields for slightly different data. The name of these fields are only modified for students, else ware in the administrative interface the field will still display it’s original pre-defined name.
  • Customize field help text
    Each field may now have the helptext customized individually. This allows the Vireo Manager to add local institution instructions to particular fields.
  • Customize sticky notes
    The sticky notes shown on the right column during the student submission process may be updated. Vireo Managers may update the text, or add / remove notes.
  • Degree codes
    Formal degree codes (i.e. Ph.D) may be specified for individual degrees under the Application Settings tab. This is primarily to support the ProQuest export format.
  • Color and branding settings
    The color and branding used on the student side may be easily customized to minimally match the branding of the individual institution.

Poster Presentation Abstracts

Building an Online Data Management Plan Tool
Michele Reilly, Anita R. Dryden, University of Houston

Abstract
Following the 2011 announcement by the National Science Foundation (NSF) that they would begin requiring Data Management Plans with every funding application, the University of Houston Libraries explored ways to support our campus researchers in meeting this requirement. A small team of librarians built an online tool using a Drupal module. This tool includes informational content, an interactive questionnaire, and an extensive FAQ to meet diverse researcher needs. This easily accessible and locally maintained tool allows us to provide a high level of personalized service to our researchers.This poster will highlight the need for such a tool, the steps taken to create the tool and outcomes from researcher involvement.

Keywords
data management plans; NSF; research tools

A Comparison of Features Among ETD Submission Systems
Stephanie Larrison, Texas State University

Abstract
My poster will provide a comparison of features offered by different Electronic Thesis and Dissertation submission systems in an easy to understand and visually appealing table. The intent of this poster is to offer a quick snapshot of what ETD systems can do to help potential and current users determine which products may best fit their needs. It will be of interest to those involved in the policies, management, publication or curation of theses and dissertations at an institution. I will be comparing Vireo, the open-source Electronic Thesis and Dissertation Submission and Management software developed by the Texas Digital Library, and Proquest’s free web-based ETD Administrator. I also plan to add additional systems to the comparison chart as information is gathered.

Keywords
ETD; ETD submission system; dissertations; Vireo; ProQuest ETD Administrator

Evolutions: Digitization vs. Digital Born Photography in Urban Architecture
Scott A. Holmes, Kris Swenson, University of Texas at Arlington

Abstract
The Fort Worth Evolutions project takes original material by Ft. Worth commercial photographer W.D. Smith and creates a modern view of those photographs. Images taken in the 1920s-1950s are retaken from the same angle in this century and compared to urban architecture of the early to mid 20th century. This project leverages the process of digitization and the art of modern digital photography to create a contrast of urban development over the last 90 years.

Keywords
digital preservation; digital photography; urban planning

Current Practices in Quality Assurance for Web Archives
Brenda Reyes, University of North Texas

Abstract
Web archiving is the process of storing and maintaining Internet resources (such as websites) to preserve them as a historical, informational, legal, or evidential record. The process involves three stages: selecting relevant resources for preservation, gathering and storing them, and providing for their access. In recent years, it has become an increasingly common practice in libraries around the world, as national libraries, such as the Library of Congress and the National Library of Australia, seek to preserve their national digital heritage. Many universities have also begun archiving the web, usually to create subject-specific collections of web sites that supplement their existing print and digital collections.

Within the web archiving community, a step that often goes unmentioned is the Quality Assurance process (QA), which measures the quality of an archived site by comparing it to a standard that must be met. Currently, each institution conducts its QA process independently, using a myriad of different standards and software tools. The result is a considerable knowledge gap: practitioners do not know if and how their peers are conducting a QA process and generally do not share this information. Consequently, there are no agreed-upon quality standards or processes.

The study presented here attempts to address this information gap in the web archiving community. To this end, we investigated how several institutions conduct their quality control processes. It is worth noting that quality control procedures are often not publicly available and not thoroughly documented, if at all. Much of the information present here has been obtained from reports, electronic communications, listserv discussions, and interviews with staff involved in the QA process. The results we obtained led us to design a survey instrument to gather information in a more thorough and structured manner. The results from this survey are included here.

Keywords
web archives; quality assurance; digital preservation

Uncovering the Mysteries of Metadata Harvesting: Optimizing Digital Library Content for Summon Discovery Tool Access 
Kelsey Renee Brett, Santi Thompson

Abstract
One of the most valuable attributes of a discovery system is its ability to harvest locally owned digital materials and make them discoverable through a single search box along with the rest of the library’s collection. At the University of Houston Libraries, we sought to make local digital content more findable in our discovery system, Serials Solutions Summon.  In order to do this, we had to better understand the way digital library content in CONTENTdm mapped to Summon and work within the constraints of each commercial product to achieve our desired outcome.  While investigating the ways Summon harvested our digital content, we discovered that a majority our digital material was being described by a catch-all content type called “Archival Material” and the extensive efforts made by our metadata librarians and staff to describe digital items were being lost in the harvesting process.  We also discovered that there was no way to limit a search to exclusively digital library materials. This motivated us to explore the details of the Summon harvesting process and develop solutions to enhance the discoverability and description of digital materials within Summon.

Our poster will outline our efforts to increase the findability of digital materials, including our research of the OAI mapping process that was happening behind-the-scenes at Serials Solutions, our discussions with them about the possibilities of modifying the default mapping that was currently in place, and the strategies we developed achieve our desired outcomes.  We will discuss the challenges and successes that we had throughout the process, and how we modified our local content to work well with our discovery product.

Keywords
discovery systems; metadata; metadata optimization; digital harvesting; Serial Solutions Summon; digital collections

CONTENTdm to DSpace – Why?
Joy Marie Perrin, Texas Tech University

Abstract
This poster outlines TTU’s decision to move all our digital library content from CONTENTdm to DSpace. The poster will include statistics and results from some of our tests that showed that for us CONTENTdm was not serving our users’ needs or our stated goal of making our collections as accessible as possible. This poster will discuss why DSpace looks like it will serve our needs better.

Keywords
content management system

Digital Curation Micro-Applications: Digital Lifecycle Management with AutoHotkey
Andrew Weidner, Robert John Wilson III, Daniel Gelaw Alemneh

Abstract
Maintaining usable and sustainable digital collections requires a complex set of actions that address the many challenges at various stages of the digital object lifecycle. Digital curation activities enhance access and retrieval, maintain quality, add value, and facilitate use and re-use over time. Accordingly, the University of North Texas (UNT) Libraries develop and adopt tools, workflows, and quality control mechanisms that streamline production and quality assurance activities. This poster demonstrates open source software tools coded with AutoHotkey that the UNT digital libraries group has developed for use during the pre-ingest and post-ingest stages of the digital resource lifecycle.

AutoHotkey is free and open source software for the Windows operating system which helps digital curators to create customized micro-applications for digital lifecycle management. Developers write scripts that embed multiple keystrokes and system commands in a single key combination, or hotkey. In addition to a portable version for script developers who frequently change workstations, AutoHotkey allows users to create executable files from source scripts for use on computers that do not have AutoHotkey installed. The AutoHotkey scripting language supports programming constructs (e.g., variables, loops, conditionals) and dynamic graphical user interfaces (GUIs).

Simple AutoHotkey scripts are particularly useful for large projects that require repetitive file management actions. For ongoing batch processing activities, more complex scripts improve efficiency for those who are unfamiliar with command line equivalents. These scripts allow digital curators to perform automated file management and data entry tasks within the operating system’s GUI. The Digital Libraries Division at the UNT Libraries has constructed a set of tools that bundle complex AutoHotkey scripts together in a convenient GUI. This poster demonstrates how AutoHotkey micro-applications streamline workflows for the creation and maintenance of digital objects and their associated metadata.

Keywords
digital lifecycle management; digital curation; programming; AutoHotkey

“This is totally going on our blog.” Using WordPress and Edublogs to Enhance Access to Digital Collections
Eric Ames, Baylor University

Abstract
From its first post on November 9, 2011 to the present, the Baylor University Libraries Digital Collections Blog has been access 10,630 times by users around the world. Increasingly, the blog serves as a major entry point into the collection, with users’ Google searches leading them to its posts detailing everything from in-depth looks at a particular digital collection to professional musings and analysis of the processes behind the creation of digital collections.

This poster presentation will provide an in-depth look at how the Digital Projects Group uses the blog to achieve a number of goals, such as:

  • Providing context for collections
  • Establishing a resource for small museums and archives to receive information on digitization trends and processes
  • Presenting unique stories and items from the Baylor University Libraries Digital Collections
  • Serving as a central clearinghouse for information related to the DPG

The poster presentation will be of interest to institutions that are considering starting a blog in conjunction with their digital collections; institutions that are currently using a blog but are looking for new ways to utilize them; and anyone interested in how to mine existing digital collections for stories to present via a blog.

Keywords
outreach; contextual research; collection curation

The Development and Implementation of an Interdepartmental Digitization Workflow
Valerie G. Prilop, R. Niccole Westbrook, University of Houston

Abstract
The science of creating workflows that involve multiple departments within a large organization can be challenging, especially when the projects handled by the workflow are complex and involve a number of stakeholders. Such a project to revise a workflow dealing with the digitization of unique library materials in a university library was undertaken by the presenters.

Libraries and similar institutions are increasingly embarking on digitization projects that involve a number of participants and stakeholders, but creation of workflows to complete these projects is not a topic that has been widely discussed or reported on. This poster will use the project as a case study for developing and implementing a digitization workflow. Among the topics to be addressed are the interdepartmental nature of the workflow, the move toward and impact of more written documentation about digitization projects, and the challenges associated with shepherding the workflow through a subsequent major departmental transition.

Keywords
workflows; project management; digitization

The Denton Declaration: An Open Data Manifesto
Spencer D. C. Keralis, Shannon Stark, University of North Texas

Abstract

On May 22, 2012 at the University of North Texas, a group of technologists and librarians, scholars and researchers, university administrators, and other stakeholders gathered to discuss and articulate best practices and emerging trends in research data management. The resulting document, The Denton Declaration, bridges the converging interests of these interest groups and promotes collaboration, transparency, and accountability across organizational and disciplinary boundaries.

This poster presentation will describe the process of developing the Declaration, crediting collaborators and participating institutions; will discuss why we chose the genre of the Manifesto (rather than a more traditional report), reproduce key principles of open data from the Declaration, illustrate the international network of cosigners that have championed the Declaration, and invite the TCDL community to join us in advocating these principles throughout the academy.

Keywords
open access; open data; manifesto

Implementation of a New Data/Time Standard in Digital Library Metadata
Hannah Tarver, University of North Texas

Abstract
Three years ago, the University of North Texas (UNT) Libraries Digital Projects Unit (DPU) chose to switch from the International Standards Organization (ISO) date standard to the Library of Congress’ Extended Date Time Format (EDTF) as the authorized format for metadata in The Portal to Texas History and the UNT Digital Library. Since the standard was still being drafted at the time, complete and appropriate use of the EDTF in the digital library system has been inconsistent. Now that the EDTF is somewhat finalized (though still in draft state), the DPU has started implementing the date format in a more organized fashion, including the preliminary use of a validation program to flag dates that do not match the specifications.

The EDTF is a date standard specifically intended to be machine-readable rather than formatted for usability by people, which means that there is occasionally a disconnect between the valid representation of dates according to standards and the ways in which partner institutions have previously expressed complex dates. This presentation would discuss the benefits and challenges of implementing the EDTF, such as: the pros/cons of choosing EDTF over ISO standards; difficulties in reconciling machine-readable date strings into dates that are understandable for users; challenges of fitting library-centric date paradigms (such as the use of square brackets) into a different standard; and making local cataloging decisions about how to implement the standard and at what level.

Keywords
metadata standards; date-time standards; metadata entry; metadata schemas

Scanning TRAIL Project Technical Reports: A Workflow for a Large-Scale Collaborative Digitization Effort
Hannah Tarver, University of North Texas

Abstract
As part of the Technical Report Archive and Image Library (TRAIL) project, the University of North Texas (UNT) Libraries Digital Projects Unit (DPU) has been working with the University of Arizona (UA) during the last three years to digitize technical reports that have oversized fold-outs and which are of non-consistent. Project managers in the DPU have developed a straightforward workflow for handling the large volume of technical reports for this collaborative effort. The poster would illustrate the process we use to digitize the reports including inventory, scanning, processing, metadata, and upload. The same workflow could be used by other institutions to manage similar large-scale digitization of text objects.

Keywords
digitization; workflows; digital preservation

Collection Development for an Institutional Repository through Collaborations between Departments
Wilhelmina Randtke, Brian Detweiler, St. Mary’s School of Law

Abstract
Poster presentation: St. Mary’s University School of Law’s Sarita Kenedy East Law Library recently launched an institutional repository. The School of Law and law library had no preexisting digital collections. In order to quickly acquire appropriate content, the law library focused on locating born digital materials, such as School of Law publications, which had not previously been formally archived. The law library also attempted to identify digitization performed as part of routine library operations, and to assess digitized material for long term archiving. The law library was able to quickly and efficiently build an online collection for the repository by collecting preexisting born digital material, and assessing for inclusion material provided digitally to professors after conversion from legacy formats such as microfilm, and audiotape. This poster presents on how interdepartmental collaborations provided the framework to populate a digital collection in the absence of resources or equipment dedicated specifically to digitization.

Keywords
collection development; collaboration; born digital; workflows; institutional repository; digital library

Mapping the Southwest Project: Putting the Region’s Maps Online 
Daniel Gelaw Alemneh, Jerrell Jones, Cathy Hartman, Mark Phillips, Ann Hodges, Ben Husman, University of North Texas

Abstract
The University of North Texas Libraries and its partner, the University of Texas at Arlington’s Special Collections, are working on a 3-year (2010 to 2013) collaborative “Mapping the Southwest” project, sponsored by a National Endowment for the Humanities (NEH) We the People grant. For this project we are digitizing 5,000 historically-significant maps and more than 80% are already processed and available online for free public access through The Portal to Texas History: http://texashistory.unt.edu/explore/collections/UTAM/browse/

All of the maps digitized for this grant meet the UNT metadata requirements, which means that all are Open Archive Initiative-Protocol for Metadata Harvesting (OAI-PMH) harvestable, and are interoperable or compliant with widely adopted standards (such as Dublin Core, MARC, MODS, and PREMIS).

On the Portal, users can find materials using a basic search, an advanced search, or through multiple browse interfaces. Digital object display on the Portal has been optimized for indexing by Google and other search engines. Overall, 44% of all traffic to the Portal is referred from search engines such as Google, MSN, Yahoo, etc.  To attract lay people to the good quality resources in our collections, we have also added extensive external links into our digital materials from Wikipedia, which has resulted in significant increases in usage of the collections.

Among other digital libraries functionalities, when users find maps in the Portal, the maps manifest within a zoom feature that does not require any special downloads or software.  The zooming software works with Flash Player, which is ubiquitous, reaching 99% of Internet users.  The zooming feature allows users to see every detail of the map; grab and move the map with a click and pull mouse motion; or move the area of selection by dropping and dragging the selection area on the object icon.

As we approach the project completion phase, this poster describes the lessons learned, and project impact not only in terms of showcasing the cartography of the region, but also in promoting best practices and advancing the capacity of academic libraries to reliably curate, preserve, and provide seamless access to such wide-format items to the diverse global user community.

Keywords
digital libraries; maps; wide format items; open access

Looking through the Eyes of Texas
Tanya N. Brassie, Ann Serrano, University of Texas at Austin

Abstract
This poster will describe the creation and advertisement of a virtual historic walking tour for the University of Texas at Austin using HistoryPin, a publicly available interactive website based on linking or “pinning” historic images to an interactive map. The project demonstrates how HistoryPin can be a powerful tool for creating meaningful, accessible exhibits; anyone with the Internet can explore the University’s rich heritage while learning more about its present.

Since its founding, The University of Texas has witnessed 130 years of dramatic change. With the virtual tour, this change can be re-experienced through the juxtaposition of historic images of student life and campus architecture with corresponding contemporary images. Historic materials were provided by The University of Texas at Austin’s Dolph Briscoe Center for American History, which serves as the University’s archive, housing thousands of historic images, books, ephemera, and manuscripts. The curation process for the tour included selecting materials from the Briscoe Center’s vast holdings for digitizaton, assigning metadata to newly digitized items, uploading photos to HistoryPin. After creating the exhibit using HistoryPin’s tour function, it will then be promoted and presented for use during Explore UT, the University’s open house event.

Keywords
descriptive metadata; digitization; archives; ephemera; social media

Google-izing Projects: Managing Large-Scale Digitization Projects via the Cloud 
Darryl Stuhr, Timothy Logan, Baylor University

Abstract
The Riley Digitization Center at Baylor University manages many projects throughout the year, digitizing thousands of items and hundreds of thousands of pages. To handle high-volume throughput, track project status, manage source materials, and ensure a high level of quality, the Digital Projects Group develops and maintains a distributed project infrastructure that supports extensive and complex workflows. Managing many projects down to the item-level through an infrastructure built to handle quality control, data flow, multi-format processing and preservation, with a wide variety of staff including interns, graduate and undergraduate students, project-specific temps, and a few full-time staff, requires the use of robust project management tools available for reporting and data entry at all digitization and processing workstations. Unable to find a single tool that met all of our needs, we began to utilize a collection of tools, many of them free, such as Google Docs (Spreadsheets and Documents), Linux utilities, Evernote, and DropBox. This presentation will address Baylor’s implementation of the variety of tools used to manage digitization projects at the Riley Digitization Center – lessons learned and opportunities for growth. We hope that the tools and procedures will help others build a framework of inexpensive tools to organize and manage digitization projects large and small.

Keywords
digital preservation; project management; digitization; workflow

On the Road Again: Mapping a Texana Collection
Ann Serrano, Stefanie Lapka, University of Texas at Austin

Abstract
Digital mapping and charting holds great potential for repositories. Mapping print collections provides visualization for digital exhibits and context for remote users. This poster presentation will showcase a prototype digital map of the Frank Caldwell Collection, 1778-1941, which contains approximately 3,000 rare books, pamphlets, manuscripts, and other printed ephemera about Texas, collected by Austin-based traveling salesman Frank Caldwell. Maintained by the Briscoe Center for American History, the Caldwell Collection represents a comprehensive scope of items published in and about Texas between the early 18th and the  mid-20th centuries.

The Caldwell Collection is a geographically cohesive collection of significant size with extensive bibliographic records, which makes it an ideal selection for this prototype. Using the online application Topo.ly, the map is populated with metadata extracted from the bibliographic records which were created using AACR2R, LCC, and LCSH descriptive standards.

Keywords
geographic information systems; descriptive metadata

When Too Many Cooks Do NOT Spoil the Broth: Selecting and Implementing a New Web Scale Discovery Tool for a Large University Library
Jane Fleming, April Kessler, Jade A. Diaz, Colleen Lyon, Sara Snow, University of Texas at Austin

Abstract
In early spring of 2011, the University of Texas at Austin Libraries convened a working group to recommend a discovery tool to be implemented on the Library’s Web site. The group’s broad membership of user services staff, technical services staff, librarians and classified staff and branch and main library staff, reflected the desire that this search function provide patrons the “Google-like” single point access to the entirety of the University’s collections and databases that our patrons now expect. A second working group was convened to implement the tool chosen. The second group included a few core members of the first and was extended to include staff qualified to consider the implementation’s information architecture, the user (patron) experience and its integration with the Libraries’ existing technology and bibliographic processes. The new search tool, dubbed “scoUT”, was fully launched in August 2012, and has been the most successful rollout of any new service at the Libraries.

The poster presentation will cover the process to the present day.  Both teams began with research into best and recommended practices for their tasks, and an evaluation of peer institutions’ experiences and implementations. Both formally sought the input of library staff and of users. The first working group developed an extensive evaluation tool for discovery tool products, evaluated the tools and made a recommendation from among the final candidate tools, and contributed to the development of an RFP and to negotiation of the final contract for the tool.

The second working group guided the customization of the discovery tool’s capabilities to the library’s needs, devised a continuing process for updating bibliographic metadata on the libraries’ holdings and subscription services, and designed the incorporation of the discovery tool into the Libraries’ Web site. The group also coordinated the marketing of the tool and the development of training in its use for library staff and the Libraries’ user community. Concurrently to other work, group members created and carried out a usability assessment plan.  Heuristic evaluation and usability testing of other libraries’ implementations of the same product, an analysis of analytics and search logs for the existing UT interface, and an assessment of technology constraints informed the information architecture and design. The prototype of the scoUT search interface and the final scoUT search interface were tested and assessed between May and August 2012 in an iterative design process leading to the version in full release. Months of staff training preceded the full launch. Feedback on staff training prompted subsequent tutorials in customizing scoUT training for different user groups. We continue to collect feedback on scoUT to enable us improve service to our users. The University of Texas Libraries’ rollout of scoUT exemplifies what can be accomplished when groups of individuals who collectively have broad knowledge from across the library structure work together.

(Presenters from both working groups will be at the poster reception to field questions.)

Keywords
discovery tools; Web scale discovery; information technology projects; technology integration; heuristic evaluation; usability testing; user interface design; library instruction; academic library

Resource Description & Access: Now or Later? 
Ann Marchock, David Melanson, University of Texas at Austin

Abstract
Cataloging & Metadata Services at the University of Texas Libraries developed a series of extensive, in-depth training sessions in an effort to prepare staff for the transition to RDA. This poster will illustrate that entire process from the beginning point of training the trainers through the various sessions offered and tailored to the various audiences within the cataloging and metadata communities on UT campus.

Keywords
descriptive metadata

Metadata Quality in Texas Woman’s University’s Digital Collection
Amber L. Bookman, Texas Woman’s University

Abstract
Accessibility to information despite physical proximity is now an expectation of many students and professionals.  However, few appreciate what it takes to translate physical media and archival items into digital representations that can be stored, searched and retrieved virtually. A collection developed around a well-formed structure and metadata schema is necessary to address the complexities of this representation issue. However, the actual content and how it is entered will ultimately determine the usability of a digital collection. The Woman’s Digital Collection curated by Texas Woman’s University Libraries is one case that exemplifies some of the challenges of creating quality metadata records for a rapidly expanding collection. Inconsistencies within this one collection are typical of those that are likely to occur elsewhere as the demand for digital content increases. In identifying the issues that adversely affect the completeness, accuracy and consistency of the information populating these records, solutions for avoiding these problems emerge that can be applied to future projects.

Keywords
digital collections; metadata; data representation; data curation & management; repository management

@InstitutionalRepository How Do I Preserve Internet Ephemera? #Twitter #Wordpress
Franny Gaede, University of Texas at Austin

Abstract
University repositories are tasked with collecting and preserving the intellectual output of their institutions. Colleges and departments are adopting social media to connect with alumni, participate in disciplinary conversations, and encourage community engagement. These ephemeral communications must be captured and preserved to ensure a comprehensive record of the university’s scholarly output. Developing a standard for ingesting and describing these materials poses a novel challenge.

This poster will present, as a case study, the specific challenges faced by the University of Texas Digital Repository when preserving the Department of American Studies’ Twitter feed (@AmStudies) and WordPress blog (AMS::ATX). It will recommend best practices for preservation and access, including content capture, file formats, and metadata standards.

Keywords
digital preservation; social media; repository management; metadata standards; twitter; wordpress

The Monster MeSH! – Taming Medical Subject Headings for an ETD Collection
Cameron J. Kainerstorfer, Heather Perkins, Misou DeWeese, UT Southwestern

Abstract
In 2012, the repository administrator and metadata manager developed a semester-long project to review and modify subject headings for the electronic theses and dissertations collection within the UT Southwestern institutional repository. This also served as a test case and springboard for future intern projects.

The focus of this poster will be the challenges encountered during the project, how those challenges were met for this project, and other options for future consideration.

Poster highlights:

  • Schedule coordination with larger intern program (rounding, departmental visits, etc.)
  • DSpace training program for intern
  • Issues with access for a non-institutional account
  • Remote work and file transfers
  • Project documentation: methods and motivations
  • Medical institutions and special needs (MeSH, HIPAA, and more)

Keywords
repository management; subject headings; retrospective quality improvement