Restrategizing Bibliographic Services and the One Good Record | |
John J. Riemer | |
John J. Riemer is Head, UCLA Library Cataloging & Metadata Center; jreimer@library.ucla.edu |
A whirlwind of reports, product announcements, and spirited discussions swept over the library landscape in 2006. Lorcan Dempsey’s blog announced the posting of the University of California Bibliographic Services Task Force (BSTF) report and the debut of the new Endeca-based online catalog at North Carolina State University, which exemplified many of the BSTF recommendations.1 The Library of Congress–commissioned Karen Calhoun report followed a couple of months later.2 The Library of Congress (LC) described a priority shift to “access to content rather than access to description.”3 LC’s discontinuation of series authority work was the first change in bibliographic control along these lines.4 At least one ILS (integrated library system) vendor announced development of a new product in response to many of the shortcomings of the existing systems. Among other benefits, Ex Libris’s Primo is expected to permit unified searches of material traditionally siloed in separate databases and enable users to access library materials from within course-management systems and institutional portals.5
Much in these reports and developments was both exciting and disturbing to library staff. In the words of one colleague reading BSTF’s interim report, “I must say it really shook up my world when I read it. Not that there were any surprises—it was very much mom-and-apple-pie stuff that has been said before—but somehow you voiced it in a more compelling and urgent way.”6
These events and the subsequent discussion stemmed from an earnest look at a question reminiscent of Freud, “What do users want?” The following assertions have been made on behalf of users.
- Users want simplicity and immediacy. Users are accustomed to a single search box on an entry screen, without an obligation to categorize search keywords.
- Users want the search interface to be intuitive; this does not mean they want the response to their searches to be simple-minded. On the contrary, as Kautzman and Ryan have noted, “An intuitive interface is not by definition ‘dumbed down’ or anti-scholarly.”7
- Users benefit from prevention of dead-end searches, provision of enhancements like tables of contents and cover art, and navigation assistance that groups results and enables jumps to “more like this.”
- Users want one system to search, to cover a wide information universe. They frequently do not know about the separate silos that exist and what is in each; even when they do, having to search them one at a time is unwelcome.
- They expect immediate access to full-text online resources wherever possible. For other materials, they want a full range of fulfillment options, not restricted to those services the library controls.
- They appreciate the opportunities to annotate, review, and “tag” resources, as well as share them with others. Users like results sorted by relevance, a key component of which is how others like themselves have used the resources.
- Instead of having to come to the library, physically or virtually, they would appreciate the above services being delivered to where they are—be it an institutional portal, course-management system, commercial search engine, and other mechanism.
Libraries determined to reform their operations and modernize themselves in response to these demands and expectations have a couple of choices of how to regard metadata. One view sees library metadata as overkill in light of what users of commercial search engines can be satisfied with. Any search functionality or data element not enjoying significant usage is a prime candidate for elimination. The granularity of library metadata probably is responsible for presenting too many searching choices on online catalog screens. With the arrival of e-resources in library workloads, the current methods of resource description appear less and less scalable. The availability of full-text online tempts the conclusion that summary-level metadata is unnecessary. The large expense going into a behind-the-scenes cataloging operation represents an opportunity cost for libraries, in terms of what they could be doing for more directly beneficial user services. If users are abandoning the catalog for Amazon and Google, it is an indictment of the outdated practices and products of cataloging activity.
Another view sees metadata as the library’s unique “value add” in this information age. In Web 2.0 terms, metadata gives those who use it a competitive advantage in the information marketplace. The richness of the metadata creates opportunities for providing a wide variety of library services. Bibliographic records support more than searching catalog inventory.8 In a Web environment, they can provide a rich set of possibilities for browsing relationships that a title has with related material. Reasonably consistent data can support precise report writing against a file of bibliographic records, along parameters not offered in live catalog searching.
A solution to poor-quality experiences using existing online catalogs is separating the front-end interface from the valuable back-end data store. An even greater means of justifying and recovering the cost of creating metadata is capitalizing on the expanding markets for it. In the future, online catalogs will be just one of the places that library metadata goes.
At the June 2006 Technical Services Administrators of Large Research Libraries Discussion Group meeting in New Orleans, an update on the efforts to make library data available to commercial search engines elicited the observation that receipt of the initial metadata increased the appetite for more of it. At an August 2006 Digital Library Federation workshop on implementation of the Open Archives Initiative, presenters confirmed the trend is moving away from seeking lowest common denominator, unqualified Dublin Core toward seeking whatever richer schema a data provider happens to have.9 Another insight from the same workshop is that metadata’s role does not end once a described resource has been discovered and delivered. Metadata also supports the use of the resource, enabling a user to obtain a citation of it, as well as to find resources related to it. For nontextual objects, the metadata explicates what the user has found.
The benefits of library classification systems seem to be steadily growing over time. Classification can provide a unified browse of library resources on similar subjects—those on open shelves, in use, in remote storage, or in digital format. High-level subject browse categories can be created for subsets of the collection such as e-resources, by mapping from existing classification data. When a set of subject browse categories needs to change, particularly being made more or less granular, this can be achieved by remapping from the classification, versus starting from scratch. Powerful analysis of collections and assessments for accreditation are supported by classification available in cataloging data.
While this future for library metadata looks very bright, there are some changes we will need to make. If we wish to sustain the level of metadata we are accustomed to now, we need to be willing to change a long-standing cataloging model. Libraries have historically approached new-title cataloging as starting with the existing bibliographic utility record and custom editing a copy of it for the local file. Revisions to existing cataloging records often are made directly in the local file. Some of the original cataloging energy involved was channeled back to the bibliographic utility’s record, through the Program for Cooperative Cataloging, the OCLC Enhance program, and others.
At a regional or national level, we need to be willing to start working in a communal file, where the work on bibliographic records need only be done once. Imagine the savings and efficiency that could be gained if serial title changes, authority work to maintain headings, and other upgrades to data could be done once and automatically shared by others. Think about the prospects of having additional resources for placing more material under better bibliographic control, gearing more of our cataloging toward more original and unique work, and being able to serve a wider variety of research needs. We have to be willing to cooperate in using each other’s records (largely) as-is, to agree on standards, to trust each other.
Libraries should not always feel that they have to be the ones to provide the metadata. If there is similarity in the uses made of bibliographic descriptions by publishers, vendors, and libraries for their inventories, and if there is significant overlap among needed data elements, then it makes sense to pursue a single metadata creation effort whose results we all can use. A good example of this is the recent arrangement between North American libraries and Casalini Libri to share a national-level bibliographic record.
The model of relying on national library provision of high-quality cataloging records may be shifting to a wiki. We will need to rely on ourselves, together with strategic partnerships with others, for sustaining a rich underlying metadata that can guide users to many bibliographic services. Users––and fulfillment of their wide range of needs––depend on good metadata for finding, identifying, selecting, and obtaining resources.
Call this the vision of the One Good Record. The least scalable aspects of metadata are the local variation and the duplication of effort.
The more a library goes it alone in metadata creation policies and efforts, the sooner it will run up against the limits of what it can sustain and the sooner it will feel the need to retrench the content of the record.
References and Notes
1. | Lorcan Dempsey, “Thinking about the Catalog,” weblog entry (Jan. 12, 2006). http://orweblog.oclc.org/archives/000919.html (accessed Aug. 30, 2006); University of California Libraries Bibliographic Services Task Force, Rethinking How We Provide Bibliographic Services for the University of California: Final Report (2005). http://libraries.universityofcalifornia.edu/sopag/BSTF/Final.pdf (accessed Aug. 27, 2006); North Carolina State University, “NCSU Libraries Unveils Revolutionary, Endeca-Powered Online Catalog” (Jan. 12, 2006). www.ncsu.edu/news/press_releases/06_01/007.htm (accessed Aug. 27, 2006). The online catalog is located at www.lib.ncsu.edu/catalog/ (accessed Aug. 27, 2006) |
2. | Karen Calhoun, The Changing Nature of the Catalog and Its Integration with Other Discovery Tools: Final Report, prepared for the Library of Congress, Mar. 17, 2006. www.loc.gov/catdir/calhoun-report-final.pdf (accessed Aug. 27, 2006) |
3. | Deanna Marcum, “The International Dimensions of Digital Science and Scholarship: Address to the American Association of Research Libraries and the Canadian Association of Research Libraries, 18 May 2006, Ottawa, Canada.” www.arl.org/arl/proceedings/148/marcum.ppt and www.arl.org/arl/proceedings/148/marcumkeynote.pdf (accessed Aug. 30, 2006) |
4. | Library of Congress, Director for Acquisitions and Bibliographic Access, “The Director for Acquisitions and Bibliographic Access Announces the Library of Congress’ Decision to Cease Creating Series Authority Records as Part of Library of Congress Cataloging” (Apr. 20, 2006). www.loc.gov/catdir/series.html (accessed Aug. 30, 2006) |
5. | Ex Libris, Primo: Find It, Get It. www.exlibrisgroup.com/resources/various/primo_A4.pdf (accessed Aug. 30, 2006); “What Is Primo?” The Ex Librian: Ex Libris Newsletter 3, no. 1(Apr. 2006). www.exlibrisgroup.com/newsletter/?cat=71&parent=19 (accessed Aug. 27, 2006) |
6. | Linda Barnhart, personal communication to the author, Aug. 24, 2005 |
7. | Amy Kautzman and Terry Ryan, “Pathway to the Future: Library Bibliographic Services for the 21st Century,” presentation made at the 6th Living the Future Conference, Apr. 5–8, 2006, Univ. of Arizona. www.library.arizona.edu/conferences/ltf/2006/documents/LTF6April2006final.ppt (accessed Aug. 27, 2006) |
8. | Tim O’Reilly, “What is Web 2.0: Design Patterns and Business Models for the Next Generation of Software” (Sept. 30, 2005). www.oreillynet.com/lpt/a/6228 (accessed Aug. 30, 2006). Web 2.0 is a “set of principles and practices” governing Web services. The article includes the advice that “Data is the Next Intel Inside. Applications are increasingly data-driven. Therefore: For competitive advantage, seek to own a unique, hard-to-re-create source of data.” This quickly brings to mind the nature of the metadata that libraries have long valued and in which they have invested |
9. | Digital Library Federation, “DLF OAI Implementers Workshop, Aug. 25, 2006, Stanford Univ.” www.diglib.org/architectures/oai/imls2004/training2/OAI-workshopOverview.pdf (accessed Aug. 30, 2006) |
Figures
|
Article Categories:
|
Refbacks
- There are currently no refbacks.
© 2024 Core