MALIBU Project

Title: Resource sharing and hybrid libraries: the MALIBU Project

Derek Law, Director of Information Services & Systems, King’s College London

British university librarians have been brought up on those twin pillars of Universal Bibliographic Control, the Library of Congress Catalogue and the British Library Catalogue. Although they filled great banks of shelves in their dress of green and blue, no research library was (or is) complete without them. But they followed somewhat idiosyncratic rules, reflecting their separate and venerable origins. The international nature of scholarly literature has been reflected in the growth of international (or at least interoperable) standards which at least complement if not replace these great works of scholarship. The Anglo-American Cataloguing Rules, the MARC standard, ISBNs and ISSNs and the growth of national bibliographies all represent a huge and successful professional effort over the last four decades to globalise and standardise bibliographic control. The dominance and constant revision of both Dewey and LC Classification schemes have also forced order and structure on an essentially chaotic publishing output. Research libraries have introduced coherent collection development policies in an effort to focus their collections and as a response to financial pressures. As library housekeeping systems became universal in the 1970’s and 1980’s co-operative catalogues grew and retrospective conversion on a quite massive scale took place. Although not yet complete it has hugely enhanced access to collections throughout the world from throughout the world. The abstracting and indexing of journals has a longer and largely commercial history. But it too has provided a near comprehensive system of access which has kept pace comfortably with the great post-war expansion of scientific literature. Political and economic instability in many parts of the globe may have prevented the total establishment of UBC, but as a profession we have developed and actively maintained a vigorous and robust bibliographic infrastructure. All of this has required a constant but largely unremarked stream of diplomacy and activity of the very highest order.

Universal Availability of Publications has been a particularly British led activity – indeed IFLA’s UAP office is based at the British Library in Boston Spa. If bibliographic control is a necessary pre-requisite of access to collections, inter-library lending is its acme. Of course libraries have lent books for centuries, but the development of standardised national and international systems has created the hard-won rather than inevitable system we see today. It is not self-evident that subject-based document delivery systems such as that of the National Library of Medicine or co-operative systems such as that run by OCLC should be interoperable with systems run by a national library. The development of a common currency; the very act of trusting fellow professionals – and even more implausibly their readers - in small libraries in countries at the other end of the globe to act in a uniformly responsible way; the meshing of different copyright and fair use traditions and legislation all represent quite remarkable acts of international co-operation. Not all is perfect, of course, since not everything is deliverable and sometimes the scholar must still move to the book. Nevertheless it is broadly true that scholars anywhere can identify and acquire or gain access to the most obscure works of scholarship, wherever held

In sum, the combined efforts of librarians throughout the world have created global systems of bibliographic description and access over recent decades. It might be argued that these are a natural consequence of our professional skills and interests, but even that view would have to acknowledge the sheer scale and complexity of what has been achieved. Yet having climbed the twin peaks of UBC and UAP, like Sisyphus we are now faced with the same task in a yet more complicated digital environment. And the task is much harder this time. In the world of print the field was largely ours to make of it what we would. In the new environment we are jostled and compete with computer scientists, publishers, authors, learned societies and agents, all of whom feel they have a role to play and skills to offer. This is perhaps most simply illustrated by the creation of the ubiquitous URL as a standard. It requires a very particular skill to create a descriptor of an ugliness and lack of structure which makes the Library of Congress Classification seem elegant and classically simple.

A brief review of the problems associated with scholarship on the Internet demonstrate that it is as yet an inappropriate medium to which to entrust scholarship. Some are a function of the medium while others are variations of old issues. A thread which runs through all of the problems is the failure of the academy to recognise that the problems exist. The Internet is seen as a great and liberating development, but it is not a neutral development and requires very substantial international effort if it is to be made usable for sustained scholarly communication rather than short-term gratification. The problems begin at the most basic levels.

The very act of naming and identifying electronic objects consistently is fraught with difficulty. A book is a static object which does not change over time. In an electronic environment there is a need to reference objects as they move and change over time and place. The ephemeral nature of URLs is notorious. Even where the URL remains constant, issues of version control and quality assurance remain unresolved. The seriousness of this problem cannot be overemphasised for the continuity of citation is central to scholarship.

The issue of naming objects is also difficult and as yet unresolved. At present anyone can name an object with no obligation to maintain names over time. This is compounded by the fact that many of the reference points we take for granted in the print world disappear. A book published by Oxford University Press implies a set of values, standards and scholarly rigour that is understood. But an address incorporating the phrase “ox.ac.uk” could be anything from a university press to a student p.c. in a rented room. The persistence of object names is a long way from having a settled structure – and there is little evidence that the official bodies in scholarship understand the threat this poses.

Metadata and the description of objects is in rather better case. The Dublin Core standard has very rapidly developed international acceptance with participation in standards work from Europe, USA and the Pacific Rim. But even here much work remains to be done. Cataloguing has historically described static and largely immutable objects. The Internet offers new genres of multimedia and even services which will require appropriate description. This work remains to be developed.

Unlike the book, terms and conditions of use must also be described for electronic materials. Many will have multiple copyright permissions, many will be licensed rather than purchased, many will have restrictions on categories of users – and these will vary according to the terms of sale rather than be inherent in the product. Although the initial success of the Dublin Core gives confidence that these problems can be resolved, a great deal of international effort will be required to create a usable system.

Searching and indexing have proved much more difficult technically than the designers of web robots would have us believe. Web indexing systems are breaking down as their architecture collapses under the weight of data. It is increasingly common to undertake a search on the major systems and to recover hundreds of thousands of hits in apparently random order. Much work is going on here but designers despair at the inability or unwillingness of the public to master Boolean searching and most systems still have a long way to go to beat a half way competent reference librarian. Web searching has undoubtedly transformed the ability of to acquire a whole range of current reference information, but is dramatically poor at discovering scholarship and research.

Unlike the print world, the electronic one will require validation of the rights of the user. User authentication is regarded as an essential element of electronic commerce, but it too lacks basic elements for the furtherance of scholarly activity. At present there are no good ways of proving membership of the “data club” when away from the parent institution. Scholars visiting another institution, students on vacation or researchers on field trips are difficult to validate. There is then a very knotty problem surrounding usage data. On the one hand commercial publishers wish to collect usage information as a marketing tool. They are, however, unwilling to release this information to libraries so that they can judge whether usage justifies subscription. Conversely many users do not wish anyone to know what they are reading or researching. Traditionally, libraries have preserved the anonymity of user data except where criminal acts are suspected. Is this a right or simply a custom?

Then there are a series of issues and old battlegrounds to revisit. Rights Management Systems are growing quickly and are promoted largely by commercial concerns. They provide many areas of philosophic contention. As mentioned above, the question of whether the user can remain anonymous conflicts with commercial need. Secondly, the issue of preservation remains technically, legally and operationally unresolved. Historically this has been the domain of the national libraries, but it is not clear that they will or can perform the same role in an electronic environment. We cannot reasonably expect preservation to be undertaken by publishers. And thirdly the whole issue of fair use is being revisited by publishers, some of whom declare that it does not or cannot exist electronically. Major battles need to be undertaken on these issues, again with little evidence that the academy understands or cares about the issues.

The preservation and archiving of electronic information has barely surfaced as a very complex issue. The Data Archive at the University of Essex has existed for some twenty-five years and has perhaps as a clear a picture as anywhere of the so far intractable problems of storing, refreshing and kite-marking information. The problems are staggeringly complex technically and staggeringly expensive to resolve. Although some progress is being made on the legal deposit of commercial material, little appears to be done on the non-commercial and primary materials of scholarship. There are no standards or control or approval mechanisms for institutions or data repositories. This position may be compared with that in the United Kingdom where archives are expected to meet the BS5454 standard and the Historical Manuscripts Commission takes an active interest in the state of repositories and where archivists have specialist professional training. A new class of electronic material, what Clifford Lynch of CNI has called “endangered content” is emerging, where the formal and informal records of disciplines are effectively at risk through neglect. University Archives collect papers, but institutions do not sample or preserve the electronic mail or word-processed files of their scholars. Lab books are routinely preserved by scientists but it is doubtful if any institution has a policy for the preservation of digitally captured images or data from research equipment.

Network topology is barely discussed as an issue due to a naïve assumption that there will be an infinitely expanding amount of bandwidth which will somehow be made available to scholarship. And yet there is no evidence to support this view. American universities have abandoned the failing Internet provided by telecommunications companies to create Internet II as a private network attuned to their needs. In Europe the relatively modest ambition of the European Union to link existing research networks through the TEN-34 Project has been “shaped by a series of non-technical influences such as non-availability of required public services”, while “standard PNO (public network operator) services in Europe could not fulfil the requirements of the R&D community in Europe”. Equally the assumption that we accept a simple commercial approach to network planning is questionable. At present in the UK, bandwidth is acquired in the light of use rather than as a result of scholarly or educational policy decisions. Thus bandwidth expands at a great rate to the East Coast of North America to meet traffic growth. There is almost no debate on whether policy should drive such acquisition and route bandwidth say to Southern Africa then India, Singapore, Australia and then the West Coast of the United States opening up markets and scholarship to what is sometimes called UK Higher Education Limited. There is a creeping form of cybercolonialiasm in the assumption that only the largest English speaking countries have digital material of value to the world. It is interesting to note the recent decision of the Australian Vice-Chancellors to use network charges to discriminate against overseas websites and in favour of Australian ones. No discussion appears to take place of how the products and output of small learned societies are to be mirrored around the world and what standards and quality controls will apply to mirror sites. Again the scholarly community is silent while the commercial giants of the STM world dictate the shape of electronic scholarly communication – despite the fact that large scientific publishers are aberrant rather than the norm.

Nor is the network yet totally robust. A recent Dilbert cartoon pointedly and uncomfortably accurately suggested that all of the time saved through automation in the information age had been lost by people sitting at computers waiting for web images to load. Networks do not yet for example give the reliable quality of service required for multicasting, while video clips have all the power, quality and assurance of early silent films. It should be self-evident that for research institutions working at the leading edge of scholarship and indeed telecommunications, the standard services provided by Internet Service Providers will always be inadequate.

A more positive element which is emerging in the electronic era is the broadening of what constitutes content. Services such as the Arts and Humanities Data Service based at King’s College London or the excellent SCRAN project funded by the museums of Scotland are much involved in the digitisation of museum and archive collections. This is happening fast and brings relevant experience in activities such as new licensing models and standards. It also highlights the role of curators in the digital environment as relating to presentation as well as preservation. But again there appears to be little concerted effort by the official organs of scholarship to build formal cross-domain linkages.

Thus we can see that the Internet lacks many of the features associated with long-term scholarly communication. It is clear that scholarly research libraries can expect to operate in an environment which mixes existing with present and future resources – paper, microform, digital sound and image and must develop models for offering robust services. The British electronic libraries programme has funded four projects to develop thinking on these so-called hybrid libraries and one of these is MALIBU (Modernising Academic Libraries in British Universities), based at King’s College London. Working with the libraries of Oxford and Southampton Universities it will take a very rich set of humanities resources from archives and incunabula to digital products and networked resources and create a seamless single access point to all the available resources. It will also explore how strategies can be developed to make resources available locally as operational services rather than unreliably over the Internet. It does not set out to create its own new tools but rather to find ways of integrating tools already in existence or being developed. The traditional library forms a perfect paradigm for the way in which organisations can manage – and more importantly support – electronic resources. For centuries organisations have identified the information resources they require, collected as much of it a possible in one place, employed staff to manage, organise and preserve it and set up controlled document delivery arrangements for those who require access to material not held locally. What they have not done is to hand out tiny sums of money to every member of the organisation and invited them to make their own arrangements for information access. As the cost of filestore and the cost of bandwidth begin to come together the library model seems increasingly attractive as a way of managing electronic resources.

TRADITIONAL LIBRARY ELECTRONIC LIBRARY

–acquisition > resource discovery

–collection building > local fileservers

–classification > knowledge management

–preservation > long-term data sinks

–user instruction > user instruction

The local/ consortially based model for delivery of electronic resources may then be seen to fit more appropriately with the historic needs and attitudes of scholarship and at least for the moment offers the most appropriate vehicle for supporting research and scholarship.