Evaluation of Digital Libraries Criteria and Problems From Users Perspectives 2006 Library Information Science Research
Evaluation of Digital Libraries Criteria and Problems From Users Perspectives 2006 Library Information Science Research
Evaluation of Digital Libraries Criteria and Problems From Users Perspectives 2006 Library Information Science Research
Abstract
Digital library research has developed over the past decade, but little has been done on the
identification of evaluation criteria, especially from users' perspectives. This article identifies users'
criteria and applies them to the evaluation of existing digital libraries. Forty-eight subjects were
instructed to develop and justify a set of essential criteria for the evaluation of digital libraries. At the
same time, they were requested to evaluate existing digital libraries by applying the criteria that they
were developing. A compilation of criteria developed by participants show that usability and collection
quality were the most important criteria for evaluating digital libraries. Service quality, system
performance efficiency, and user opinion solicitation were also deemed essential criteria. The author
further compares digital library evaluation criteria identified by users and researchers and applied in
previous research. The article reveals problems in current digital library design and development,
discusses suggestions for improving digital library design, and presents areas for further research.
© 2006 Elsevier Inc. All rights reserved.
1. Introduction
The emergence of digital libraries provides more opportunities for users to access a variety
of information resources. There are different definitions in the literature as to what constitutes a
digital library. Chowdhury and Chowdhury (2003) place digital libraries into two major
categories based on Borgman's (1999) discussion of competing visions. One approach focuses
0740-8188/$ - see front matter © 2006 Elsevier Inc. All rights reserved.
doi:10.1016/j.lisr.2006.06.002
434 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
on access and retrieval of digital content and the other focuses on the collection, organization,
and service aspects of digital resources. The Association of Research Libraries (1995)
identifies the common elements of digital library definitions: the digital library is not a single
entity; the digital library requires technology to link the resources of many; the linkages
between many digital libraries and information services are transparent to the end users;
universal access to digital libraries is a goal; and digital library collections are not limited to
document surrogates, they extend to digital artifacts that cannot be represented or distributed in
printed formats. Digital library research has developed rapidly over the past decade, and
millions of dollars have been spent on building digital libraries. However, previous research
indicates that many potential users may still not use them (Thong, Hong, & Tam, 2002).
Research on the evaluation of digital libraries is still in its infancy. Researchers are still
investigating who should evaluate, when to evaluate, what to evaluate, how to evaluate, and
why to evaluate. As Saracevic and Covi (2000) point out, the evaluation of digital libraries is a
complex undertaking that is conceptually and pragmatically challenging. Borgman, Leazer,
Gilliland-Swetland, and Gazan (2001) further suggest that technical complexity, variety of
content, uses and users, and the lack of evaluation methods contribute to the problem. Any
evaluation is based on the conceptual model of the evaluators: their understanding of the goals
of the system and of users' needs and behaviors. Evaluation itself is a form of sense making,
and it is also situated (Van House, Butler, Ogle, & Schiff, 1996).
As to who should evaluate digital libraries, users of digital libraries should have their
voices heard. After all, the ultimate goal of the development of digital libraries is to serve
users and to facilitate their effective use of information and services. Marchionini, Plaisant,
and Komlodi (1998) emphasize that all efforts to design, implement, and evaluate digital
libraries must be rooted in the information needs, characteristics, and contexts of the people
who may use those libraries. Research on digital libraries has moved from the technical
aspects of building digital libraries to how to design digital libraries to satisfy user needs.
One way to assess user needs is to investigate digital library evaluation criteria from the user
point of view.
2. Research problem
• What criteria do users identify as important for the evaluation of digital libraries?
• What are the problems with the existing digital libraries?
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 435
3. Literature review
studies and collected feedback about the users, the interface, system functionality, and content
of the digital library. Cherry and Duff (2002) conducted a longitudinal study of a digital library
collection of Early Canadiana Materials, focusing on how the digital library was used and the
level of user satisfaction with different features of the digital library, such as response time,
browsing capabilities, comprehensiveness of the collection, print function, search capabilities,
and display of document pages. These studies provide a basis for improvements to the digital
library so that it could meet, and possibly exceed, end user needs and expectations.
Digital library service evaluation received relatively little attention in the early phases of the
Internet revolution when developments tended to be technology driven (Heath et al., 2003).
Recently, digital library evaluation has begun to include to service evaluation. Heath et al.
(2003) discussed the benefits of a combined methodology that would provide greater potential
for evaluation of digital library services but failed to conduct an actual evaluation study.
Monopoli, Nicholas, Georgiou, and Korfiati (2002) evaluated the electronic journals service of
a digital library and emphasized understanding the users and their use patterns. The majority of
digital library service evaluation has focused on digital reference services. For example, Carter
and Janes (2000) analyzed logs of over 3000 questions asked of the Internet Public Library on
the basis of questions asked, and how those questions were handled, answered, or rejected.
Another type of digital library evaluation is to examine the impact of digital libraries on
users and their communities, typified by the longitudinal study of the Perseus Digital Library,
evaluating its impact on users and the humanities community (Marchionini, 2000). Bollen,
Luce, Vemulapalli, and Xu (2003) analyzed the usage patterns derived from log analysis and
further identified research trends that evolves in an institution's user communities over time.
This research has implications for the acquisition and retention of digital content. In addition,
research on digital library evaluation explores what determines digital library acceptance. The
results showed both perceived usefulness and perceived ease of use were determinants of user
acceptance of digital libraries (Thong et al., 2002). Digital library evaluation also involves
comparing use of digital libraries and other types of systems. MacCall, Cleveland and Gibson
(1999) reported a preliminary evaluation of a classical digital library model by analyzing usage
logs and comparing users' experience in using a digital library with their experience in general
Internet use.
Although the volume of published research on the evaluation of digital libraries has
increased, little of it discusses criteria for evaluation. Currently most of the research on digital
libraries evaluation is based on existing evaluation criteria for traditional libraries,
performance of information retrieval systems, human–computer interaction, digital technol-
ogies, and so on. Marchionini (2000) suggests digital libraries are extensions and
augmentations of physical libraries. Researchers might use existing techniques and metrics
to evaluate digital libraries, for example, circulation, collection size and growth rate, patron
visits, reference questions answered, patron satisfaction, and financial stability. Evaluation
criteria for digital technologies can also be useful, such as response time, storage capacity,
transfer rate, user satisfaction, and cost per operation. However, digital libraries provide new
services, products, and capabilities, making it more difficult to compare them with physical
libraries. After reviewing evaluation criteria for libraries by Lancaster (1993), for library and
information services by Saracevic and Kantor (1997), information retrieval systems and
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 437
4. Methodology
4.1. Participants
A total of 48 subjects participated in the study. To enhance the representation of the data, the
author collected data twice, in 2003 and 2004. Subjects were recruited from the School of
Information Studies, University of Wisconsin-Milwaukee. The recruitment message stated that
a study to investigate digital library evaluation would be conducted in order to create digital
libraries that truly support effective use. Users' inputs were needed regarding their perceptions
of digital library evaluation criteria and their evaluation of the strength and limitation of
existing digital libraries. Twenty-three subjects were recruited and their responses were
collected in fall 2003 and another twenty-five were recruited and their responses were
collected in fall 2004. These subjects were graduate and undergraduate students who were
interested in the use of digital libraries. Female subjects account for 69% of the participants.
While the majority of the subjects (87.5%) were graduate students, 2.5% were undergraduate
students. All of them (100%) had basic knowledge of digital libraries and had used and
searched digital libraries.
The objective of this study is to develop criteria for the evaluation of digital libraries. Before
data collection, the subjects were informed of the common elements of digital library
definitions identified by the Association of Research Libraries (1995). Data were collected
mainly from an open-ended self-administrated survey. The same instruction was given to all 48
subjects. All the subjects had approximately one week to respond to the survey. The survey
438 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
included two parts. In the first part, the subjects were instructed to develop a set of evaluation
criteria for the evaluation of digital libraries. To be more specific, they were asked to identify
the essential criteria that were important for the development and use of digital libraries. They
were also instructed to justify their choices. For each of the criteria that a subject identified, he
or she needed to discuss why that criterion was important.
In the second part of the survey, subjects were asked to report on their evaluation of existing
digital libraries. They were instructed to evaluate an existing digital library by applying the set
of evaluation criteria developed by themselves. In order to conduct the evaluation, the
participants were provided with a list of existing digital library projects that represented a
variety of digital libraries developed by different agencies and organizations, including
governments, associations/organizations, universities, museums, and companies. These
included such collections as the Library of Congress American Memory Project (http://
memory.loc.gov/), the ACM Digital Library (http://www.acm.org/dl/), and theElectronic
Poetry Center at SUNY-Buffalo (http://wings.buffalo.edu/epc/). Each subject was instructed to
select one digital library with which he/she was familiar and evaluate it. A subject could also
select a digital library that was not on the list as long as it satisfied the requirement of a digital
library. Table 1 presents detailed information on the types of digital libraries selected by the
subjects. Digital libraries developed by associations/organizations, governments, companies,
and universities were the major ones evaluated.
Before providing their evaluation, participants were instructed first to explore the digital
library thoroughly. To be more specific, they were told to browse through (e.g., whether they
can find one item from two paths) and search their own questions in the selected digital library.
After that they were asked to discuss problems with the digital library by going through every
criterion they identified in the first part. They were also asked to provide examples to illustrate
the problems they encountered in the digital libraries they selected to evaluate.
Both quantitative and qualitative methods were used to analyze the data. Quantitative
methods were employed to conduct descriptive data analysis, such as frequency and mean.
Content analysis (Krippendorff, 2004) was applied to develop categories and subcategories of
evaluation criteria and problems. According to Krippendorff (2004), categorical distinctions
define units by their membership in a category by having something in common. In this study,
Table 1
Types of digital libraries (DL) selected for evaluation
Types of DLs No. of DLs selected No. of subjects selected DLs
Government 3 7
Association/organization 7 27
University 3 3
Museum 1 1
Company 3 9
International 1 1
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 439
each category and subcategory of evaluation criteria and problems with digital libraries is
defined and discussed by citing responses directly from the participants. Instead of creating an
existing structure, types of criteria and subcriteria and types of problems were derived directly
from users.
The data analysis proceeded as follows: (1) Each subject's responses about evaluation
criteria were extracted from the survey. (2) All the responses that shared the same properties
were classified into one category. For example, all the responses about interface use were
classified into one category, and all the responses about content were classified into another.
(3) A name for each type of criterion was assigned based on the content of the responses. The
types of criteria identified generally used terminology common in the library and information
science field (e.g., usability, collection quality, service quality, etc.). (4) After all the
responses were classified into general types of evaluation criteria, all the responses for one
type of criterion were further analyzed to differentiate subgroups based on their properties.
For example, all the responses about different ways of finding information were classified
into one subcategory while the responses about the importance of help functions were
categorized into another subcategory. (5) A subcategory name was assigned to the subgroup
based on the content of the responses and its general type name. For example, in the usability
category, search and browse, navigation, help features, view and output, and accessibility
emerged as subcategories. Some of the responses discussed the types of evaluation criteria in
general; therefore, usability in general, content in general, and IR in general were also
identified as separate subcategories. The same procedures were followed for the analysis of
problems. Examples of categories and subcategories of evaluation criteria and problems with
digital libraries are further discussed in the Results section. In order to provide a better
understanding of participants' views of different types of digital library evaluation criteria,
the author also calculated the number of subjects and the percentage of the total subjects who
identified each criterion. The results of the data analysis of types of evaluation criteria are
presented in Table 2.
5. Results
Table 2 presents frequency and percentage of types of criteria that participants developed.
Both content and format are important to users. Interface usability and collection quality were
considered as important criteria by a majority of the participants. In addition, service quality,
system performance efficiency, and user opinion solicitation were also deemed as essential
criteria by many of the participants.
5.1.1. Usability
Almost all the participants considered usability as the most important criterion for a useful
digital library. One subject stated it, “I think the most important criterion that I listed above is
usability. Users need to be able to navigate through a digital library with a certain sense of ease,
440 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
Table 2
Types of evaluation criteria (N = 48)
Types of criteria Types of subcriteria Number of responses Percentage
Usability Usability in general 25 52
Interface usability 26 54
Search and browse 34 71
Navigation 24 50
Help features 20 42
View and output 6 13
Accessibility 4 8
Other 10 21
Collection quality Content in general 37 77
Quality 13 27
Scope 19 40
Authority 18 38
Accuracy 17 35
Completeness 15 31
Currency 12 25
Other 7 15
Service quality Mission 15 31
Intended user community 12 25
Traditional library service 5 10
Unique services 5 10
Other 6 13
System performance efficiency IR in general 9 19
Efficient and effective 4 8
Relevance 3 6
Precision and recall 2 4
Other 6 13
User opinion solicitation User satisfaction 16 33
User feedback 7 15
Contact information 5 10
Other 2 4
or else they may become frustrated and decide to go elsewhere” (S32). More than half of the
participants (52%) discussed usability in general as the key evaluation criterion. Participants
cared the most about interface usability. Fifty-four percent of the participants emphasized the
importance of interface design. As one subject pointed out, “This is the connection between the
user and the system. It is the first thing they see. It is their tool to receive output from the digital
library” (S36). A couple of participants emphasized the importance of the interface as a
gateway between a user and a digital library. One subject concluded that interface evaluation is
important because “if the users can't access the DL effectively or finds it cumbersome, the
information is not really accessible to them” (S24).
The majority of participants (71%) identified search and browse functions as essential
criteria for the evaluation of usability of digital libraries. As one subject stated (S39), “Search
features are important in any digital library, to enable the quick retrieval of information, when
certain details of that information are already known. The browse feature is important for
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 441
retrieval of information that is not already known in part. Browsing enables people to look
through a digital library and discover things they had no previous knowledge of. ” Participants
indicated that the search and browse functions enhanced a user's ability to find relevant
information effectively.
For half of the participants (50%), navigation efficiency is another key criterion for interface
usability. Being intuitive is the basic requirement for navigation. One subject stated that “a user
must be able to navigate and find his/her ways to information quickly and easily” (S47). To
ensure effective navigation, “multiple access points” and “site map” were identified as
essential from the responses, especially “easily understood access points and good
organization structure which must abide by good traditional cataloging and classification
methods as well as principles of HCI” (S31).
Another important finding is the identification of the availability of help features as one
of the significant criteria. Forty-two percent of the participants chose help feature
availability as an evaluation criterion, mainly because they needed support in using a digital
library. This relates to their general feelings regarding digital libraries. According to one
subject (S1), “Help features allow a user to feel comfortable within a digital library. If a
user is experiencing a problem, a help page or a FAQ page will normally provide the
answer”. Limited by the existing help features, many of the participants only considered the
use of help features for technical problems. There is no comment about how help systems
can help clarify user need. Here is a typical comment from a subject, “Help system is
important to help with technical difficulties. It should be easy to find and easy to
comprehend” (39).
Other criteria related to usability were also identified by the participants. For example, 13%
of the participants identified view and output issues. A subject explained the reason for the
view criterion, “The user needs to have the appropriate player or viewer to access the items in
the collection. The user needs to know which player or viewer is needed” (S37). Another one
discussed the output criterion, “For instance, can the user print, save, and email the desired
information? If the user can't ‘take’ the information in the form that they want, it has a negative
impact on them” (S36). Four participants mentioned accessibility as an issue, indicating that
people would not use resources if there were too much required in the way of logging on. Other
criteria mentioned by participants were HCI principles, reliability of the digital library site,
uniformity, learnability, etc.
When participants discussed collection quality, they immediately related it to the traditional
library evaluation criteria since collection is a major component in both digital and physical
libraries. One subject stated, “A further look at the criteria for the evaluation of the content of
this DL will be aspects that would also apply to a more traditional library. These would include
subject, scope, coverage, and authority” (S9). In addition, the characteristics of electronic
materials also lead people to worry about their quality. For example, one subject emphasized
that “the quality of the information should be able to be assessed because anyone can put
anything on the Web” (S26).
Based on participants' responses, the criteria of quality of a collection can be divided into
scope (40%), authority (38%), accuracy (35%), completeness (31%), and currency (25%). It is
important for a digital library to clearly define its scope so users can immediately judge
whether they have accessed the right digital library. One subject argued that the reason for
choosing authority as one of the evaluation criteria was that “Authoritative sources comprise
an important aspect of collection development in a traditional library” (S26). In terms of
accuracy, it is obvious that “if the information is inaccurate, there is no reason for people to use
it” (S17). “How comprehensive is the content of the library, does it cover all subject fields and
to what degree?” (S7) One subject further defined “completeness”, “A good library will cover
its subjects thoroughly and be able to provide information that meets the demands of users with
varying levels of information need” (S38). Currency is another key criterion for evaluating the
quality of a collection. It relates to the usefulness of the material. One subject stressed its
importance, “Collections need to be updated daily, monthly, and yearly in order for them to be
any value to people” (S32).
A subject asked, “Does the digital library provide unique services that are not available
elsewhere” (S17)?
5.2.1. Usability
Based on the criteria they proposed, participants reviewed and identified problems in the
existing digital libraries. Usability is still a major problem, especially with respect to interface
design. Navigation efficiency, availability of search functions, and help features emerged as
main areas that need to be improved. In regard to navigation, the lack of intuitive interfaces
was identified as a problem for some of the digital libraries. Participants found them difficult to
navigate. One subject pointed out, “Project Gutenberg's interface, the intermediary between
the user and the system, is not as intuitive to navigate and use as it should be” (S42). Another
one mentioned a specific problem in one of the digital libraries, “Another omission is that there
was no easily discernable homepage” (S9).
There are several problems related to search features. First, some of the search functions
lack user control and are not transparent to users, especially for experienced users. “The
Project Gutenberg is somewhat unsophisticated and, because it is unable to be manipulated, is
more restrictive than it seems at first glance. It also does not tell the user how it decided that
there was a match between the query and the item”, one subject (S42) explained in detail.
Second, some of the digital libraries do not support the Boolean operators with which some
users are familiar. For example, one subject complained, “When it came to running searches, I
found that I was able to use natural language but the digital library does not support Boolean
searches” (S6).
444 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
In terms of help features, the first problem is that not all the digital libraries provide help.
One subject said, “There is no defined ‘Help’ portion to this digital library. Nothing is offered
about any of the technical needs to access portions of the site” (S39). Another subject backed
up this statement, “Access to information is impeded because there is no site map or help page”
(S26). Some existing digital libraries do not provide context-sensitive help; users have to
figure our how to find the help and how to ask the question. For example, “The Internet Public
Library offers users a link to ‘Ask a Question’, and this link takes users to a form where they
have to fill out their name, email address and a host of other questions including the question
that they want to ask. This can be a hassle to many users who just want to do a quick click to
find an answer” (S43).
In addition, not all the digital libraries provide output options. One of the subjects
complained, “The library does not provide any output features such as print, email or save,
since the site contains many large images, this can be a problem if trying to save to a disk due
to the large file sizes. The site contains some great images and information but the lack of these
features deters the overall user satisfaction” (S10).
one of the characteristics of digital libraries, many of the digital collections “lack multimedia
formats” (S34).
a bit since basketball is a very popular sport in the United States. I also know of quite a few
popular basketball magazines which means either a lot of them are not provided online or the
IPL just chose not to put them in their collection” (S32).
In short, a majority of the subjects of this study considered usability and collection quality
as essential criteria for digital library evaluation. In addition, service quality, system
performance efficiency, and user opinion solicitation were also identified by some of them as
important evaluation criteria. After evaluating existing digital libraries, subjects identified
some problems with the existing digital libraries. They posed the following challenges for the
development of digital libraries: how to design an interface to be easy to use, how to make it
easy for people to assess a collection, how to provide both traditional as well as new services,
how to increase precision of retrieval results, and how to involve users in digital library
development.
6. Discussion
6.1. Similarities and differences of evaluation criteria between users’ and researchers’
perspectives
The evaluation criteria developed by participants of this study echo the criteria that
researchers have proposed (Chowdhury & Chowdhury, 2003; Marchionini, 2000; Saracevic,
2000; Saracevic & Covi, 2000) and the criteria that have been applied in previous studies to
evaluate digital libraries (Bishop et al., 2000; Bollen et al., 2003; Borgman et al., 2001; Carter
& Janes, 2000; Cherry & Duff, 2002; Heath et al., 2003; Hill et al., 2000; Kassim &
Kochtanek, 2003; MacCall et al., 1999; Marchionini et al., 1998; Monopoli et al., 2002; Thong
et al., 2002; Van House et al., 1996). Table 3 presents types of digital library evaluation criteria
identified by subjects of this study and researchers and applied in previous studies. Here
researchers refer to those researchers who have worked on and identified evaluation criteria for
digital libraries. Previous studies refer to those studies that have evaluated digital libraries or
prototypes of digital libraries.
Table 3 does not cover all the criteria identified by researchers and applied in previous
studies, but it does highlight the focus of digital library evaluation research and studies.
Researchers connect digital library evaluation criteria with previous research in evaluating a
variety of information retrieval systems, including libraries. Digital library evaluation studies
focus on examining the criteria from selected categories, mostly to do with interface usability.
Participants of this study associate their evaluation criteria with their experience in using
physical libraries and other types of information retrieval systems. There is a commonality in
the overall categories of the evaluation criteria. To be more specific, the same major categories
(e.g., usability, collection quality, service quality, system performance, and user opinion
solicitation) have been identified by participants of this study and researchers and applied in
previous evaluation studies.
However, participants emphasize more the usefulness of the digital libraries from the users'
perspectives and less from the perspectives of developers and administrators. For example,
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 447
Table 3
Types of evaluation criteria from this study, researchers, and previous studies
Types of criteria Types of subcriteria This study Researchers Previous studies
Usability Usability in general X X
Interface usability X X
Functionality X X
Search and browse X X
Navigation X
Help features X X
View and output X X
Accessibility X
Failures X
Collection quality Content in general X X X
Quality X X X
Scope X X X
Authority X X
Accuracy X X X
Completeness X X X
Currency X X
Cost X
Format X X
Treatment X
Preservation X
Uniqueness X
Service quality Mission X
Intended user community X X X
Traditional library service X X X
Unique services X
System performance efficiency/Use IR in general X X X
Efficient and effective X X X
Relevance X X X
Precision and recall X X X
Success X
Usage X X
User opinion solicitation User satisfaction X X X
User feedback X
Contact information X
Impact Social impact X
On user community X
Other Standard X
cost, treatment, and preservation are some of the important criteria for digital library
evaluation from researchers' perspectives, but the participants in this study did not identify
these criteria as important ones. Participants placed more emphasis on the ease of use of the
interface and high quality of the collection in evaluating digital libraries. Compared with
previous research and studies, this study offers not only detailed information about evaluation
criteria but also why they are important to users. In participants' discussions of the
justification of usability criteria, ease of use and more options were the main concerns. They
448 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
also mentioned the importance of the availability of help features in digital libraries.
Limited by the existing digital libraries, their criteria for evaluation were still quite basic
and tended to focus on whether a help feature existed. They did not further consider the
effectiveness of these features. For example, participants emphasized the importance of a
digital library providing help for technical problems but did not further discuss the need for
help with non-technical problems, such as a help feature that could assist them in clarifying
their need. As to collection criteria, participants discussed evaluation criteria in multiple
dimensions from coverage, authority, and accuracy to currency. Many participants
considered authority and accuracy of a collection more important than comprehensiveness
and currency of a collection. The understanding that everything on the Web might not be
authoritative and accurate affected their perceptions of the importance of these evaluation
criteria.
The participants of this study did identify some of the criteria that researchers and
previous studies have neglected. For the service quality criteria, they considered not only
traditional services used for evaluating traditional libraries but also unique services for
evaluating digital libraries. Participants suggested that digital libraries needed to take
advantage of the opportunity to provide services that traditional libraries cannot provide.
They also considered user opinion solicitation as a critical element for digital library
evaluation, a criterion that researchers and previous studies have not discussed. Participants
further suggested that user feedback needed to be collected actively by tracking user
satisfaction as well as passively by providing clear contact information. However,
participants of this study failed to discuss the social impact of digital libraries on a society
and communities as researchers and previous studies have emphasized. Their perspectives
again were limited by their emphasis on the use of digital libraries and their approach to
digital libraries from the point of view of an individual user rather than that of a researcher
or an institution.
The problems identified by the participants represent their concerns with existing digital
libraries. Fig. 1 presents the digital library evaluation criteria, problems, and suggestions to
solve the problems. According to participants, ease of use is a main problem for existing
digital libraries. While the majority of the participants were mainly concerned with the ease
of use of the interfaces in digital libraries, others, especially experienced ones, sought more
control in finding the information they wanted. They complained that some of the digital
libraries were designed too basically and were not sophisticated enough, therefore they
were unable to manipulate the systems. To be more specific, they could not conduct more
accurate and specific searches. Digital library design needs to balance ease of use and user
control. The basic requirement for ease of use requires providing search, browse, and help
functions as well as a variety of output options. Some existing digital libraries do not have
these features. At the same time, experienced users also want to have more control, such as
more advanced features for them to customize the way they search, access, view, and
disseminate information.
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 449
Fig. 1. Digital library evaluation: Criteria, problem, and implications for design.
One unique finding of this study is that collection quality is less an issue. The major
problem identified by participants was how digital libraries can assist users in making
judgments about quality of the collection. Specifically, they need information to help them
make judgments about currency, copyright, authority, and accuracy of the collection.
Currency and copyright concerns can be easily solved by consistently providing this
information in digital libraries. Users also need assurance on the authority and accuracy of
the collection. Digital libraries need to implement mechanisms to assure users that their
collections are authoritative and accurate. Transparency of collection building might be one
way to help users understand how the collection has been selected and processed and what
measures have been taken to ensure the accuracy of the information. Furthermore, the
procedures for collection selection should be innovative and convincing and could involve
both human intelligence and technology. For example, technologies used to test the
integrity of Web items can be enhanced to ensure users of the accuracy of a document in a
digital library. Standards can also be developed and used to ensure the authority of a
document.
Many participants value the role that traditional libraries play as community centers, and
they wonder if the unique technologies offered in digital libraries can be used to not only
enable them to play the same role but also to enhance that role. However, existing digital
450 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
libraries fail to provide community services that users prefer, such as making announcements
about community activities, and promoting the exchange of ideas about topics of interest to the
community. Synchronous interactions and services are needed. Some of the Web technologies
in video conferences and online chats can be adapted to digital libraries. In addition, more
publicity is needed to inform the community of the existence of digital libraries. Collaboration
between the community and digital libraries for a variety of activities is also an effective
approach to promote community services and uses of digital libraries.
It is essential to offer relevant and useful information to users efficiently. Users generally
care more for precision than recall. Subjects of this study were not satisfied with the precision
of the search results from some digital libraries. One way to solve this problem is to use some
mechanisms to increase precision, such as PageRanking used by Google. Another way to solve
the problem is to facilitate the evaluation of results (e.g., highlight keywords in the display).
Users like to apply the least effort principle to finding useful information to solve their
problems. They are also more sensitive to response time in digital libraries than other types of
information retrieval systems because the collection contains multimedia materials. It is
important to explore ways of enhancing multimedia storage and retrieval.
Since the ultimate goal of digital libraries is to serve their users, it is essential to have a
mechanism to solicit opinions. Based on the responses from the subjects, there are digital
libraries that have not provided channels for users to send their feedback. More important,
user opinions should not be solicited only for evaluation purposes. It is critical to involve
users in the process of design and development of digital libraries. Multiple channels need
to be opened to facilitate communications between designers/developers and users along the
way. Moreover, these channels should support the active solicitation of information (e.g.,
distribute user surveys to request user feedback) as well as the passive solicitation of
information (e.g., provide clear contact information for users to send their feedback).
The contribution of this study is limited by its sampling. The selection of this group of
subjects has both benefits and problems. On one hand, the subjects of this study are users of
digital libraries, and they have basic knowledge and skills in the use of digital libraries. They
are the ideal subjects for the study. On the other hand, they do not represent a variety of user
groups since the subjects are all students of the School of Information Studies. Their
knowledge and experience have an impact on their perceptions of evaluation criteria for digital
libraries. For example, some of them might have been exposed to evaluation criteria for
traditional libraries or IR systems. Furthermore, it is a convenience sample since the subjects
are volunteers and they were not randomly chosen. Another limitation of this study is that the
evaluation of the digital library is largely based on the subjects' own report on the open-ended
survey. This study asked subjects to use the digital libraries, but did not record their use of
digital libraries. It would be better if in a further study their use were captured and obtained.
Finally, the subjects of this study each selected the digital library he or she would like to
evaluate. While this could provide a more comprehensive evaluation of existing digital
libraries, it does not offer a focused view on one specific digital library or one type of digital
H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452 451
library from a group of users. In short, this can be considered as a preliminary study of digital
library evaluation from users' perspectives. Further research is needed.
7. Conclusion
Although researchers have developed digital library evaluation criteria, and previous studies
have evaluated existing digital libraries or prototypes of digital libraries, researchers are still
investigating what to evaluate. Furthermore, there is not much input from users in terms of
what are the essential evaluation criteria. The major contribution of this study is that it is one of
the few studies that examine digital library evaluation criteria based on users' perceptions. By
applying their own criteria, participants evaluated five types of digital libraries developed by
different organizations. Their evaluation reveals some of the problems in current digital library
design and development. As digital libraries normally target different levels of user groups, it
is essential for digital libraries to support ease of use as well user control. It is also critical for
digital libraries to be transparent in design. Digital libraries not only need to provide high-
quality collections but, more importantly, need to assist users in their efforts to make
judgments about the quality of the collection. It is also important for digital libraries to provide
both traditional and unique services.
Considering the limitations of this study, further research might focus on the evaluation of
one or two types of digital libraries, and the participants should be the targeted users of these
digital libraries. Digital libraries need to be evaluated in their social and organizational context.
Further research also needs not only to identify evaluation criteria but also measure the degree
of importance of each criterion to participants. In addition, multiple methods can be used to
extend the study from the identification of evaluation criteria to the evaluation of actual use of
digital libraries.
Acknowledgments
The author would like to thank Esha Datta for her assistance in data analysis and to the
anonymous reviewers and Candy Schwartz for their constructive comments.
References
Association of Research Libraries. (1995). Definition and purposes of a digital library. Retrieved July 17, 2002,
from http://www.arl.org/sunsite/definition.html
Bertot, J. C., & McClure, C. R. (2003). Outcomes assessment in the networked environment: Research questions,
issues, considerations, and moving forward. Library Trends, 51, 590−613.
Bishop, A. P., Neumann, L. J., Star, S. L., Merkel, C., Ignacio, E., & Sandusky, R. J. (2000). Digital libraries:
Situating use in changing information infrastructure. Journal of the American Society for Information Science,
51, 394−413.
Bollen, J., Luce, R., Vemulapalli, S.S., Xu, W. (2003). Usage analysis for the identification of research trends in
digital libraries. D-Lib Magazine, 9(5). Retrieved March 15, 2005, from http://www.dlib.org/dlib/may03/bollen/
05bollen.html
452 H. (Iris) Xie / Library & Information Science Research 28 (2006) 433–452
Borgman, C. (1999). What are digital libraries? Competing visions. Information Processing and Management, 35,
227−243.
Borgman, C. L., Leazer, G. H., Gilliland-Swetland, A. J., & Gazan, R. (2001). Iterative design and evaluation of a
geographic digital library for university students: A case study of the Alexandria Digital Earth Prototype
(ADEPT). Proceedings of the 5th European Conference on Research and Advanced Technology for Digital
Libraries (pp. 390−401). London: Springer-Verlag.
Carter, D. S., & Janes, J. (2000). Unobtrusive data analysis of digital reference questions and service at the Internet
Public Library: An exploratory study. Library Trends, 49, 251−265.
Cherry, J.M., Duff, W.M. (2002, January). Studying digital library users over time: A follow-up survey of Early
Canadiana online. Information Research, 7(2). Retrieved March 12, 2005, from http://informationr.net/ir/7-2/
paper123.html
Chowdhury, G. G., & Chowdhury, S. (2003). Introduction to digital libraries. London: Facet Pub.
Heath, F., Kyrillidou, M., Webster, D., Choudhury, S., Hobbs, B., Lorie, M., et al. (2003). Emerging tools
for evaluating digital library services: Conceptual adaptations of LibQUAL+ and CAPM. Article No. 170,
2003-06-09, Journal of Digital Information 4(2). Retrieved November 15, 2005, from http://jodi.tamu.edu/
Articles/v04/i02/Heath/
Hill, L. L., Carver, L., Larsgaard, M., Dolin, R., Smith, T. R., Frew, J., et al. (2000). Alexandria Digital Library:
User evaluation studies and system design. Journal of the American Society for Information Science, 51,
246−259.
Kassim, A. R. C., & Kochtanek, T. R. (2003). Designing, implementing, and evaluating an educational digital
library resource. Online Information Review, 27(3), 160−168.
Krippendorff, K. (2004). Content analysis: An introduction to its methodology. Thousand Oaks, CA: Sage
Publications Inc.
Lancaster, F. W. (1993). If you want to evaluate your library. (2nd ed.) Urbana-Champaign, IL: University of
Illinois.
MacCall, S. L., Cleveland, A. D., & Gibson, I. E. (1999). Outline and preliminary evaluation of the classical digital
library model. Proceedings of the American Society for Information Science Annual Meeting, 36, 33−46.
Marchionini, G. (2000). Evaluation digital libraries: A longitudinal and multifaceted view. Library Trends, 49,
304−333.
Marchionini, G., Plaisant, C., & Komlodi, A. (1998). Interfaces and tools for the Library of Congress National
Digital Library Program. Information Processing and Management, 34(5), 535−555.
Monopoli, M., Nicholas, D., Georgiou, P., & Korfiati, M. (2002). A user-oriented evaluation of digital libraries:
Case study: The electronic journals service of the library and information service of the University of Patras,
Greece. Aslib Proceedings, 54(2), 103−117.
Nielsen, J. (1993). Usability engineering. San Diego, CA: Academic Press.
Saracevic, T. (2000). Digital library evaluation: Toward an evolution of concepts. Library Trends, 49, 350−369.
Saracevic, T., & Kantor, P. (1997). Studying the value of library and information services: II. Methodology and
taxonomy. Journal of the American Society for Information Science, 48, 543−563.
Saracevic, T., & Covi, L. (2000). Challenges for digital library evaluation. Proceedings of the ASIS Annual
Meeting, 37, 341−350.
Shneiderman, B. (1998). Designing the user interface strategies for effective human–computer interaction. (3rd ed.)
Reading, MA: Addison-Wesley.
Su, L. T. (1992). Evaluation measures for interactive information retrieval. Information Processing and
Management, 28, 503−516.
Thong, J. Y. L., Hong, W. Y., & Tam, K. Y. (2002). Understanding user acceptance of digital libraries: What are the
roles of interface characteristics, organizational context, and individual differences? International Journal of
Human–Computer Studies, 57, 215−242.
Van House, N.A., Butler, M.H., Ogle, V., Schiff, L. (1996, February). User-centered iterative design for digital
libraries: The Cypress experience. D-Lib Magazine, 2. Retrieved February 3, 2005, from http://www.dlib.org/
dlib/february96/02vanhouse.html