文档介绍:Informing Science Journal Volume 8, 2005
Developing a Framework for Assessing
Information Quality on the World Wide Web
Shirlee-ann Knight and Janice Burn
Edith Cowan University, Perth, Australia
s.******@, j.******@
Abstract
The rapid growth of the as an environment for information exchange and the lack of en-
forceable standards regarding the information it contains has lead to numerous information qual
ity problems. A major issue is the inability of Search Engine technology to wade through the vast
expanse of questionable content and return "quality" results to a user's query. This paper attempts
to address some of the issues involved in determining what quality is, as it pertains to information
retrieval on the . The IQIP model is presented as an approach to managing the choice and
implementation of quality related algorithms of an crawling Search Engine.
Keywords: Information Quality, IQIP, Data Quality, Information Retrieval, Search Engines
Introduction – The Big Picture
Over the past decade, the 1 – or World Wide Web (Technically the is a huge collection of
puters using TCP/IP protocol to exchange data. The World-wide Web (WWW) is in essence only part
of work puters, however its visible status has meant that conceptually at least, it is often used inter-
changeably with "" to describe the same thing.) – has established itself as the key infrastructure for
information administration, exchange, and publication (Alexander & Tate, 1999), and
Search Engines are the monly used tool to retrieve that information (Wang, 2001). The
deficiency of enforceable standards however, has resulted in frequent information quality prob-
lems (Eppler & Muenzenmayer, 2002).
This paper is part of a research project undertaken at Edith Cowan, Wollongong and Sienna Uni-
versities, to build an Focused Crawler that uses "Quality" criterion in determining returns
to user queries. Such a task requires that th