Where is Meaning when Form is Gone?
Knowledge Representation on the Web

Terrence A. Brooks

Information School
The University of Washington
Seattle, WA 98195
e-mail: tabrooks@u.washington.edu

Legacy knowledge representation methods have assumed document-like embodiments (Svenonius, 2000, p.8) of information: A librarian tags a book with a subject heading. How well does this paradigm transfer to the Web? New extensible Web technologies for marking up resources seem to have effectively antiquated the document as a vessel of information. If the document paradigm does not transfer to the Web, where shall we locate our signifiers of meaning such as subject headings, or the Web equivalent, meaning-bearing metadata?

This essay suggests that new locales of meaning in extensible technologies may be:

  • The structure of the information resource itself, implicitly in the markup tags, and explicitly as meaning-bearing attribute qualifiers


  • The relationships among information resources: implicitly in the links, and explicitly as meaning-bearing attribute qualifiers


  • Situational expertise that orients information seekers to the semantic norms of a specific meaning community

Legacy Technologies of Knowledge Representation

Typical technologies of knowledge representation in the 20th Century were the Machine Readable Cataloging (MARC) record structure, the Library of Congress Subject Headings (LCSH) and the WorldCat database. These technologies reflect a historic transition from paper-based to digitial systems, and express assumptions such as:

  • A uniform catalog entry can represent holdings The MARC record acts as a uniform record structure for the representation of disparate information types, including manuscripts, archives, cartographic material, musical scores, serials, sound recordings and so on. MARC is an "integrated format defined for the identification and description of different forms of bibliographic material." (The MARC 21 formats)


  • A single list of subject terms can provide subject access. The LCSH acts as a general list of subject terms and phrases. "As an increasing number of other libraries have adopted the Library of Congress subject headings system, it has become a tool for subject indexing of library catalogs in general." (Library of Congress Subject Headings)


  • A single database can store information. The WorldCat database acts as a single repository for information. As of June 2000, WorldCat hosted 40 million MARC records. It "is the most consulted database in academe." (Smith, 1996, p. 1)

These three technologies are merely representative. Other dominant 20th Century information providers such the Dialog Corporation reflect similar assumptions. The Dialog Corporation vends access to approximately 500 databases, each of which may have a unique record structure, and a list of subject terms or descriptors. In the early days of online information retrieval, it was common to refer to specific databases as "the medical database" (i.e., MEDLINE, Dialog database 154) or "the education database" (i.e., ERIC, Dialog database 1).

In retrospect, the 20th Century appears as an era when information oligarchs amassed large, singular databases and then vended access. This strategy provides a legacy technological environment of knowledge representation characterized by

  • Universal and time-invariant record structures


  • A disinterested group of information professionals (i.e., librarians) who developed and employed subject terms and phrases


  • A database-oriented assumption that the most efficient strategy for dealing with information was aggregating it into large information stores.

The Technological Environment of the World Wide Web

Since the introduction of the Hypertext Markup Language (HTML) in 1990, the World Wide Web has become a major information utility. Theoretically at least, the Web permits anyone anywhere to post pages on any topic and in any language. Such extreme decentralization makes estimating the Web's size and rate of growth difficult. A survey by Lawrence and Giles (1999) estimated 800 million public Web pages available in February 1999. In June 2000 the search engine Google claimed an index of 1 billion URLs (Google, 2000). Jacob Nielsen (1995) suggests that the growth rate of the Internet is 100 percent per year. Whatever the exact figures may be, the Web is a large, heterogeneous, decentralized phenomenon with a explosive rate of growth.

So does this mean Google is the first search engine to give 100 percent coverage of the web? No. For one thing, that 1 billion page estimate is several months old, and the web has almost certainly increased in size since then. Nor does that estimate include the millions of pages that search engines typically don't crawl, such as those behind password protected areas or served up by identifiable dynamic delivery systems. How big the web is now is anyone's guess. (Sullivan, July 5, 2000)

Some parts of the Web exhibit a high rate of content churn (Brewington and Cybenko, 2000). Speaking of his survey, Brewington estimated that 20% of Web pages are less than twelve days old, while only 25% are older than one year (Markoff, May 29, 2000). An earlier survey by Douglis, Feldmann and Krishnamurthy (1997) found 16.5% of Web pages to be under constant update. An increasingly large number of Web pages are produced "on the fly" by database manipulations. Sherman (1999) calls this "the invisible Web" and concludes that "this trend is going to make it even harder for search engines to be comprehensive Web indexes." The size of the invisible Web is essentially unknown, but may be vast (Abreu, September 11, 2000). BrightPlanet estimates that the invisible Web is five times the size of the visible Web. "Using Google as a benchmark, that means BrightPlanet would estimate there are about 500 billion pages of information available on the web, and only 1/500 of that information can be reached via traditional search engines" (Sullivan, August 2, 2000).

As the Web grows in size, timely delivery of content becomes a problem. Fisher (April 17, 2000) describes two strategies used to speed content delivery: caching popular content (the approach used by Inktomi) and using distributed servers (the approach used by Akamai). Many Web pages are assemblages of cached and variously distributed material. "When a user in Singapore, say, clicks on a popular page in Yahoo, only the first request goes to Yahoo's server in Palo Alto, Calif.; the balance of the page is then delivered from an Akamai server with the shortest, fastest connection to the person in Singapore" (Fisher, 2000, p.C1).

The Web's technological environment is a wholly decentralized phenomenon that springs from many communities, many authors, many languages and points of view. The Web is growing quickly with a significant proportion of information sources exhibiting high content churn. What appears in a Web browser as a static, "document-like" object may have been produced by a combination of dynamic scripts or programs, various database manipulations, with content possibly originating from caching and/or distributed processing. Finally, it should be noted that Web browsers, themselves, exhibit idiosyncratic qualities that may alter the appearance of Web pages depending on their abilities to support scripts, Applets, cookies, dynamic HTML, cascading style sheets, extensible markup and so on.

Legacy Technologies Applied to the Web

It is a truism, perhaps, that we seldom recognize the radical nature of new technologies and prefer to view them as merely extensions of more familiar, older technologies. This impulse expresses itself in the attempts to catalog the Web or develop a single subject topical scheme for Web pages.

The NetFirst database is an attempt to catalog the Web by creating a database of MARC records. To date, volunteer Web surfers have contributed approximately 150,000 MARC records (Greene, June 16, 2000). The CORC project combines the efforts of 489 libraries in 24 countries to build a database of Web pages useful to libraries. "The integration of CORC and WorldCat will create a rich, robust database shared on a global scale, making each library's unique material available to library users worldwide" (Surface, 2000, p. 33). CORC has approximately 26,000 records. The present size and rate of growth of the Web described above compared to the small size of these projects underscores the labor-intensive quality of Web cataloging, and why it is a strategy appropriate for only small pools of relatively static Web content.

Resource discovery on the Web has developed into a major problem with many searches swamped by thousands of false drops. Considerable activity developing metadata schemes has attempted to address this problem. HTML metadata are terms and phrases located in the <HEAD> element of a Web page using the NAME and CONTENT attributes of the <META> element. The ambition of HTML metadata is the addition of subject topical terms and phrases to Web pages, thus emulating the legacy strategy of adding subject topical terms and phrases to cataloging records.

Metadata is data about data. The term refers to any data used to aid the identification, description and location of networked electronic resources. Many different metadata formats exist, some quite simple in their description, others quite complex and rich. IFLA Digital Libraries: Metadata Resources

Numerous user communities have attempted to employ metadata schemes to control their particular data, examples being the Nordic Metadata Project, the Arts and Humanities Data Service, and the United States Federal Statistics project. Readers are directed to the IFLANET International Federation of Library Associations and Institutions Web site for a more complete listing. Since 1995, a series of workshops has promoted the Dublin Core Metadata Initiative as the standard metadata tag set. Dillon has provided a very useful and comprehensive discussion of how Dublin Core Metadata might address the problem of identifying Web resources. He strongly urges a refocus towards the development of a "MARC version of Dublin Core."

To date the employment of meaning-bearing metadata has been meagre.

O'Neill, Lavoie and McClain (May 26, 2000) sampled 1, 024 homepages and found only seven using Dublin Core metatags. Lawrence and Giles (1999) reported low metadata use, finding only 0.3% of sites use the Dublin Core metadata standard. At this time no major Web search engine supports topical metadata. (Taylor, April 1, 1999)

Current metadata usage patterns are a long way from comprehensive document description at the page level. Finally, most metadata usage is still ad hoc; with a few exceptions, most sites do not adhere to a well-defined set of metadata elements. (O'Neill, Lavoie and McClain, May 26, 2000)

The idea of a particular user group customizing its data is a powerful one, as is the idea of a controlled set of terms and phrases used to advantage in Web resource discovery. Two powerful false assumptions, however, seem to block the success of current metadata efforts:

  • False Community Assumption The technological legacy of knowledge representation assumed the existence of a class of disinterested information workers (i.e., librarians) to develop and apply subject cataloging. The decentralized Web lacks such a disinterested class of information workers. Quite the contrary, the Web is composed of millions of individuals who can markup their pages in any manner they wish. Web resource providers, competing for attention to their Web sites, will use meaning-bearing metadata to gain an advantage for site promotion. The Web lacks community norms to prevent this behavior. Search engines avoid meaning-bearing metadata because the unscrupulous will spoof their meaning-discovery algorithms with untrustworthy information (Taylor, April 1, 1999).

  • False Document Assumption Current metadata strategies are designed for "high-level document properties" (Lander, 1998). Placing topical terms and phrases in the <HEAD> element of an HTML document assumes that the semantic content of the <HEAD> element will maintain a time-invariant relationship with the semantic content of the <BODY> element. This assumption is reasonable if one conceives of Web pages as merely paper documents digitized. The preceding sketch of Web technology suggests that the legacy metaphor of paper documents and record structures does not fit Web pages very well. While there may always be a residue of static HTML pages, the majority of future Web pages will reflect the efficiencies of database manipulations and extensible markup. For example, Guernsey (July 18, 2000) describes the deconstruction and vending on the Web of "chapters, maps and even paragraphs" that in the legacy information environment would have been indivisible parts of the static information unit called a book.

This review of legacy knowledge representation technologies applied to the Web suggests that meaning-bearing metadata would be best employed within a strongly normative community, and in a manner that did not rely on the legacy concept of the document.

The family of extensible markup technologies permits specific communities to set norms as to the structure and semantics of their data, and is furthermore free of any legacy document-like assumptions. It is argued below that meaning-bearing metadata would be better placed as part of the attributes of an Extensible Markup Language (XML) representation of information.

Extensible Information Technologies

HTML mixes content and presentation tags, a design that reflects its original purpose of displaying scientific papers, but makes general data sharing awkward. Separating content from presentation permits data to be gathered without the baggage of presentation tags, and eases the consistent styling of data from different sources. XML heralds the arrival of the "second-generation Web" (Bosak and Bray, May 1999) and "The era of the distributed object" (Cagle, October 26, 1999). The reader is referred to Qin (2000) who traces the development of information technology from MARC records to XML.

Extensible Markup Language (XML) 1.0

An XML resource is a file of text strings, a format that facilitates data sharing. The text strings are defined in semantic markup, arbitrary tags that express a particular user's semantics of the data. In Figure 1,

Figure 1: An XML resource

<?xml version="1.0" standalone='yes'?>
<recipe>
     <name>Banana Bread</name>
          <ingredient>water</ingredient>
          <ingredient>flour</ingredient>
          <ingredient>bananas</ingredient>
</recipe>

a banana bread recipe is represented with various arbitrary tags of <recipe>, <name> and <ingredient> that might suit a baker's application.

XML elements are modified by attributes, which are string name-value pairs. Figure 2 illustrates three attributes that emulate Dublin Core metadata and one that is user defined :

Figure 2: An XML resource with attributes

<?xml version="1.0" standalone='yes'?>
<recipe DCCreator="Susan Cheney" DCDate="1990" DCSubject="Cookery(Bread)">
     <name>Banana Bread</name>
          <ingredient>water</ingredient>
          <ingredient>flour</ingredient>
          <ingredient number="4">bananas</ingredient>
</recipe>

In his recipe XML resource, a baker could add many recipes, notes, observations and other types of data.

The revolutionary aspect of XML is the modularization of information. Information presents itself as a self-describing unit that can does not inhibit processing, storing or display. Topical subject qualifiers (e.g., attributes) are placed at the appropriate level of granularity: Recipe qualifiers are placed the at <recipe> level while ingredient qualifiers are placed at the <ingredient> level.

Extensible information technologies antiquate the legacy concept of document:

As more and more information becomes available in XML format (and as the mechanism for referencing them gets sufficiently defined) then applications become truly transparent to the notion of servers -- a single XML 'document' could conceivably span hundreds or thousands of servers, in such a way that the physical task of locating a document becomes a secondary consideration at best.
The upshot of this is that the paradigm that we used to think about the Internet, about documents, and about the nature of information changes radically. Agents, XML code blocks that retain their integrity irregardless of their point of origin, roam the Internet as autonomous units in a sea of contextual relationships. (Cagle, October 26, 1999)

This essay springs from the question where meaning may reside when form is gone. The preceding survey of the structural form of XML provides part of the answer. Meaning resides in the semantic structure of information. Meaning can also reside in the meaning-bearing terms and phrases placed at the appropriate level of grandularity that serve to qualify a specific element of information.

XML Linking Language (XLink) 1.0

XLink is a candidate recommendation as of July 3, 2000 that describes the linking relationships among XML resources. It generalizes the HTML unidirectional links to multidirectional links among two or more resources, or portion of resources. XLinks can be qualified by attributes, which is pertinent to this enquiry of the residence of meaning.

Figure 3, an example from the XLink recommendation, illustrates two XLink standard attributes and one user-defined attribute:

Figure 3: An XLink resource with both XLink and non-XLink attributes

<my:crossReference
  xmlns:my="http://example.com/"
  my:lastEdited="2000-06-10"
  xmlns:xlink="http://www.w3.org/1999/xlink"
  xlink:type="simple"
  xlink:href-"students.xml">
Current List of Students
</my:crossReference>

XLinks facilitate rich links among extensible resources. Linkages themselves carry some semantic information that can be magnified by terms and phrases that provide a semantic context for linking. XLinks provide another locale for meaning. Meaning can also reside in the qualifications of the relationships among resources.

Normative Meaning Communities Using Extensible Information Technologies

This essay has argued that metadata would best be employed in a strongly normative community that did not rely on the legacy concept of the document. Examples of strongly normative communities are internet-based electronic marketplaces (Bakos, 1998). The ambition of an electronic marketplace is to share information in the most efficient possible manner, ultimately creating "friction-free" marketplaces for goods and services.

The participants in a marketplace for specific goods and services compose a more strongly normative community than a random collection of Web content providers. The assumption is that the trust required in a selling/buying transaction would penalize spoofing behavior. In short, the assumption is that a seller who spoofed his buyers would generate bad word of mouth and would not be in business for long.

As noted above, many groups are developing their own metadata, as opposed to using the Dublin Core set. The appropriate place of these metadata qualifiers is not in the <HEAD> element of an HTML document, but as element attributes in extensible information resources as illustrated in Figure 2. A motivating feature of this employment of metadata is that it is at the appropriate level of grandularity, targeting only specific descriptions of goods and services.

Table 1 lists several consortia that promote electronic marketplaces. Consortia provide services such as the registration of XML schemas. A repository of XML schemas provides models for the newcomers and detailed specifications for sharing or searching for data. As an example, OASIS, the non-profit XML interoperability consortium, maintains an XML.ORG Registry

In the five days since we began accepting registrations, OASIS has had organizations and companies from Australia, Canada, Germany, India, Japan, Korea, Pakistan, the Ukraine and the United States--all wanting to register as submitters. We are working now to validate their submissions and will be soon be inviting users to access the XML.ORG Registry to find schemas for their particular needs. Goldfarb (June 26, 2000)

Table 1: Consortia promoting extensible information technologies
BizTalk A Microsoft-backed consortium for the development and distribution of the BizTalk flavor of business-oriented XML schemas
CommerceNet Defines specifications to facilitate the interoperability of information and integration of content and services across and between vertical markets
FinXML A consortium supporting the creation and management of the FinXML language for the integration and exchange of digital information in capital markets
Organization for the Advancement of Structured Information Standards Nonprofit, international consortium of companies and organizations dedicated to accelerating the adoption of product-independent formats based on public standards
RosettaNet Standardizes the mechanisms used to define the business processes of vertical markets

Table 2 gives examples of extensible information intiatives beyond consortia.

Table 2: Example normative communities based on extensible information technologies
Get There Internet-based B2B travel procurement solutions for corporations, travel suppliers, portals and corporate mobile travelers
Acord Software Directory Financial services industry
Commerce XML Commerce resources
Financial Information eXchange Real-time electronic exchange of securities transactions
American Institute of Certified Public Accountants XML-based specification for the preparation and exchange of financial reports and data
adXML An international, open standard organization, which is defining an advertising XML schema for both on-line and off-line media
loanupdate Collaborative transaction management product for the mortgage industry

Situational Expertise

Knowledge sharing sites on the Web function as forums or brokers for the exchange of expert or everyday wisdom. MindCrossing may be considered a model. It has a stable of subject topical experts who have created "MindStores." A MindStore is a Web site with articles, best practices, case studies, technical specifications and so on. Some of this content is free, some requires payment.

Situation expertise is triggered by visiting a Web information marketplace and searching on a term. A responding tablet on the browser screen alerts the novice user that expertise about this subject is available (for example, see the MindCrossing demonstration)

Context-sensitive situational expertise can orient users to appropriate metadata, concepts and technical vocabulary.

Table 3: Sources of situation expertise
Allexperts Created in 1998, was the very first large-scale question and answer service
Askme Provides custom answwers to specific questions
Epinions Offers unbiased advice on over 100,000 products and services
EXP EXP connects individuals to experts in hundreds of categories

Busch and Reisman suggest that the most successful Web marketplaces are those that develop "deep, industry-specific knowledge or specialized, industry-specific supply-chain capabilities." This essay has presented the case that the place of meaning is no longer in legacy information forms, but in extensible information technologies and situated expertise that will facilitate such deep, industry-specific knowledge.

Summary

Extensible information technologies are creating new ways of structuring information and linking information resources. This essay has argued that the legacy knowledge representation methods are no longer appropriate since they are based on an antiquated, technological paradigm of massive, singular databases of highly structured, identical records.

Extensible information technologies enjoy significant advantages such as the modularization of information, semantic information structures, qualifiers (i.e., metadata) placed at the appropriate level of grandularity, and semantic relationships among information resources.

As we lose familiar metaphors such as the "document" and address ourselves to the problem of locating meaning in the second-generation Web, this essay suggests that future residences of meaning will be:

  • The structure of an information store

  • Qualifiers of elements of an information store

  • Relationships among information stores

  • Expert opinion interpreting the structure, qualifiers and relationships of an information store.

References

Abreu, E. (September 11, 2000). "Diving into the deep web." The Industry Standard, v.3, no.35, 119.

Bakos, Y. (1998). "The emerging role of electronic marketplaces on the Internet." Communications of the ACM, v.41, no.8, 35-42

Bosak, J. and Bray, T. (May 1999). "XML and the second-generation Web." Scientific American. [Accessed July 7, 2000]

Brewington, B. E. and Cybenko, G. (January 29, 2000). "How dynamic is the web?" [Accessed June 28, 2000]

Busch, J. and Reisman, L. (July 24, 2000). "B-to-B exchanges: know your domain" The Industry Standard, v.3, no.27, 96.

Cagle, K. (October 26, 1999). "Why XML? A look at XML and how it will change the world." [Accessed June 19, 2000]

Dillon, M. (2000). "Metadata for web resources: How metadata work on the web." [Accessed September, 2000]

Douglis, F., Feldmann, A., & Krishnamurthy, B. (1997). "Rate of change and other metrics: A live study of the World Wide Web." [Accessed June, 2000]

Fisher, L.M. (April 17, 2000). "2 companies take separate paths to speed delivery of Web pages." The New York Times, p. C1-C4.

Goldfarb, C. F. (Monday, June 26, 2000). "XML community rallies behind XML.ORG Registry public clearinghouse for XML schemas and vocabularies takes off." [Accessed June 26, 2000]

"Google Launches World's Largest Search Engine." (June 26, 2000). [Accessed June 28, 2000]

Greene, R. (June 16, 2000). "Database load and diacritics and UNICODE." Personal e- mail from Richard Greene, greenr@OCLC.ORG

Guernsey, L. (July 18, 2000). "Books by the chapter or verse arrive on the Internet this fall." The New York Times, p. 1-C6.

Digital Libraries:Metadata Resources IFLANET: International Federation of Library Associations and Institutions. [Accessed July 7, 2000]

Lander, R. (January 1, 1998). "The search for metadata." [Accessed June 19, 2000]

Lawrence, S. & Giles, L. (1999). "Accessibility and distribution of information on the web." Nature, 400, 107-109. [Accessed June, 2000]

Library of Congress "Subject Headings - Principles of Structure and Policies for Application: Contents" [Accessed June, 2000]

The "MARC 21 Formats: Background and Principles Revised November 1996." [Accessed June, 2000]

Markoff, J. (May 29, 2000). "As web expands, search engines puff to keep up." The New York Times, CXLIX, p. C3.

Nielsen, J. (1995). "Kill the 53-day meme." [Accessed June 28, 2000]

O'Neill, E.T., Lavoie, B.F., & McClain, P.D. (May 26, 2000). "An analysis of metadata usage on the web." [Access July 7, 2000]

Qin, J. (2000). "Representation and organization of information in the Web Space: From MARC to XML." Informing Science, v.3, no.2.

Sherman, Chris (June, 1999). "The Invisible Web." [Accessed June 29, 2000]

Smith, K. W. (1996). "OCLC - Moving toward the next stage of the electronic library." In Proceedings of the Fourteenth Annual Conference of Research Library Directors. Tomorrow's Access-Today's Decisions: Ensuring Access to Today's Electronic Resources (pp. 1-5). Dublin, OH: OCLC Online Computer Library Center.

Sullivan, D. (July 5, 2000). "The search engine report", number 44, part 1 of 2. [Accessed July 6, 2000]

Sullivan, D. (August 2, 2000). "The invisible Web gets deeper." Accessed August 4, 2000.

Surface, T. (July/August 2000). "CORC: Build locally, share globally." OCLC Newsletter, no. 246.

Svenonius, E. (2000). The Intellectual foundation of information organization. Cambridge, MA: The MIT Press.

Taylor, C. (April 1, 1999). "An introduction to metadata." [Accessed March 17, 2000]

###