The following online article has been derived mechanically from an MS produced on the way towards conventional print publication. Many details are likely to deviate from the print version; figures and footnotes may even be missing altogether, and where negotiation with journal editors has led to improvements in the published wording, these will not be reflected in this online version. Shortage of time makes it impossible for me to offer a more careful rendering. I hope that placing this imperfect version online may be useful to some readers, but they should note that the print version is definitive. I shall not let myself be held to the precise wording of an online version, where this differs from the print version.

Published in International J. of Lexicography 13.54–9, 2000.


Christiane Fellbaum, ed., WordNet: An Electronic Lexical Database.  MIT Press, 1998, xxii + 423 pp.  ISBN 0-262-06197-X.

 

Reviewed by:

Geoffrey Sampson, School of Cognitive and Computing Sciences, University of Sussex.

 

 

 

For automatic natural-language processing researchers who need a reasonably comprehensive machine-usable database of the meanings of English words, at present WordNet is the only available possibility. 1

 

There are electronic dictionaries derived from printed English dictionaries, including machine-usable versions of various editions of the Oxford Advanced Learner’s Dictionary (OALD) and of the Longman Dictionary of Contemporary English (LDOCE), but these are useful mainly for grammatical and pronunciation information; the definitions are either left as unformalized passages of ordinary English, as in the printed originals, or even omitted altogether (in the case of Roger Mitton’s computer-usable version of the third edition of OALD, Mitton 1986).  The electronic LDOCE does includes some limited semantic classifications, independent of the contents of the printed version; but this resource has the drawback, for many researchers, of being a commercial product to which access is tightly controlled by the proprietors and, I understand, very expensive.  (Chapter 16 of the volume under review suggests that LDOCE may be the most widely-used machine-readable dictionary, nevertheless.)

 

By contrast, WordNet, initiated by George Miller of Princeton University in 1985 and developed under a series of publicly-funded research grants, is freely available to all.  A flyer from MIT Press slipped into my review copy of this volume offers a CD-ROM containing PC, Mac, and Unix versions of WordNet Release 1.6, with associated software, for US $25, with no mention of restrictions forbidding commercial exploitation or the like; alternatively, Internet addresses are given from which the packages can be downloaded by ftp.  (Release 1.6 is one version on from the release described in the volume under review.) One cannot really call WordNet a “dictionary” in the usual present-day sense, because (apart from a very simple part-of-speech classification) it contains only semantic information; but, within its purview, it is rather comprehensive.  According to this book, the then-current WordNet release contained about 168,000 entries, on the order of double the corresponding figures for electronic versions of OALD or LDOCE.  For research that needs grammatical or phonological information in addition to meanings, perhaps it might be possible to link the semantic data in WordNet with the grammatical and pronunciation data in one of the traditional electronic dictionaries.

 

The volume under review contains an introduction and sixteen chapters by a total of 24 authors (the overwhelming majority of them based in North America) who have been involved either in different aspects of the development of WordNet, or in attempts to exploit it for various NLP applications.  It thus offers an excellent opportunity to take stock of the resource and assess its value.  At present, as already said, there is no real alternative for NLP  researchers who need semantic data for more than a toy subset of English vocabulary.  But does that mean that WordNet is an interim stopgap solution, or is it an approximation to what one would ideally hope to have available?

 

In view of the huge quantities of time and resources that have been devoted by traditional dictionary publishers over many decades to the task of documenting the English lexicon in detail, it seems surprising that a database constructed manually by academics with no access to a dictionary-publisher’s archive could be a serious contender as the leading tool in this domain.  Traditional printed dictionaries and machine-usable lexical databases have different functions, but one would have thought it far easier to achieve the latter by adapting the former rather than by building from scratch, so that Oxford University Press or another publishing house would years ago have made the efforts of this academic team redundant.  That has not happened; and the fact that the Princeton team’s work remains unrivalled seems to me a remarkable tribute to their achievements.  I shall have occasion, below, to make a number of criticisms of WordNet, but my critical remarks should be interpreted in the context of that fundamental point.

 

There is a good reason why George Miller and his team developed their database from scratch rather than basing it on an existing dictionary: they planned it not as a service to NLP research but as a testbed for psycholinguistic theories about mental processing of vocabulary in humans (cf. p. 4 of the editor’s Introduction).  A point which emerges clearly from this book is that the eventual WordNet user base, consisting largely of computer scientists interested in developing practical language-engineering applications, is almost disjoint from the cognitive science clientele for whom the resource was intended.  (George Miller notes on p. 43 that “WordNet has been largely ignored by psycholinguists”.)  This again makes it all the more remarkable that no rivals to WordNet have been produced by organizations more directly involved with information technology, and it explains why the structure of WordNet is not always ideally adapted to current users’ purposes.

 

The structure of WordNet, in brief, is a set of separate networks for different parts of speech, in each of which the nodes of the network correspond either to individual words or to “synsets” (sets of approximate synonyms), and the links between nodes correspond to semantic relationships.  Prominent among the relationships encoded is hyponymy.  Thus, nouns are represented by means of networks for 25 broad classes of noun such as “animal”, “event”, “food”, “shape”; each network consists in large part of a tree structure whose root node corresponds to the general concept, and in which paths leading down from the root traverse nodes representing increasingly specific concepts.  For instance, horse is coded as a hyponym of mammal, and mammal as a hyponym of the top-level concept animal (there are intermediate nodes in both cases).  Hyponymy is not the only relationship coded – the noun networks also encode e.g. relationships of antonymy, and meronymy (part-whole); and the subsets of the networks limited to hyponymy relationships are not perfectly tree-structured (some concepts are treated as hyponyms of alternative concepts which are incommensurable with one another, for instance killing is a hyponym both of termination and of felony).

 

An additional complication is that the network of hyponymy relationships between nouns apparently requires some nodes which correspond to no single items of English vocabulary; for instance, the nouns offender and libertine are treated not as immediate hyponyms of person but as hyponyms of the phrase bad person, which is itself a hyponym of person.  According to Christiane Fellbaum (p. 8), without the bad person node the words offender and libertine “would be (infelicitous) sisters to adventurer, lover, and worker”.  This point perhaps needs a little more spelling out than it receives in the book.

 

The networks for parts of speech other than nouns are based on different relationships.  For verbs, the main relationship is described as lexical entailment: for instance, snore lexically entails sleep because the proposition he is snoring entails the proposition he is sleeping.  (Hyponymy in nouns might be regarded as a special case of entailment.)  The network for adjectives is organized very differently from those for other parts of speech, because the hyponymy concept is felt to be irrelevant to adjectives.  (According to Katherine Miller, p. 48, “it is not clear what it would mean to say that one adjective ‘is a kind of’ some other adjective”.  Is scarlet not a kind of red?  But perhaps this type of relationship is less pervasive among adjectives than among other words.)  For adjectives, the main relationship coded is antonymy, and this reflects the psychological interests of the project initiators: when subjects are asked to produce a word associated with a stimulus word, the response to an adjective stimulus is commonly its antonym.

 

Some of the earlier chapters in the book discuss the organization of the database in terms of what is displayed on screen when a word is looked up.  This seems a peripheral issue: WordNet is not really a tool for manual investigation of individual words (a traditional dictionary is better for that).  But later chapters discuss experiments which exploit WordNet to develop automatic systems for executing various NLP tasks that require semantic data, for instance information retrieval, or spelling correction in cases where a mistyped word coincides with the correct spelling of another word.  These experiments are all quite preliminary – at this stage, they could hardly be otherwise; but they do give the reader a clear impression of what WordNet has to offer.  (The experimenters also make many suggestions about ways in which the resource might be improved.)

 

The fact that WordNet has not appealed to psychologists seems unsurprising.  To be taken seriously as a scientifically valid model of how the language actually functions semantically for its speakers, it would need to embody some awareness of philosophical logic, a topic which is never mentioned in this volume.  This gap is admittedly not unique to the WordNet project.  It is a long-standing puzzle to me how completely the insights of “ordinary language philosophy” have been overlooked in recent decades, after forming one of the dominant academic paradigms from the Second World War until about 1975.  Anyone who engaged in abstract thought about language in those years found himself confronted at all turns by the principle, expounded in England by Ludwig Wittgenstein and in America by Willard Quine, that there is no real difference between “dictionary knowledge” and “encyclopaedia knowledge” (a distinction whose reality is assumed again and again by the writers in the volume under review).  As Wittgenstein put it, “what to-day counts as an observed concomitant of a phenomenon will to-morrow be used to define it” (Wittgenstein 1958: §79).  This principle has never, I believe, been seriously challenged, just forgotten.  It seems to imply both that an enterprise like WordNet is ultimately doomed as a model of psychological reality, because semantic relationships between words are too fluid and mutable to be pinned down in the way it aims to do, and also that the temporary relationships which do obtain are not restricted to any limited set of relation-types such as hyponymy and meronymy: the fact that one statement entails another, and hence helps to define the meanings of the words involved, tells us nothing about the grammatical or logical structures of the statements.  Within the WordNet community, this issue is apparently known as the “tennis problem” (the meanings of tennis, ball, racquet, net, etc. are perceived to be connected but are not related by hyponymy or other relationships recognized by WordNet); it is not a problem which the WordNet approach could hope to solve.  In fact, with respect to philosophical logic the system is so naive that it evidently (e.g. p. 134) recognizes no distinction between the species/genus relationship, as in horse/animal, and the individual/universal relationship, as in Shakespeare/author, treating both indifferently as cases of “hyponymy”.

 

But these obstacles to accepting the network system as a scientifically valid model of speakers’ psychology are not reasons for dismissing the usefulness of the system in practical language-engineering contexts.  After all, traditional paper dictionaries do an excellent job of defining the meanings of words, despite Wittgenstein.  They use rules of thumb rather than scientific theories to decide where to draw boundaries between “dictionary” and “encyclopaedia” knowledge.  Not all the properties mentioned in traditional dictionary definitions can be reduced to the WordNet relations of hyponymy, meronymy, etc.; but perhaps a useful proportion of them can be.  For NLP purposes, systems that work fairly well for much of the time are much better than no systems.

 

Even at this banausic, practical-approximation-to-God’s-truth level, the book under review does raise doubts in my mind about some aspects of WordNet, but these relate to possible faults which in principle could be remedied.

 

In the first place, some of the relationships between particular words which are quoted as examples seem to assume rather specialized registers of discourse.  For instance, George Miller identifies beak as a hyponym of jaw (p. 38), and seems to imply that fish is a hyponym of animal (p. 32).  It may be that, to an evolutionary biologist, a bird’s beak is a kind of jaw, but I do not think an average speaker of English understands the words that way – the example came as a surprise to me when I read it.  Likewise, in everyday usage the word fish contrasts with animal rather than entailing it.

 

In cases like these the response might be that, if the average speaker does not count a beak as a jaw, biology has shown that he ought to.  Whether this is a satisfactory response will presumably depend on the nature of the applications for which WordNet is used.

 

More worrying are cases where the claims made about English seem just wrong, for any discourse register.  The book contains examples relating both to meaning and to usage.

 

In the case of meaning relationships, for example, Katherine Miller (p. 50) uses weighty v. weightless to exemplify antonymy, and airy v. heavy to exemplify “conceptual opposites that are not lexically paired”.  Neither of these examples works for any style of English I know.  Weighty is an abstract term referring to significance, real or purported; weightless refers exclusively to physical lack of gravitational effects, commonly in connexion with space travel.  Airy has different meanings in, say, “airy room” and “airy comment”; in the former case, a rough opposite might be stuffy, in the latter perhaps considered or the like, but heavy seems well off target in either case.

 

In the area of grammar and usage, things get worse.  Shari Landes et al. quote, as their paradigm example of erroneous part-of-speech tagging, the classification of “the adverb beyond in the phrase rich beyond her wildest dreams” as a preposition (pp. 209-10).  There is some looseness in the way that different schools of linguists fit the traditional part-of-speech names to usage in awkward cases, but, in any version of grammar I have encountered, beyond in this use is a preposition – it is not an awkward case.  Again, chapter 6, by Karen Kohl et al., discusses aspects of the WordNet verb networks which reflect the clause patterns in which different verbs occur.  The authors contrast clauses which are “ungrammatical”, such as the witch turned him from a prince (i.e. changed him from a prince into something else), with clauses which allegedly are grammatical, such as the spaceship revolves around the earth.  This latter example is used repeatedly in the chapter; but to my mind (and I have confirmed this with one other English native speaker) it too is strange English.  A satellite can “orbit” or “circle” the Earth, or “travel” or “move” round the Earth, but it cannot “revolve (a)round” the Earth.

 

Obviously, in a system on as large a scale as WordNet that was built up from nothing over a few years, there are bound to be many errors waiting to be eliminated.  But it is a little disappointing that the researchers responsible for developing the system are not sensitive enough to the language to avoid such cases when choosing illustrative examples.  Quite often in this book one encounters passages which read as if modern English is a remote, imperfectly-understood phenomenon, rather than the everyday medium of communication of almost all these writers.

 

So, in summary, on the evidence of the book I suspect that WordNet may be more a stopgap than an ideal solution to the need for a machine-usable dictionary of English word meanings.  But, as stopgaps go, it is clearly a very rich and comprehensive one.  Its creators deserve admiration for what they have achieved over so short a period.

 

Finally, I should like to make a plea that relates not to the content of WordNet but to the language used in discussing it.  The writers frequently need to refer to the converse of “hyponym” – the relation which animal bears to horse.  The editor’s Introduction uses the term “superordinate”, and Sanda Harabagiu and Dan Moldovan at one point use the term “subsumer”, but otherwise all these writers repeatedly use the neologism “hypernym”.  This is a barbarous formation: the Greek root for “name” is onym-, not nym-, so obviously the form ought to be “hyperonym”.  It is also thoroughly impractical: “hyponym” and “hypernym” are opposites with identical pronunciations.  Sir John Lyons’s standard work Semantics argues that even “hyperonymy”, although current, is too similar to “hyponymy” to avoid confusion, and he advocates the term “superordination” (Lyons 1977: vol i, p. 291).  But “hyperonym” is nowhere near as confusing as “hypernym”.  Can we hope that this horrible word will be banished from future lexical discourse?

 

 

References

 

Lyons, J.  1977.  Semantics (2 vols.).  Cambridge University Press.

Mitton, R.  1986.  “A partial dictionary of English in computer-usable form”.  Literary and Linguistic Computing 1.214-15.

Wittgenstein, L.  1958.  Philosophical Investigations, 2nd ed.  Blackwell (Oxford).



1I thank my Sussex colleague Diana McCarthy for comments based on her experience of working with WordNet.  Any errors in this review are mine.