A Grammar of English on Mathematical Principles

Zellig Sabbettai Harris

John Wiley and Sons, New York. 1982. Pp. xvi, 429. $43.50.

Reviewed by Bruce Nevin

[Computational Linguistics, Volume 10, Numbers 3-4, July-December 1984 203-211]

This is the most comprehensive exemplification to date of Zellig Harris’s formal linguistics. As such, it merits careful study by anyone seriously interested in the scientific study of language, and in particular by anyone working with computers to analyze or use natural language. The intrinsic interest Harris holds for computational linguistics stems chiefly from:

Together, these characteristics suggest an approach to computational parsing and synthesis that could be both highly efficient and semantically sensitive. Beyond that, they indicate avenues for design of artificial languages and language-like systems that have yet to be tried.

Somewhat less than the first half of the book (chapters 1-3, pages 1-185) presents H’s model of language. The remaining 228 pages (chapters 4-9 plus the Appendix) restate the categories and phenomena of traditional grammar in some detail as both a demonstration and a test of that model.<1> It is a densely written book, and astonishingly comprehensive. Every page is filled with information and insights enough to be expanded to a journal article, reflecting a grasp of syntactic and semantic data that is of extraordinary depth and breadth, both synchronically and diachronically. The reader should not assume that H neglects a particular problem simply because he does not summarize his solutions under a familiar label, such as Neg Raising.<2> On the contrary, even readers who are loath seriously to entertain an alternative paradigm in linguistics will profit from study of this encyclopedic restatement of grammar.

There are a number of reasons, to be sure, why scholars may prefer to ignore H’s work. Many readers have found H’s prose tough sledding. As Jane Robinson once observed (pc):

Let those readers who are for this reason reluctant to tackle another Harris opus be assured. The writing here, in addition to being explicit and unambiguous, as always, is also quite clear and straightforward. Even in his justifications of the more complex derivations, where what he is trying to say is indeed difficult, one’s attention span need be only somewhat longer than usual to hold the strands of argumentation in mind. Indeed, complex argumentation can be no stop to any student of generative linguistics! No, remaining difficulties for the reader will be due to the unfamiliarity of the approach, which H himself cites (page vi) as his reason for discussing some ‘methodological considerations and grammatical constructions’ redundantly in more than one section of the book.

More serious is the question of familiar jargon and shared terms for debate. As I suggested above, he does not use the jargon of generative linguistics. In a field, and indeed in a social climate, where there is great competition to be ‘more state of the art than thou’, the obvious assumption is that H has simply not kept up.

Not to beat around the bush, the central issue in any evaluation of H’s work in linguistics is that it has been eclipsed by that of his most famous student, Noam Chomsky. More especially, it has been obscured by the spirit of polemic that has consumed the field of linguistics in this country for the past twenty or twenty-five years. H is both stubbornly unpolemical and stubbornly autonomous. These characteristics have unfortunately contributed to H’s contributions going largely unnoticed in the United States (though not abroad). Perhaps it is time to allow the curtain to fall on this Oedipal drama and evaluate H’s work on its merits.

After all, H’s isolation from ‘mainstream linguistics’ has also had its advantages. Undistracted by academic politics and the demands of cross-paradigmatic refutations, he has been painstakingly "formulating as a mathematical system all the properties and relations necessary and sufficient for the whole of natural language [rather than] investigating a mathematically definable system [such as phrase-structure grammar and transformational systems based upon it] which has some relation to language, as being a generalization or subset" (Harris 1968: 1). He now offers results that linguists and computational linguists should pay close attention to, at a time when many far less well-developed alternatives rush to fill a perceived theoretical vacuum in the field.

The limitations of a review preclude a detailed comparison of H’s model with more familiar generative models, but a sketch at least is in order. In this review, I refer to the Harrisian model of language as ‘constructive grammar’ and to the Harrisian paradigm for linguistics as ‘constructive linguistics’.<3> A constructive grammar has at least the following six characteristics:

1 . The semantic primes are words in the language, a base vocabulary that is a proper subset of the vocabulary of the language as a whole.<4>

2. Generation of sentences in the base is by word entry, beginning with entry of (mostly concrete) base nouns. The only condition for a word to enter is that its argument requirement must be met by some previously entering word or words, generally the last entry or entries, which must not already be in the argument of some other word. The base vocabulary has thus a few simple classes of words:

N — base nouns with null argument
On, Onn, Onnn — operators requiring base nouns N as arguments
Oo, 0oo — requiring operators as arguments
Ono, Oon — requiring combinations of operators and base nouns

In addition to these classes, almost all of the operators require morphophonemic insertion of ‘argument indicators’ such as -ing and that. (These were termed the ‘trace’ of ‘incremental transformations’ in Harris 1965 and 1968.)

3. The base generates a sublanguage,<5> which is informationally complete while containing no paraphrases. This is at the expense of redundancy and other stylistic awkwardness, so that utterances of any complexity in the base sublanguage are unlikely to be encountered in ordinary discourse. As in prior reports of H’s work, base sentences are all assertions, other forms such as questions and imperatives being derived from underlying performatives I ask, I request, and the like.

4. A well-defined system of reductions yields the other sentences of the language as paraphrases of base sentences.<6> The reductions were called the ‘paraphrastic transformations’ and ‘extended morphophonemics’ in earlier reports. They consist of permutation of words (movement), zeroing,<7> and morphophonemic changes of phonological shape. Each reduction leaves a ‘trace’ so that the underlying redundancies of the base sublanguage are recoverable.<8> Linearization of the operator-argument dependencies — in English either ‘normal’ SVO or a ‘topicalizing’ linear order — is accomplished by the reduction system, not the base. The reduction system includes much of what is in the lexicon in generative grammar (cf. Gross 1979).

5. Metalinguistic information required for many reductions, such as coreferentiality and lexical identity, is expressed within the language by conjoined metalanguage sentences, rather than by a separate grammatical mechanism such as subscripts.<9> Similarly, ‘shared knowledge’ contextual and pragmatic information is expressed by conjoined sentences (including ordinary dictionary definitions that are zeroable because of their redundancy.

6. The set of possible arguments for a given operator (or vice-versa) is graded as to acceptability. These gradings correspond with differences of meaning in the base sublanguage, and thence in the whole language. They diverge in detail from one sublanguage or subject-matter domain to another. Equivalently, the fuzzy set<10> of ‘normal’ co-occurrents for a given word differs from one such domain to another within the base sublanguage.

In informal, intuitive terms, a constructive grammar generates sentences from the bottom up, beginning with word entry, whereas a generative grammar generates sentence structures from the top down, beginning with the abstract symbol S. The grammatical apparatus of constructive grammar (the rules together with their requirements and exceptions) is very simple and parsimonious. H’s underlying structures, the rules for producing derived structures, and the structures to be assigned to surface sentences are all well defined. Consequently, H’s argumentation about alternative ways of dealing with problematic examples has a welcome concreteness and specificity about it. In particular,<11> one may directly assess the semantic well-formedness of base constructions and of each intermediate stage of derivation, as well as the sentences ultimately derived from them, because they are all sentences. By contrast, in generative argumentation, definitions of base structures and derived structures are always subject to controversy because the chief principle for controlling them is linguists’ judgments of semantic relations (such as paraphrase) among sentences derived from them. Even if one could claim to assess the semantic well-formedness of abstract underlying structures, these are typically so ill-defined as to compel us to rely almost totally on surface forms to choose among alternative adjustments to the base or to the system of rules for derivation. And as we all know, a seemingly minor tweak in the base or derivation rules can and usually does have major and largely unforeseen consequences for the surface forms generated by the grammar.

H has always given primacy to semantics over syntax. Even in his earlier structural linguistics, H’s distributionalism was a study of semantics and not the empty taxonomy of the ‘structuralist’ stereotype; H’s two empirical touchstones, contrast and differential acceptability, are both semantic notions; and H’s aim in reducing redundancy in grammar (see below) is to get syntax out of the way of semantics. For Chomsky, on the other hand, syntax has always been central and semantics must be effected by a separate interpretive mechanism. Munz (1972) sketches the history of this divergence of the two paradigms.

The mention of paraphrase under characteristic 4 may trouble some readers. The use of judgments of paraphrase as a criterion for grammatical relationship has given rise to endless confusion and dissension in linguistics. This is because paraphrase in any strong sense is an exceedingly rare phenomenon in natural language. For H, paraphrase (‘weak paraphrase’) is rather an interpretation of the fact that the semantic dependencies of the base are preserved by (are recoverable under) the reductions. The reductions in turn are defined with respect to the more primitive notion of graded acceptability (characteristic 6).

On first exposure to H’s system, readers are frequently puzzled as to the motivation of characteristic 6, and are most likely to object to stylistically peculiar sentences postulated as sources (and as intermediate constructions in certain derivations) under characteristics 3 and 5. Characteristic 6 was of course previously H’s criterion for transformation<12> and gives him a metric for ‘weak paraphrase’: preservation of the objective information being transmitted, with expressive and communicative nuance<13> attributed to the reductions. This metric works as follows: Supposing two sentence forms<14> A and B to be related by some derivational step in H’s grammar, if any two sentences among the satisfiers of sentence form A differ on some scale of acceptability, the corresponding sentences among the satisfiers of sentence form B always differ in the same way.<15>

The relation to information content follows from this observation: given a base sentence with a valid variable (N or one of the subclasses of 0) in place of one of its words, the base sentences resulting from various valid substitutions of words for the variable are not all equally sayable, and this is a direct reflection of the meanings of the words. By this means, then, Harris’s grammar captures what the dust jacket refers to as ‘weak semantics’<16> without the cost of a separate semantic component. (For meatier aspects of meaning, we turn to discourse analysis, but that is beyond the scope of this review.)

H’s grammar is the product of long evolution, rather than of sudden revolution. H’s method of improving his grammar is to extend existing, established reductions to new argument domains, rather than to throw out the previous version of the grammatical apparatus for a revolutionary new one. (See Sager (1981) for examples of this within an earlier form of H’s paradigm.) H’s method mirrors the way language itself changes over time, and indeed incorporates the mechanism of language change by analogic extension. H’s formulation of analogic extension in his reduction system accounts in a unified way for synchronically productive analogic extension, such as metaphor, as well as for diachronic change. Then, because constructions that are at first difficult to account for turn out typically to be at the diachronic or analogic ‘growing edge’ of the language, reachable by refinement and extension of the domain of well-attested operations and reductions, the same mechanism provides an elegant solution to the sticky problem of the ill-defined and shifting boundaries of language. As an added dividend, his derivations of many constructions fit their history.<17>

Analogic extension is possible, indeed quite natural, when rules are defined in string-grammatical terms, and when the base generates a subset of the language. H’s grammar works from the inside out, as it were, generating a subset and extending to include the rest of the language. The generative paradigm works rather from the outside in: an excessively powerful system of rules generates a superset of objects containing the set of sentences, and a major task is pruning and filtering; degrees of grammaticality and degrees of acceptability (the two are often not distinguished) are problematic rather than being fundamental data of linguistics; and dynamic processes of metaphor, analogy, and language change are difficult to integrate with one another and with synchronic descriptions of languages.

Put another way, H’s reports and books since 1955 have always presented a complete grammar with a program for refinement and extension. H’s work cannot be understood and properly evaluated using critical tools that are appropriate for generative grammatical writings, which with few exceptions present a fragment of grammar with a program for either generalizing it or integrating it with other like proposals.

A crucial difference between constructive grammar and generative grammar is the importance H has always assigned to freeing his system as much as possible from redundancy.

This has not been a major concern in the generative paradigm;<18> indeed, with a phrase-structure base, there is no principled way to control the proliferation of nonterminal symbols or to control the structure of the apparatus of the grammar in general, as for example in derived constituent structure. It is therefore almost inevitable in generative grammatical work that artifactual, extrinsic structure intrudes upon and masks the intrinsic structure of the language that the grammatical apparatus seeks to explain. It follows that some putative language universals may be artifacts of generative grammatical apparatus, found in various languages just because the same apparatus is used, much as the padres of old wrote grammars of Latin with exotic vocabularies.

Although universals are not H’s explicit concern — presumably, he would determine them empirically after working out the transformational structure of other languages in constructive terms rather than making them a prerequisite for such work — one may draw some conclusions about them from this book. Surely, the very simple information-transmitting dependency structure of H’s base is universal, and having been found for English (and in some measure for French, German, Korean, Takelma, and other languages) may now be presumed as an objective of work in other languages. And surely the major types of reduction involving permutation and zeroing, such as his length permutation, include many universals, though details of the 'morphophonemic' types of reduction are more likely to be language specific.<19>

Comparison with generativist findings of semantic universals is more problematic. Differences of acceptability or likelihood among the arguments of a given operator, or among the operators (or operators and co-arguments) for a given word, are notoriously variable and vague. H has found that selection restrictions within the major dependency classes of base nouns and operators form at best fuzzy subsets that are inherently beyond the reach of the binary semantic features of interpretive semantics. To be sure, there are some hard yes/no selection restrictions that can be successfully encoded by binary semantic features, but these are produced by the reduction system (cf. Harris 1976:251 for a succinct statement), and therefore for H are not a problem of lexicon at all.

H does little with the classifier words<20> that are explicitly part of the vocabulary — the hierarchies of words that make up the taxonomies, both folk and scientific, that were the original motivation for componential analysis and semantic features. Classifier words among the operators (act is an example) go unmentioned except for the ‘appropriate’ operator needed for the sources of certain derived nouns (5 1 .25, p. 224). H does show how novel utterances may be related to familiar ones of known acceptability by substituting classifier words, as in the following example, from page 6:

Substituting classifier words:

It is not clear to the reviewer at least whether H can find further well-established reductions to derive the first sentence above from the last one (that appears not to be his aim in citing this example). Note, however, that a similar mechanism underlies Sager’s computer formatting of sublanguage discourse (Sager 1981, 1982; Hirschman and Sager 1981). Some classifier words presumably embody semantic universals (e.g. animal, human, object) so that such use of classifiers could be a device for machine translation. One may hope that H or another worker in the constructive paradigm will take up these undeveloped themes in some future report on the lexicon for a constructive grammar.

Presumably, the next important phase of H’s work includes a specification of the lexicon for his grammar and a return to discourse analysis from the point of view of operator dependencies. Characteristic 6 above suggests that the form of the lexicon of a constructive grammar differs in important ways from that of other approaches. Obviously, much that has exercised students of generative lexicography is captured by Harris’s reduction system. Even more obviously, there is no need to map the underlying base sentences onto some semantic representation, since the base sentences are themselves the semantic representation. Finally, much of the formalism of the generative lexicon, motivated as it is by characteristics of phrase-structure grammar, has no place here.<21>

Lacking H’s specification, I suppose that the lexicon of a constructive grammar would list a subset of words and affixes (the base vocabulary of N and operators, and the argument indicators and traces of reductions), together with their form class, if any. (It might also list some words that are products of particular reductions, together with their derivations, depending upon how much one wanted to include for pragmatic reasons relating to computability.)

For the acceptabilities, one could start with an ‘acceptability model’, a finite set of base sentences — actual sequences of words — that are known to be acceptable for a given subject-matter domain. These sentences would be of three sorts. The primary ones for establishing acceptabilities would use only classifier nouns, as in the onion-skin example. There would be a set of ‘dictionary sentences’ for each classifier N giving acceptabilities for its co-argument N under be. (For compactness, many of these sentences could have classifiers in both argument positions. These acceptabilities are well-defined, or nearly so, only for the classifier words in sublanguages of science.) The grammar would draw upon these dictionary sentences to extrapolate the fuzzy set of acceptable arguments for a given operator, or of operators (and, where required, co-argument N) for a given N (see Harris 1976, p. 247 and fn 13). Metaphor and analogic extension fall naturally out of the same process.

Finally, another type of dictionary sentence would be needed to meet the word-sharing requirement under conjunction that distinguishes coherent discourse from incoherent strings of sentences (pp. 13, 163-4; also Harris 1976, p. 241 and fn. 6; 1982, pp. 232-3). These ‘shared knowledge’ sentences are normally zeroed since they add only information that the hearer is presumed already to have. The three sets of acceptability-model sentences would be unique for each definable sublanguage, presumably with greater overlap between related subject-matter domains and with some shared ‘common usage’, and could function as a knowledge base for particular computer applications.

Much highly successful work has been done with computer grammars using string-based parsers of an earlier Harrisian type.<22> The present work suggests that a different sort of parser might use the predetermined constructive dependency class of each word, the operator indicator for each operator, and the traces of the reductions, as ‘handles’ to facilitate the parsing process. Such an approach was not feasible before the precise form and interrelationship of the base and the reduction system were determined. It should make possible computer grammars that are both more efficient and more semantically sensitive than any now implemented.

As I observed above, this is a complete grammar, and one whose internal integrity and parsimony is such that we examine individual fragments of the grammar out of context at our peril. With that caveat, here are some examples drawn from the book in the interest of giving more substance to the general remarks above.

Among the familiar semantic problems<23> that H covers is the middle voice (pp. 368-69) which we may exemplify by hit vs. break, where only the latter may become seemingly intransitive (The window broke, *The window hit). H derives the intransitive form of verbs like break from a zeroable higher operator that strongly selects a subject identical to the object of the lower verb:<24>

We may exemplify a second classic problem of semantic analysis by the relation between find and seek, where Fred found Pegasus presupposes the existence of Pegasus, and Fred sought Pegasus does not. Seek is in the dependency class Ono, and its apparent status here as Onn, is due to zeroing of the lower operator find.<25> The source of the example sentence is thus Fred sought to find Pegasus, where seek is asserted (under the I say operating on the whole sentence), but find is not asserted, and hence its presupposition about the existence of Pegasus is not asserted.<26>

A third familiar semantic problem is the scope of quantifiers. For H, quantifiers are not semantic primitives, but are reductions from complex sources:

The passive constructions Someone was opposed by everyone and Everyone was opposed by someone illustrate how H handles questions of quantifier scope. To understand them, we will first have to look at H’s treatment of the passive.

The first transformation isolated in the 1940s (or earlier), the passive has generally been assumed to be a single (paraphrastic) transformation. Many anomalous examples of passives have raised questions about the semantic properties of transformations in general. However, each of the three physical components of passive sentences occurs elsewhere in the grammar.

The first of these components is the apparent permutation of object and subject. H argues that the object, rather than being permuted, is the subject of a higher verb of the semantic set that make their subject a recipient (including have, need, undergo). As elsewhere in the grammar (such as the hit vs. break example, above), when this higher verb operates on a sentence containing an object N that is metalinguistically identified with its subject, the redundant lower object may be zeroed:

The second component of the passive is be -en before the original subject. The -en occurs with stative meaning in the have … -en (perfect) construction, as well as in the diachronically moribund construction seen for example in He is risen. H takes these to be occurrences of the same suffix, a suppletive form of the underlying operator be in a state: He is in the state of his scolding by me.

The third component is by before the original subject. (Sometimes another preposition is used, as in is interested in, is tired of, is tired from.) The by (or other preposition) appears in nominalizations such as The chopping of the trees by the settlers.

H shows that the domain and selection of the passive is precisely the product of the domains and selections of these components. As a bonus, he thereby accounts for those semantic idiosyncrasies that have made the passive problematic. The full discussion defies brief paraphrase, but the following examples<27> illustrate the point:

"Even though the source is a dubious kind of sentence", observes H,

A final example is the definite article. H analyses this not as a modifier or adjunct of the following noun, but vice-versa: the is itself a noun, a variant of the ‘indefinite’ noun that,<28> to which the following noun with its modifiers is in apposition: That which is a small book fell —> The small book fell.<29> This is the form of the source for apposition in general, so that many of its characteristics are not peculiar to the definite article. The special restrictions and semantic peculiarities of various uses of the fall out of the contexts of the underlying that which is. (Section 5.36 gives details.) For example in When H. W. retired the other day, he had sweet revenge on the alarm clock which had awakened him at 6 AM daily for 47 years, " … where the revenge was on anything that had been an alarm clock and had awakened him", (page 243), a condition that is met by the underlying that which is an alarm clock. Similarly, with the generic the in the family doctor is fast disappearing, it is that (of that which is) and not doctor that is the subject of disappear (loc. cit.).

From these examples, one may perhaps get a sense of how unfamiliar H’s derivations are — similar in form and intent to those of Wierzbicka (e.g. her 1982) though much more parsimonious — and perhaps even of how tightly interwoven the argumentation for each derivation is with that for other parts of the grammar. The challenges of full coverage are not addressed by the episodic, anecdotal examples that are typical of other approaches.

Some final comments on the book as a product are in order. It is fitting tribute to the publisher’s craft, published as it was in John Wiley & Sons’ celebratory 175th year. It is well produced, with the attractive layout and good-quality printing, paper, and binding that one expects from Wiley. With due consideration for the costs of publication today, this helps to justify the steep price. The book is marred by some typographical errors, but they are generally not confusing. Cross-references and citations are inconveniently buried in the footnotes that follow each chapter, so that one must resort to the index to find them. Because of the complexity of the subject, and the repetition of particular topics in various locations, the index is an especially important help to the reader. It is comprehensive and well arranged, although one might quibble about certain lapses.<30> The book is well organized and laced with cross-references, showing the pains that author and publisher have taken to make this very important work accessible to readers willing to put similar care into reading it.

Altogether, we have here an impressive analysis of the grammar of English as a whole, embodying a model of language and a paradigm for linguistics that is clear, explicit, and verifiable. Linguists of all persuasions would do well to study it carefully and on its own terms.

Bruce E. Nevin

Bolt Beranek and Newman Inc.

References

Chomsky, Noam. 1964. Current Issues in Linguistic Theory. Mouton, The Hague.

Corcoran, John. 1972. Harris on the Structures of Language. In P1ötz (1972).

Gross, Maurice. 1973. About the French verb TO KNOW. Studia Linguistica 122-124.

Gross, Maurice. 1975. Methodes en Syntaxe. Hermann, Paris.

Gross, Maurice. 1979. On the Failure of Generative Grammar. Lg. 55(4): 859-885.

Harris, Zellig S. 1965. Transformational Theory. Lg. 41: 363-401.

Harris, Zellig S. 1968. Mathematical Structures of Language. John Wiley & Sons, New York.

Harris, Zellig S. 1969. The Two Systems of Grammar: Report and Paraphrase. TDAP, Philadelphia. Reprinted in Papers in Structural and Transformational Linguistics (1970). Reidel, Dordrecht.

Harris, Zellig S. 1976. A Theory of Language Structure. Am. Phil. Quart. 13(4): 237-55.

Harris, Zellig S. 1982. Discourse and Sublanguage. In Kittredge and Lehrburger (1982), pp. 231-6.

Hirschman, L. and Sager, N. 1981. Automatic Information-Formatting of a Medical Sublanguage. In Kittredge and Lehrberger 1982 pp. 27-80.

Hiz, Henry. 1979. On some general principles of semantics of a natural language. Syntax and Semantics 10: 343-352.

Jespersen, Otto. 1909-1931. A Modern English Grammar on Historical Principles. George Allen and Unwin, Ltd., London.

Kendall, Daythal Lee. 1977. A Syntactic Analysis of Takelma Texts. Ph.D. dissertation, Department of Linguistics, University of Pennsylvania, Philadelphia, Pennsylvania.

Kittredge, Richard and Lehrburger, J., Eds. 1982. Sublanguage: Studies of Language in Restricted Semantic Domain. de Gruyter, New York.

Langerhans, Ilse. 1971. Unrestricted Predicates as Transformational Sources in German. TDAP 86, University of Pennsylvania, Philadelphia.

Munz, Jim. 1972. Reflections on the Development of Transformational Theories. In P1ötz 1972, pp. 251-74.

Newmeyer, Frederick J. 1980. Linguistic Theory in America. Academic Press, New York.

Plötz, Senta, Ed. 1972. Transformationelle Analyse. Athenaum, Frankfurt.

Postal, Paul M. 1978. Traces and the Description of English Complementizer Contraction. LI 9(l): 1-30

Sager, Naomi. 1981. Natural Language Information Processing. Addison-Wesley, Reading, Massachusetts.

Sager, Naomi; Claris, P.; and Clifford, J. 1970. French String Grammar, String Program Reports, No. 8. Linguistic String Project, New York University, New York.

Salkoff, Morris 1973 Une Grammaire en Chaine du Francais. Dunod, Paris.

Wheeler, Eric S. 1984. Review of Zellig S. Harris, A Grammar of English on Mathematical Principles. Computers in the Humanities 17(3): 88-92.

Wierzbicka, Anna. 1982. Why Can You Have a Drink When You Can’t Have an Eat?. Lg. 58(4): 753-99.

Zadeh, L.A. 1965. Fuzzy Sets. Information and Control 8: 338-353.

Notes

1. Jespersen was reputedly H’s principle guide for this second portion of the book. In his preface, H cites Jespersen’s monumental Modern English Grammar on Historical Principles (1909-1931), together with the OED, as "the two indispensable aids on the English language". The parallelism of titles is surely no accident. I believe that H also culled the literature to ensure that he was accounting for all the examples that other linguists had found for one reason or another to be problematic.

2. Another pitfall for the unwary reader is, ironically, a consequence of H’s longstanding rhetorical practice of repeating key information in succinct form so that each chapter or section is relatively autonomous. He does this in an effort to make his writings easier to read and use. However, a superficial reader might first encounter a controversial topic (for example, H’s proposed bisentential metalinguistic operator co-state in the source of and, or) in such a recapitulation, and reject it out of hand, having missed the main section where it is properly developed and justified. Unfortunately, the editors have failed to indicate the most important entry where the index has several, so that one must cross-check the page numbers of the index against those of the table of contents to find the main entry.

3. ‘Harrisian’ imputes to Harris some obscure responsibility for work that others do in this paradigm. His use of ‘constructive grammar’ (e.g. Harris 1968, pp. 17f, 20, 32, 89, 121) is apparently allied with intuitionist usage in mathematics. The term ‘constructive’ provides a counterpoise to the (equally irrelevant) rhetorical overtones that have grown up around the term ‘generative’ as a kind of trademark. A constructive grammar such as the one under review is of course generative in the technical sense of producing an explicit structural description for every sentence of the language. Furthermore, at least since his 1965 report in Language H has been explicitly concerned with generative rules for deriving sentences from base structures. (Newmeyer 1980:37 misinterprets Corcoran 1972:279 on this point: he ignores Corcoran’s further discussion in that paper, as well as H’s own later writings, in particular section 8.8 of Harris 1969, Corcoran’s principal reference). This notwithstanding, few would claim that H is a generative grammarian, and a correlative term other than ‘Harrisian grammar’ seems in order.

4. The suggestion outlined in Harris (1969), and exemplified in Gross (1973), that some of the base vocabulary may be suppletives factored from partial homonyms and synonyms, is not taken up systematically in this book. Echoes of this program may be found here and there, however, as for example in H’s observations on the derivation of -hood, -dom, which were once free words in English (suppletive forms of something like state and estate, respectively). Note here also the discussion (p. 73) of classifiers in the sublanguage of a science, in particular the sublanguage of grammar (which is the metalanguage of the language as whole).

5. A sublanguage is a subset of sentences in the language, mathematically defined by closure under a subset of the operations that are defined for the language as a whole (cf. Harris 1968, e.g. p. 152). The notion of sublanguage, with its semantic interpretation of subject-matter specialization, runs somewhat counter to rationalist notions of innate ideas and may be difficult to characterize in generativist terms. Nevertheless, sublanguages are as important as social and regional dialects for an understanding of language change, and are essential to an understanding of semantics of natural language as opposed to formal languages. See Kittredge and Lehrburger (1982) for a brief survey.

6. Weak paraphrases: see below.

7. Zeroing is similar to deletion with a condition of recoverability (but contrast the Constructive treatment at pp. 50 ff. and fn. 10 with Chomsky (1964), which introduced the latter notion). The term ‘deletion’ is often applied to ‘replacement’, as in pronominalization. This and other points of difference between the two terms seem to be an artifact of generative derivations being defined on abstract phrase-structure trees and Constructive derivations being defined on lexical dependencies. For an example of problems with deletion as defined on nodes of phrase-structure trees, see for example Postal (I 978), pp. I 0- II, and ref. cit. For Postal (p. 23), there is no way to formulate in a transformational framework the fact that ‘to contraction’ — phonetic reduction of to under the seven verb forms want, going, have, ought, used, got, and supposed is determined by ‘subject sharing’, though that is precisely how H’s transformational grammar did and his constructive grammar does account for these phenomena.

8. These traces, and those noted under 3 above, were the basis for reformulation of transformations as elementary sentence-differences, a crucial step in development of the constructive model. They are of course unrelated to the trace marker proposed in Chomsky (1973), which, as Gross (1979: 874) points out, is simply a computational device, a pointer to a memory address, such as is employed in programming languages.

9. See Harris (1968: 17 ff.) for reasons why the metalanguage must be within the language. "The one science that is part of the structure of the language itself is the grammar of that language" (p. 73). This property of language of course does not in itself preclude use of other types of notation for the metalanguage. Convenient notational reductions of other specialized sublanguages to formal symbol systems come to mind, as for example the notational systems of logic and mathematics, whose most complex formulae are nevertheless always stateable somehow in sentences of the corresponding sublanguages. The motivation for such a notational system (computational convenience) is offset here by H’s demonstration that metalinguistic information is normally zeroed under reductions defined for the language as a whole. There is also the open question whether the presence of low-information and metalanguage conjuncts in the base sublanguage constrains the reduction system, and whether subscripts and other computationally convenient notations might therefore give rise to some thorny problems as an artifact (see note 21 below). Furthermore, since sentences containing overt (unzeroed) assertions of metalinguistic information do occur in normal discourse, and cannot be excluded as ungrammatical, any grammar must account for them whether it uses them in this way or not, so it might as well use them.

10. See Zadeh 1965.

11. See also the discussion of analogic extension, below.

12. It was an alternative to well-defined selection or ‘co-occurrence sets’ as a criterion for two sentence-sets to be transforms in 1957, and it became the only criterion after H determined that simple well-defined co-occurrence did not stand up, as he reported in his 1965 paper.

13. These more subtle aspects of meaning, which are more difficult to characterize formally, are thus segregated for study together with other essentially gestural systems of communication, as distinct from transmission of objective information. In this way, H’s metric provides a principle by which to control the more baroque excesses of abstract syntax without giving up the power of the base to generate underlying semantic structures directly.

14. A sentence form is a sequence of variables and constants, where the variables are word-classes and the constants are particular words or affixes, such as those of the sequence be … -en by of passive sentences. Any n-tuple of words that may more or less acceptably be substituted for the n-tuple of variables constitutes a satisfier of the sentence form.

15. A seeming exception is where a sentence form itself — that is, all of its satisfiers — has reduced acceptability. (These sentences H marks † with a dagger.) However, even where some acceptability differences are collapsed and become difficult for linguists and their informants (or experimental subjects) to access, the corresponding sentences are never reversed in their positions on the given scale of acceptability, but at most are only reduced to identity as ‘equally marginal’.

16. There is no mention of this ‘weak semantics’ in the book proper, though it does occur in Hiz (1979). This ‘weak semantics’ must be construed as a form of ‘propositional meaning' with the proviso that a great deal of pragmatics, implied meaning, and the like, may appear as conjoined 'common knowledge' sentences that are zeroable because so utterly redundant (characteristic 5). This makes sense only in context of a knowledge base, a theme to which I will revert at the end of this review. Examples of H’s treatment of some familiar semantic problems follow this comparison of the two paradigms.

17. See e.g. pp. 27-28 and 377-78 for discussion. However, H’s derivation of please (p. 351 and n 7) could be improved with more historical perspective. He suggests:

Better would be:

I am not certain what support there is in H’s system for the permutation at the end of this derivational sequence; the last, commaless sentence does preserve the lowered intonation that please has in the preceding sentence. Contrast the derivation of the corresponding sentence with a comma:

Nuances of deference, irony, and the like associated sometimes with please and more often with if you please arise historically from normally construed context of a superior granting a boon and thereby pleasing an inferior.

18. H’s reduction of redundancy has in its intent some relation to the notion of simplicity in generative grammar, but is both more sensitive and intuitively more satisfying than gross measures of formalisms, such as symbol-counting.

19. For Korean, see e.g. Harris (1968:109-113). For French, see Sager (1981: 10.2); Sager, Claris, and Clifford (1970); and Salkoff (1973). For German, see Langerhans(1981). For Takelma, see Kendall (1977).

20. Except insofar as H’s indefinites may be thought of as classifiers with the broadest selection. There is some parallelism between the notion of a 'designated representative' of a category introduced in Chomsky (1964) and H’s use of indefinites where a zeroed argument is not specified by context.

21. For computer work, it might seem convenient to introduce some special notational conventions, such as subscripts for metalinguistic identity statements preserving a straightforward formal relationship to the words used to express them within the language, as noted above under characteristic 5. Such a program is questionable, however, insofar as the presence of such zeroable sequences in sources may interact with the normal restrictions defined for the reductions and thus impose some significant constraints on derivations. (There is some evidence for this in certain island phenomena.) In a system with metalanguage statements of coreference and the like represented in a separate notation outside the language, these constraints have to be restated in the form of special rules or exceptions to rules, introducing extrinsic redundancy precisely of the sort that we wish to avoid.

22. See for example Sager (1981), Gross (1975, 1979), Salkoff (1973), and references cited in those works.

23. The first three examples were suggested by Wheeler (1984).

24. With other verbs, such as stand in He stood the box on end, The tree stood in the forest, and †He stood the tree in the forest, H sets up a causative source (Zero Causative, 6.8). In both cases, the higher verb imposes restrictions on the derived form, such as we see reflected in the limited acceptability of the sentence just cited with a dagger † (which might occur acceptably, for example, in a myth or folk tale). These restrictions betray the zeroed presence of the higher verb and indicate which form is primitive and which derived.

Following H’s practice, I am using a dagger to mark a sentence of reduced acceptability that is still grammatical. It marks the relatively lower acceptability of marginal but still sayable sentences, rather than indicating their position on some absolute scale of acceptability. In some of the examples that follow where one sentence in a derivation is stylistically strange and another is even more difficult on semantic grounds, sayable only in some enabling special context, the dagger marks only the one with lowest acceptability to indicate the contrast. Note that the acceptability of such sentences is reduced precisely because the less redundant and more conventional reduction is available and obvious.

25. In some contexts, perhaps in the argument of capture or some other operator. Note that one may seek to do anything, but in most contexts only finding is so predictable (as the purpose of one’s seeking) that the verb find is zeroable under seek.

26. In Harris 1976, H argues that problems of intension vs. extension do not arise in his inscriptional approach to semantics, as they do in the more familiar interpretive and set-theoretic approaches.

27. Even here, I have omitted some details. The arguments for the ‘counting’ sources of plural quantifiers such as everyone are developed at length in section 5.5 (pages 244-263). H there refers us to section 7.15 for the derivation of each, any, every from modifiers on the metalinguistic I say that is the highest operator on a sentence, but I can find no account of each there or anywhere else in the book. This is the more astonishing because of the important role of for each in H’s metalanguage. For this and some other reasons, H’s description of quantifiers seems somewhat less thoroughly worked out and complete than other parts of the grammar. One may well guess that this area might be more problematic than most.

28. An ‘indefinite’ noun is one with exceptionally broad selection. This derivation accounts for the being almost always the leftmost prenominal, and explains precisely what may precede it and why (pp. 237, 263).

29. This in turn is derived from a base sentence containing the requisite metalinguistic identification of the noun with some explicit or implicit nearby repetition: A thing mentioned (nearby) — said thing is a small book — fell. More explicitly: A thing [prior is same as mentioned (nearby)] — a thing [prior is same as penult] is a small book — fell. The words in square brackets are interrupting metalinguistic asides about the words of the sentence themselves, for convenience abbreviated by said in the first version. In this small and precisely defined metalanguage, prior refers to the word before the current word (that is, the word before the word prior itself), and penult refers to the word before the ‘prior’ word. The two occurrences of thing are thus made each to refer to the other, yielding the ‘peculiar self-referring effect of the restrictive that which’ (p. 95). All anaphoric and epiphoric reference is handled by the same sort of zeroable metalinguistic sameness statement. So-called crossing coreference, as in The man who shows he deserves it will get the prize he desires, is not problematic for H’s reduction rules, since they are not defined in terms of constituent structure.

30. For example, the range 40-54 for operator indicator (rather than the individual pages 40 and 54), be as operator on p. 69, everyone on p. 366, nonrestrictive relative clause on p. 162, zeroing of which is on pp. 201, 238, zero causative on pp. 317 f, and numerous instances where a well placed f or ff would help to indicate the major entry.