1 Introduction
Syntax and semantics play different roles in the relationship between a predicate and its arguments. From a semantic perspective, predicates require different arguments to assume particular semantic functions or roles. This is referred to as semantic selection (s-selection) (Pesetsky Reference Pesetsky1996). Syntactic selection (c-selection) determines the syntactic categories that realize the predicate’s arguments. A related notion, subcategorization, pertains to the idea that lexemes, which are categorized according to syntactic categories, can be further subcategorized according to the type and number of syntactic complements that they require. Finally, linking describes the matching between arguments and syntactic functions. Thus, for example, active verbs link their Agent argument with the Subject role, while with passive verbs, the Patient argument assumes this role.
The question of whether s-selection, c-selection, and linking are basic or whether one is reducible to another has been worrying linguists since at least Grimshaw (Reference Grimshaw1979), who argues that semantic selection and syntactic selection are two independent sets of restrictions. She focuses on the selection of interrogative and exclamatory complements, and shows that syntactic subcategorization alone cannot account for the data.
As an illustration of her argument consider the following sentences:
Both ask and wonder can appear with an embedded question as a complement, yet only ask is compatible with an NP complement. In this context, this NP is referred to as a ‘concealed question’ since it can be intuitively paraphrased as an embedded question.Footnote [2] Nevertheless, this set of sentences raises the question of what drives the distinction between the subcategorization options of the two verbs. Grimshaw (Reference Grimshaw1979) claims that these differences cannot be derived from semantic differences, and thus suggests that syntactic and semantic selection are independent.
Predicates, in Grimshaw’s system, impose two types of restrictions on their complements. Semantic selection involves semantic categories such as Q (question), P (proposition), or E (exclamation). Syntactic subcategorization targets syntactic phrase types such as CP, PP, NP, AdjP. The two sets of restrictions pertaining to ask and wonder are demonstrated in Table 1.
Although Grimshaw’s claim is that the two restrictions are independent, naturally the association between them is not completely arbitrary. Grimshaw (Reference Grimshaw1979) suggests that each semantic category is associated with a Canonical Structural Realization (CSR), which is its ‘default’ syntactic realization. Thus, for example, the CSR of Q arguments is CP, and the CSR of ‘Thing’ is NP. Nevertheless, NPs can also realize Q arguments, hence the ‘concealed question’ NP in (1b).
The ultimate constraint, however, stems from the interplay between the two sets of restrictions. Thus, although Q arguments can be realized as CPs or NPs, their actual realization depends on the syntactic subcategorization of the predicate. With ask, the two potential realizations of Q are possible, while the subcategorization property of wonder restricts it to CP. This accounts for the data in (1).
Grimshaw focuses on interrogative and exclamative complements, yet her claim regarding the independence of syntactic and semantic selection is general. A fully fleshed-out theory of argument structure in the spirit of Grimshaw must define a complete set of syntactic categories, a complete set of semantic categories and the CSR of each semantic category, as well as a set of its alternate realizations (e.g., NP as Q). These building blocks should be sufficient to determine the semantic selection and syntactic subcategorization of all predicates in the language, as well as the linking between the semantic selection and the syntactic subcategorization.
The relationship between the syntactic and semantic aspects of argument selection features in a study by Jackendoff (Reference Jackendoff1985), which discusses the phenomenon of multiple subcategorization, whereby a predicate can appear in a number of different subcategorization frames. Jackendoff focuses on the verb climb, which can appear in three different configurations:
The sentences above are each prefixed with their respective subcategorization frame, represented in the notation originally proposed by Emonds (Reference Emonds1991). The lines in this notation indicate the position of the verb, and the NP and PP indicate the syntactic category of the complements.
Bearing in mind that the three configurations all pertain to one lexeme, a more compact and telling notation could be used, with parentheses indicating the optionality of the complement:
However, as Jackendoff notes, this representation will not rule out the following ungrammatical sentence, where the two complements are realized:Footnote [3]
A possible solution would be to use the set notation to indicate that only one complement is possible:
Grouping the NP and PP complement together, Jackendoff claims, misses a semantic distinction that differentiates between the two complement types. In both cases, the verb climb indicates motion. With the intransitive frame and the NP complement frame, the motion is understood as upwards (cf. (2a) and (2b)). With the PP complement, the direction of the motion is specified by the preposition heading the PP (2c). All of this information is part of the speaker’s knowledge of the predicate.
The question, then, is how to capture the correspondence between syntactic arguments and semantic relations. Gruber (Reference Gruber1965) was the first to refer to the semantic relations between predicates and their arguments as thematic roles. Subsequently, different sets of thematic roles (also known as theta roles or $\unicode[STIX]{x1D703}$ -roles) have been proposed in the literature.Footnote [4] The most common roles adopted in the literature are Agent, Patient, Theme, Location, Source, and Goal.
Nevertheless, in his discussion of cases such as the verb climb, Jackendoff (Reference Jackendoff1985, Reference Jackendoff1987) contends that thematic roles cannot accurately capture the semantic relations between predicates and arguments. For example, it is plausible to assume that the semantics of climb involves a Goal. Nevertheless, it is the top of the ladder in (2b) and not the ladder itself, that is the Goal. Thus, Jackendoff claims, attributing this thematic role to the complement of the verb is not adequate. More generally, Jackendoff concludes, thematic roles are not primitives of semantic theory, and thus a more fine-grained semantic formalism is required.
This kind of phenomenon and argumentation led Jackendoff (Reference Jackendoff1985, Reference Jackendoff1987) to propose an alternative theory, Conceptual Semantics. A basic assumption in this theory is that the semantic structure of a sentence is built up from conceptual primitive (Jackendoff Reference Jackendoff1987). One type of primitive is ‘semantic parts of speech’ such as Thing, Place, Path, Event, State, Manner, and Property. Correspondence rules map semantic parts of speech to syntactic ones. Each conceptual constituent has a syntactic phrase type which is its unmarked realization.Footnote [5] Thus, for example, the unmarked realization of Thing is NP, of Event is S, and of Property is AdjP. However, there are also marked realizations: Property can be realized by PPs (e.g., out of luck) or NPs (e.g., bummer).
Atomic semantic expressions can be expanded into more complex expressions by the application of semantic functions such as GO, BE, STAY, TO, FROM, and TOWARD. Thus, for example, a Path can be expressed by applying the TO function on a Thing such as the house. Following is an example of a conceptual structure and its corresponding syntactic structure.
An important principle of Conceptual Semantics is that there is no assumption of one-to-one correspondence between syntax and semantics. While this correspondence holds in the case of to the house in (6), this is not necessarily so. A case in point is the lexical entry that Jackendoff proposes for climb. The lexical entry expresses the fact that the verb denotes a motion event by specifying a semantic function, GO, and a Path specification. The Path specification is realized in two different ways, depending on the complement. The PP complement provides the total semantic content of Path in a compositional fashion, similarly to to the house. However, when the complement is an NP, its Path denotation is implicit and is expressed by the use of primitive semantic functions, as is illustrated below.
Syntactically there is only one argument, and it denotes a Thing. The fact that the Path is directed towards the top of the Thing is not expressed explicitly in the syntax. Rather, it is incorporated into the meaning of the verb climb.
The exact notation and specifics of the lexical entry are immaterial for our purpose.Footnote [6] Nevertheless, the point that we mean to stress is the observation that a formal syntax-only representation of multiple subcategorization is not sufficient to tell the full story. Rather, the interplay between syntax and semantics is a major factor in understanding argument structure. This observation guides the solution that we propose in Section 3.
Our solution distinguishes between syntactic selection and semantic selection. Consequently, constraints can be defined for each level individually, yet their interaction ultimately accounts for the argument structure of predicates. This, as we show, extends the expressive power of the mostly syntactically based Head-driven Phrase Structure Grammar (HPSG, Pollard & Sag Reference Pollard and Sag1994) approach to argument structure. As a result, generalizations that are currently overlooked can be made explicit in the grammar. Moreover, the analyses produced by the grammar are semantically informative and can serve as input to natural language processing tasks which require information related to the syntax and the semantics of linguistic expressions, such as machine translation (e.g. Bond et al. Reference Bond, Oepen, Nichols, Flickinger, Velldal and Haugereid2011), intelligent text understanding, etc.
We explore in Section 2 a number of different approaches for capturing the semantic relations that hold between predicates and their arguments. This serves as background for Section 3, where we present our proposal and illustrate its benefits by considering various phenomena that our proposal facilitates better, more general accounts for. In Section 4 we delve into more technical HPSG details, describing how our proposal is implemented in an LKB-based (Copestake Reference Copestake1999, Reference Copestake2002b) computational grammar of Modern Hebrew. We conclude with directions for future research.
2 Semantic selection
The previous section reviewed the theories of Grimshaw (Reference Grimshaw1979) and Jackendoff (Reference Jackendoff1985) regarding the relationship between syntax and semantics in the context of predicates and their argument structure. The two theories share a general view, namely that both semantic and syntactic categories are needed in order to account for the argument structure phenomena found in natural language. Furthermore, in both theories the relationships between the two domains are subject to correspondence rules. The main question, then, is what constitutes a possibly universal, finite, exhaustive, and well-defined set of semantic categories that can capture the semantic relationships between selecting predicates and their arguments. This is still an open question which has received many different answers over the last fifty years.
Dowty (Reference Dowty1991) distinguishes between two types of understandings of thematic roles. The approach adopted by Jackendoff and Grimshaw views thematic relations as notions of conceptual structure, existing independently of syntactic or interface notions. A different understanding is referred to by Dowty as the ‘argument-indexing’ view of thematic roles and is tightly connected to syntax.
A strong formulation of the argument-indexing view is found in the $\unicode[STIX]{x1D703}$ Criterion (Chomsky Reference Chomsky1981):
While there is no consensus regarding the first clause of the $\unicode[STIX]{x1D703}$ criterion, the principle expressed in the second clause figures in all the approaches we review.Footnote [7] Thus, regardless of the number or content of the semantic roles proposed in each approach, the principle of ‘one argument per role’ is maintained.
In what follows we review some of these approaches and discuss their applicability to the purpose of this study.
2.1 Fillmore’s deep cases
Fillmore’s seminal paper ‘The Case for Case’ set the stage for all subsequent work on semantic roles. Inspired by Gruber (Reference Gruber1965), Fillmore (Reference Fillmore1968) makes one of the first attempts to establish a list of semantic roles, or, as Fillmore referred to them, ‘deep cases’. The list, as envisioned by Fillmore, is ‘a set of universal, presumably innate concepts which identify certain types of judgments human beings are capable of making about the events that are going on around them, judgments about such matters as who did it, who it happened to, and what got changed’ (Fillmore Reference Fillmore1968: 24). Fillmore identifies an initial set of six deep cases:
However, he also notes that ‘Additional cases will surely be needed…’ (Fillmore Reference Fillmore1968: 46–47), and subsequently proceeds to add Benefactive and Time, in his discussion of prepositions, and Comitative in the context of coordination.
Fillmore (Reference Fillmore1971) revises the list of semantic roles: the Dative role is replaced with the new Experiencer case, Locative case is split into three cases (Location, Source, and Goal), Factitive is subsumed under Goal, and Comitative is dropped. The result is a nine-case system (Agentive, Instrumental, Experiencer, Object, Location, Source, Goal, Time, Benefactive).
This early attempt at providing a comprehensive analysis of the semantic relations between predicates and their arguments has since taken different directions by different researchers. Broadly speaking, some approaches have extended the roleset and enriched the semantic representation of argument structure, while others, mostly computationally oriented, have attempted to reduce the set. In the following sections we describe a number of alternative approaches, and discuss whether they are suitable for representing the syntax–semantics relations in a wide-coverage computational grammar.
2.2 Semantic roles in HPSG
A number of different approaches to semantic roles are found in the HPSG framework. We describe below three approaches which vary in the level of generalizations they express and in the richness of their semantic representation.
2.2.1 The original HPSG analysis
The semantic relations assumed by Pollard & Sag (Reference Pollard and Sag1994: 29) consist of the feature relation, whose value is atomic (e.g., love), and a number of predicate-specific role features (e.g., lover, loved). With relation-specific role labels, the total number of unique roles is proportional to the size of the lexicon. Moreover, an obvious shortcoming of this approach is that it does not provide any means of expressing generalizations regarding the arguments and the association between semantic roles and syntax.
Acknowledging this, the authors propose an alternative approach. They sketch a solution to the generalizability problem by suggesting (in their Section 8.5) to define a type hierarchy of relations, where shared role features are introduced for non-maximal relation types (i.e., more general types which dominate subtypes), and are ultimately shared by their subtypes. A sketch of this hierarchy, which is dominated by the general quantifier-free parameterized states of affairs (qfpsoa) type, is shown in Figure 1.
With such a hierarchy, instead of individually defining for each verb its semantic relation, verbs with a similar semantic structure inherit from a general semantic relation type, for whom the shared features are defined. For example, the semantic relations denoted by verbs expressing influence (e.g., persuade, urge) will all be subtypes of a more general influence relation, and as such their semantic relation will include three semantic roles: Influence, Influenced and State-of-Affairs (SOA). A similar conceptualization of general semantic relations that are inherited by specific lexical items is found in the FrameNet project (see Section 2.4).
2.2.2 Linking as constraints on word classes
The idea of defining a hierarchy of semantic relations is further developed by Davis & Koenig (Reference Davis and Koenig2000), who use general semantic categories such as Actor, Undergoer, and State-of-Affairs (SOA) to express more complex semantic relations. A sketch of this hierarchy is given in Figure 2 (Davis & Koenig’s Figure 5). In this hierarchy, each supertype introduces one semantic role. More complex relations inherit feature specifications through multiple inheritance from multiple supertypes.
Consider as an example the act-und-rel relation, which includes two semantic roles, act(or) and und(ergoer), which it inherits from the general types act-rel and und-rel, respectively. This relation in turn dominates instances of semantic relations, such as hit-rel. A more complex relation, cause-change-of-state-rel, inherits from the act-und-rel relation type and from soa-rel, which introduces the feature soa, whose value is a relation which denotes the state caused by the action of the Actor.Footnote [8] Instances of this semantic class are the relations denoted by cut and break. Examples of the general cause-change-of-state-rel relation type and the specific cause-break-rel are illustrated in Figure 3.
The main purpose of the semantic representation that Davis & Koenig (Reference Davis and Koenig2000) propose is to account for linking patterns, that is, the systematic mapping between semantic arguments and syntactic functions. The semantic relation type hierarchy that is illustrated in Figure 2 is mirrored by a parallel word-class hierarchy, where linking is defined. For example, the act-und-rel semantic relation type is associated with the act-und-vb verb type, where the Actor argument is linked to the subject and the Undergoer argument to the object. Since accounting for linking is the goal of the study, the set of semantic relations that Davis & Koenig assume are those that affect just that. Consequently, the set is limited to a few roles: Actor, Undergoer, SOA, Figure, Ground, Property-bearer. The authors demonstrate how linking patterns of various verb types and constructions that are considered a challenge for linking can be accounted for based on this set of roles.
A primary concern in this context is the lack of correspondence between syntactic arguments and semantic arguments. For Davis & Koenig (Reference Davis and Koenig2000) this is an important feature of their proposal, as they argue that semantic representations should not be reduced to ‘syntactic diacritics’, to borrow their term. Consequently, they do not assume one-to-one correspondence between semantic roles and syntactic arguments. For instance, although cause-change-of-state verbs have two syntactic arguments, the semantic relation that they denote includes an additional role, soa, which denotes the end-state and which is an implicit part of the content of the verb.Footnote [9] This is shown in Figure 3, where the end-state of the verb break is represented by the broken-rel relation, which does not have a syntactic counterpart. Conversely, the referent of the NP complement assumes two semantic roles: the Undergoer of the ‘main’ relation, and the Property-bearer of the embedded relation.
On the one hand, the system proposed by Davis & Koenig (Reference Davis and Koenig2000) captures fine-grained semantic distinctions. For example, it characterizes the difference between two transitive verbs such as hit and break, whose subcategorization frames are identical. The semantic difference between the two is that the former has a simple actor–undergoer denotatum, while the latter is assumed to have a more complex caused-change-of-state semantics. On the other hand, although the inventory of semantic relations may be sufficient to account for non-trivial linking phenomena, it is quite limited. For example, their semantic representation does not distinguish between basic semantic roles such as Source and Goal (cf. Fillmore Reference Fillmore1968). Consequently, it is not clear how scalable this schema is; the set of semantic roles may not be sufficient to account for large and diverse data, and, furthermore, the creation of a wide-coverage lexicon with such sophisticated representations is not a simple task.
2.2.3 The delph-in approach
HPSG grammars implemented in the framework of the delph-in initiative, such as the English Resource Grammar (Flickinger Reference Flickinger2000), use Minimal Recursion Semantics (MRS; Copestake et al. Reference Copestake, Flickinger, Pollard and Sag2005). MRS adopts a ‘flat semantics’ representation which allows for different levels of specification of scope (negation, quantifiers, adverbials, etc.). Underspecification is used to capture scopal ambiguities. The semantic roles in predicate–argument structure are represented in delph-in grammars by what is referred to by Copestake (Reference Copestake2009) as ‘slacker semantics’, in contrast to frameworks such as FrameNet (Section 2.4) which aim to provide elaborate, semantically rich representations. Slacker semantics, Copestake (Reference Copestake2009) claims, is appropriate from a grammar engineering perspective.
Four general argument role features are used in feature structures of predicates: arg1 through arg4. Copestake (Reference Copestake2009: fn. 5) notes that ‘arg4 occurs very rarely, at least in English (the verb bet being perhaps the clearest case)’. The approach is syntactically rather than semantically driven. Arguments are specified on the basis of the syntactic obliqueness hierarchy (Pollard & Sag Reference Pollard and Sag1994) and are assigned semantic roles consecutively. Thus, beginning with the least oblique argument (the subject in non-passivized predicates), arguments are assigned numbered roles (arg1 through arg4). There is no predicate that is specified for an arg n argument without selecting for arg n-1. Nevertheless, any argument may be syntactically optional, and thus not necessarily realized in the syntax. Moreover, the roles of predicates are relative to a predicate and remain constant across valence alternations. Consequently, active/passive and double object/dative pairs will have the same predicate–argument structure.
The main shortcoming of this system is that across predicates there is no semantic significance to the semantic labels with which arguments are associated. For example, being an arg2 argument simply means that the argument is the second least oblique argument in the argument structure of the verb, regardless of whether it denotes a Theme, Goal, or Attribute (9).
Moreover, the class of syntactic phrases that realize a semantic role, say arg2, most likely contains the entire inventory of syntactic phrases in the grammar. There is no principled restriction on the syntactic or semantic properties of the argument bearing this semantic role.
2.3 PropBank
A similarly limited set of semantic roles is used in the PropBank project, whose goal is to add a semantic layer to the syntactic trees of the Penn Treebank (Kingsbury & Palmer Reference Kingsbury and Palmer2003). The semantic representation is therefore closely tied to the syntactic structure. The annotation scheme employed by PropBank attempts to maintain a small set of five roles, while providing consistent argument labels across different syntactic realizations of a verb sense. The scheme uses numbered roles due to a lack of consensus regarding semantic role labels, as well as the desire to limit the number of labels. Table 2 presents PropBank’s set of role labels and corresponding semantic roles.Footnote [10]
The annotation of Arg0 and Arg1 across predicates is consistent. Following Dowty (Reference Dowty1991), the choice between Arg0 and Arg1 is made by comparing the number of proto-Agent and proto-Patient entailments that are valid for an argument. The greatest number of entailments determines whether the argument is labeled Arg0 (Proto-Agent) or Arg1 (Proto-Patient). The distinction between roles Arg2, Arg3, and Arg4 is verb-specific, yet is consistent across different syntactic realizations of the same verb sense. The lack of consistency across verbs is an inevitable consequence of the desire to define a small set of roles, while observing the principle of ‘one argument per role’.
The annotation scheme provides a way of distinguishing among different senses of a given verb. For example, the verb leave has a number of senses, of which two are ‘move away from’ and ‘give’. Each one of the senses is associated with its own roleset, where each role is explicated by the use of a sense-specific role, such as ‘place left’ or ‘thing given’.
In addition to the descriptions of the different rolesets, PropBank provides examples of the different syntactic argument frames in which the verb in a particular sense appears. Following are examples of two syntactic argument frames in which leave in its ‘give’ sense is found: the ditransitive frame (11a) and the double object frame (11b).
Consistency in annotation across predicate sense is maintained, as (for) Mary is labeled as Arg2, regardless of its syntactic realization and its position with respect to the other complement.
The PropBank schema is more semantically motivated than the ‘slacker semantics’ approach of delph-in grammars, in that each of its five roles is associated with a number of common semantic roles, and a sense-specific characterization of these roles is defined specifically for each lexical entry. There is, however, no attempt to maintain consistent labels across different verbs and even verb senses. For example, Arg2 represents very different semantic roles in the two senses of the verb leave, illustrated in (10) above. Moreover, there is considerable overlap between the semantic roles associated with each label, as can be seen in Table 2. Thus, for example, Arg2 and Arg3 are associated with ‘Benefactives’ and ‘Attributes’. This inconsistency is evident in the English framesets illustrated on the PropBank website for the predicates give,Footnote [11] bake,Footnote [12] and buy.Footnote [13]
Although the chefs, Mary, and his mother are all the recipient/benefactive arguments of their respective verbs, they are labeled differently for each verb. Thus, their labels are only meaningful in the context of the particular roleset in which they occur, where their actual meaning is listed. As was previously mentioned, Arg0 and Arg1 are exceptions to this, as they are used fairly consistently across the predicates.
2.4 FrameNet
Semantically rich representations are proposed by FrameNet and VerbNet (Section 2.5), two lexical databases which contain detailed syntactic–semantic descriptions of lexical items, as well as annotated examples of how the words are used. The goal of the two projects is much greater than to provide a set of semantic roles which capture the semantic relations between predicates and their arguments and which can be used to represent these relations in a computational grammar. Nevertheless, the identification and categorizations of these relations is an essential part of the two projects, and for this reason they are relevant to this study.
FrameNet is based on a theory of meaning called Frame Semantics (Fillmore Reference Fillmore1982), which evolved from Fillmore’s theory of deep cases (or Case Grammar), described in Section 2.1. The first step in this evolution was the realization that cases can be used to define situation types. Thus, for example, a ‘caused change’ situation is associated with an AIO (Agent–Instrument–Object) case frame (e.g., I fixed it with a screwdriver). Consequently, a large number of situation types were defined, which ultimately led to a new conceptualization: ‘making frames primary, and defining roles in terms of the frames’ (Fillmore Reference Fillmore2012: 711), or, in other words, Frame Semantics.
The FrameNet project (Fillmore, Johnson & Petruck Reference Fillmore, Johnson and Petruck2003) involves the construction of a database which includes a list of Frames. Fillmore (Reference Fillmore2012) reports that FrameNet has 1200 frames. Each frame is associated with Frame Elements (FEs), which are semantic labels for ‘things worth talking about when a given frame is relevant’ (p. 714). The list of FEs associated with a frame is divided into core FEs, which are required by the frame, and peripheral FEs, which are traditionally described as adjuncts. In addition, FrameNet includes a list of 13,000 lexical units – nouns, verbs, and adjectives – which are associated with frames. Each lexical unit appears with a set of annotated corpus-based sentences which illustrate the different ways in which the FEs can be realized.
As an illustration, consider the compliance frame, its FEs, lexical units, and annotated sentences (Fillmore Reference Fillmore2012).
Each of the boxed expressions in the definition is an FE. Among the lexical units associated with compliance are the positive abide, observe, and obey and the negative breach, flout, and violate. Following is an example of annotated sentences with the lexical unit violate:
Similar FE labels may be found in different frames. Thus, for example, there could be other frames with a Protagonist. Nevertheless, the theory assumes that FEs are local, or, in other words, only relative to the frame in which they appear. For this reason, it is meaningless to evaluate the number of roles used by FrameNet. Moreover, it is not possible to isolate FrameNet’s semantic roles from the entire framework. Thus, this approach is not suitable for our goal.
2.5 VerbNet
The structure of the VerbNet lexicon (Schuler Reference Schuler2005) is based on Levin’s (Reference Levin1993) verb classification. According to Levin, verbs can be classified according to the types of syntactic alternations in which they can appear. Consequently, entries in VerbNet are associated with a particular verb class, and pertain to a number of verb members that share syntactic and semantic properties.
One well-known verb class is the Spray/Load class, which includes verbs that participate in the locative alternation.
Some other verbs that belong to this class are splash, scatter, drizzle, and sprinkle.
VerbNet augments Levin’s verb classes by (1) listing the thematic roles (e.g., Agent, Theme, Destination) associated with the verb class, (2) assigning a thematic role to each syntactic argument in each argument frame, and (3) describing the meaning of the frame by using semantic primitives (e.g., Motion, Location, Cause). In addition, when relevant, thematic roles appear with selectional restrictions which address the existence or absence of properties such as CONCRETE, SUBSTANCE, ANIMATE, or ORGANIZATION.
Two frames are associated with the Spray class, each pertaining to one of the alternates illustrated in (15) above.
While both alternations share the same set of thematic roles, the syntactic realization of these roles is distinct (hence the alternation). Moreover, the semantics of the two alternations differ: whereas (15b) implies that the location/container is completely filled or covered, (15a) does not. This difference is expressed in the semantic representation in the respective frames.
Unlike FrameNet, VerbNet assumes a general list of semantic roles, which apply to all verb classes and verbs. Altogether, VerbNet’s roleset includes 36 roles.
The roles are defined in an inheritance hierarchy which is dominated by four general roles: Actor, Undergoer, Time, and Place. The Actor role dominates two more specific roles: Agent and Cause, distinguished by the negative and positive values of the feature INTENTIONAL. The Cause role further dominates the Stimulus role. More specific roles are often used for specific types of verb classes. Thus, for example, the Topic role, a sub-role of Theme, which in turn is dominated by Undergoer, is associated only with information transfer or communication verbs. Roles that are in a parent–child relationship cannot co-occur.
VerbNet’s sizable roleset provides a way to represent the semantic relations between predicates and their arguments in a relatively fine-grained manner. Moreover, this roleset can be adopted as an independent component, without necessarily adopting Levin’s classification of verbs. However, with fine-grained distinctions between semantic relations it is at times difficult to decide between a number of alternative relations (e.g., the precise difference between Undergoer and Patient). Indeed, Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014a) report low inter-annotator agreement regarding the use of VerbNet’s roles in Polish. Consequently, with large-scale computational grammars, using a smaller, more broadly defined set of roles may be more feasible.
The hierarchical structure in which the roles are defined in VerbNet addresses the need for more general categories; a number of lower-level roles can be collapsed into one role, the one that dominates them. This, however, has proven to be quite tricky, since the hierarchy includes cases of multi-inheritance where a given role is dominated by more than one role (a case in point is Result, which is dominated by both Goal and Patient[+affected], where the former is dominated by the general role Place, and the latter by Undergoer). Consequently, there is no straightforward way to extract a smaller, more general set of roles from the VerbNet roleset.
2.6 Syntactic approximation of semantic roles
Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014a) experimented with adopting VerbNet semantic roles for Polish. Seven annotators used VerbNet’s set of roles to annotate sentences containing 37 randomly selected verbs. Altogether there were 393 occurrences of the verbs, with a total of 843 arguments. An analysis of the annotations revealed low inter-annotator agreement ( $\unicode[STIX]{x1D705}$ score of 0.617).Footnote [14] The two main reasons the authors give for this low score are (1) numerous cases where more than one argument role seemed to fit and (2) cases where no suitable role was found. They conclude that VerbNet’s schema may not be suitable for their grammar engineering task, which is to add a semantic component to a Lexical Functional Grammar (LFG, Kaplan & Bresnan Reference Kaplan, Bresnan and Bresnan1982) of Polish.
As an alternative, Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b) propose a method for assigning semantic roles to arguments in languages that have rich morphosyntax. They exploit the rich morphological case system of Polish (with seven cases) in order to define a small set of 11 semantic roles, or labels (see Table 3). Each category is associated with a number of prepositions and/or morphological cases. The authors concede that these categories are only approximations of semantic roles, and consequently refer to them as ‘semantic roles’ (in scare quotes) and name them R0, R1, etc.
This semantic roleset proposed by Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b) is richer than the rolesets used by PropBank and in computational HPSG, and it allows for consistent labeling across predicates. While semantically coherent, it is strictly ‘argument indexing’, in that it does not introduce implicit semantic features. Moreover, each preposition or morphological case is matched with only one role, thus making the labeling process quite deterministic.Footnote [15] However, due to the nature of the task, this schema is closely tied to the Polish language with its rich morphosyntactic system and what Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b) view as a semantically coherent use of prepositions. As such, it may not be appropriate for representing semantic relations across languages. For example, R10 ‘Manner Argument’ is associated with a single Polish preposition, według (‘according to’), which does not seem to correspond to a core semantic role in Hebrew. Nevertheless, as was mentioned at the beginning of this section, the question of whether there is a possibly universal, finite, exhaustive, and well-defined set of semantic categories remains an open question.
2.7 Conclusion
This section presented a number of different approaches to analyzing and representing the semantic relations between predicates and their arguments. One significant difference between the aforementioned schemas for the semantic categorization of arguments lies in the tension between two desiderata:
-
∙ Coherent, semantically driven argument labels.
-
∙ A small, manageable set of argument labels.
While the motivation behind each of these desiderata is clear, one cannot ignore their conflicting nature. Consequently, we can identify a cline, ranging from the semantically rich approach adopted by FrameNet (Fillmore et al. Reference Fillmore, Johnson and Petruck2003), whereby semantic labels are frame-specific and frames are organized in a large and intricate hierarchy, to the minimalist approach employed in delph-in grammars (e.g., The English Resource Grammar; Flickinger Reference Flickinger2000), where a small set of four consecutively numbered labels (Arg1, Arg2, $\ldots$ ) is assigned to arguments. The remaining approaches occupy different positions along this cline.
Moreover, a crucial factor which distinguishes between the various approaches is the extent to which the semantic representation is dependent on the syntactic structure. As mentioned, Dowty (Reference Dowty1991) distinguishes between conceptual approaches and ‘argument indexing’ approaches. However, even within the latter approaches, there are varying degrees of overlap between syntax and semantics in the domain of argument structure. The schemas adopted by PropBank and Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b) are designed to add a semantic component to syntactic representations (Penn Treebank and LFG f-structures, respectively). As such, these schemas are more tightly connected to syntax, yet they are not involved in the parsing process. Conversely, Davis & Koenig (Reference Davis and Koenig2000) argue that a true semantic representation cannot be reduced to ‘syntactic diacritics’, and consequently in their system there is no one-to-one correspondence between semantic roles and syntactic arguments.
The representation of argument structure that we propose is designed to capture the syntactic and semantic aspects of argument structure while striking the right balance between the two levels. As such, our approach is closest in spirit to that proposed by Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b). It is syntactically driven, in that the semantic representation is directly related to the syntactic structure. Nevertheless, the inventory of semantic roles is sufficiently extensive to be meaningful, and is arguably universal. Moreover, as we show in Section 3.2, the approach that we propose extends the expressive power of the grammar by providing a way to state generalizations that are not captured by ‘standard’ HPSG (e.g., Sag, Wasow & Bender Reference Sag, Wasow and Bender2003).
3 Representing the syntax–semantics interface in argument structure
The relationship between predicates and their arguments involves both syntax and semantics. Following Grimshaw (Reference Grimshaw1979), we view syntactic selection and semantic selection as two separate yet correlated phenomena. Representing this relationship involves accounting for each of the levels separately, as well as for the correspondences between the two levels. The rest of this paper is dedicated to doing just that.
Generally speaking, there is broad consensus regarding the types of syntactic categories that are involved in syntactic selection. The semantic side, however, as is evident from the previous discussion, has received many different analyses, with no one accepted approach.
The representation that we propose is guided by a number of principles:
-
∙ The semantic roles need to be semantically contentful and clearly defined.
-
∙ The semantic roles should be language-independent.
-
∙ The semantic roles should be consistent across all predicates, that is, a semantic label should have the same denotation regardless of the predicate.
-
∙ There should be one argument per semantic role.
As shown, such principles cannot be maintained with a very limited set of roles (e.g., four in delph-in grammars, or five in PropBank). However, bearing in mind the cautionary words of Copestake (Reference Copestake2009) (i.e., ‘slacker semantics’), it is essential to maintain a realistic implementationally viable number of roles which will be compatible with the requirements of implementing a broad-coverage grammar. Thus, with respect to the richness of the semantic representation we attempt to find an optimal compromise between elaborate semantic representations and realistic grammar engineering.
Moreover, aside from the size of the roleset, an additional design decision stems from the nature of the task. Our main efforts in the implementation of the grammar are focused on accounting for the syntax of the language, not on implementing a semantic theory. Consequently, we adopt a ‘pure’ argument indexing approach, whereby semantic roles are necessarily linked to syntactic arguments. While this approach does not capture the fine-grained semantic distinctions that some of the semantics-centered approaches we reviewed express, the analyses produced by the grammar are semantically informative and potentially useful for tasks that require a deeper understanding of natural language.
Finally, the correspondence between the syntactic and semantic levels is captured in our system by the association of each semantic role with a class of syntactic categories that are potentially interchangeable with respect to the realization of that role. We will refer to them here as realization classes. This idea is close in spirit to Grimshaw’s and Jackendoff’s association of semantic primitives with syntactic categories. Yet, contrary to them, we do not assume unmarked or canonical structural realizations.
3.1 Semantic roles and syntactic realization classes
The semantic roleset is derived from the study of existing rolesets, described in Section 2, and corpus-based data. The corpus data include at least 100 randomly selected examples of sentences containing each of the 50 most frequent verb lemmas in the 60-million token WaCky corpus of Modern Hebrew (Baroni et al. Reference Baroni, Bernardini, Ferraresi and Zanchetta2009). First, the arguments of each verb in all its instances were identified.Footnote [16] Then, we attempted to label each argument with a semantic role, first considering the applicability of the roles that were proposed in the literature we reviewed. Consequently, there is considerable overlap between our roleset and existing rolesets. Most notably, the list proposed by Jaworski & Przepiórkowski (Reference Jaworski and Przepiórkowski2014b) bears the most resemblance to our proposed list, since it conforms with most of our desiderata.
Following are the semantic categories we propose, along with their corresponding characterizations. We illustrate each category with an English example to simplify (and shorten) the presentation.
- ( $\mathit{Arg}1$ )
-
The prototypical member of this category is an Actor. Also included are Perceiver and Causer.
- ( $\mathit{Arg}2$ )
-
This category corresponds to what is often referred to as Theme. Members of this category are negatively characterized as not belonging to the other, more semantically coherent, categories.
- ( $\mathit{Arg}3$ )
-
Members of this category are prototypically animate beings which are positively or negatively affected by the action or state. It includes arguments that can be labeled Affectee, Benefactive, Malfactive, or Recipient.
- ( $\mathit{Arg}4$ )
-
This category includes semantic arguments that are Attributes, or, in other words, predicative of another argument in the clause.
- ( $\mathit{Arg}5$ )
-
This category includes semantic arguments that denote Source in both its physical and metaphorical meanings.
- ( $\mathit{Arg}6$ )
-
This category includes semantic arguments that denote Goal in both its physical and metaphorical meanings.
- ( $\mathit{Arg}7$ )
-
This category includes semantic arguments that denote Location in both its physical and metaphorical meanings.
- ( $\mathit{Arg}8$ )
-
This category includes arguments that denote the Topic (of Communication).
- ( $\mathit{Arg}9$ )
-
Members of this category denote Instrument arguments, usually an inanimate entity causally involved in the action.
- ( $\mathit{Arg}10$ )
-
Members of this category denote Comitative arguments, usually an animate being who accompanies another participant of the event.
The coherence of this roleset was put to the test by having two annotators (one of whom was the original annotator) annotate a new set of 200 sentences, which included 10 sentences each for 10 verbs from the original frequent verb set, and 10 each for additional (new) verbs. The annotation of semantic roles was compared, and there was 86% (172/200) agreement between the two annotators, where ‘agreement’ was defined as annotating the exact same semantic roles. There were a few discrepancies, but all were easily resolved by a short discussion between the annotators. These include distinctions between complements and adjuncts, between unaccusative and unergative verbs, between Recipient (Arg3) and Goal (Arg6), etc.
The association of the arguments of the 50 most frequent verbs with the semantic roles described above revealed syntax–semantics correspondences, or, as we refer to them here, realization classes. The list of the semantic roles along with their realization classes is presented in Table 4, and the number of instances of each role is presented in Table 5.
Each verb lexeme in the grammar is associated with two types of argument structure information: semantic frames and syntactic realization choices. Semantic frames are similar to Fillmore’s case frames, which list the semantic arguments of the verb (e.g., AIO: Agent–Instrument–Object). In our schema they are expressed by referring to the numbered labels of semantic roles (e.g., arg129 for Agent–Object–Instrument).Footnote [18] Corresponding to the semantic frame is a syntactic component which states for each semantic argument which syntactic phrase types are used to realize it. The choice of syntactic phrase types is restricted by the realization class associated with each semantic role.Footnote [19]
Whereas the semantic classes are expected to be more or less universal, some language-specific differences are predicted to be found in the syntactic realizations. Corpus investigations in the context of the development of AraGram, an HPSG grammar of Modern Standard Arabic, confirmed these expectations (Arad Greshler et al. Reference Arad Greshler, Herzig Sheinfux, Melnik and Wintner2015). The Arabic verbs that correspond to the 50 most frequent Hebrew verb lemmas were found to share the same semantic frames. Slight differences were found in the syntactic realization classes of Arg2 and Arg6.
As was shown by Grimshaw (Reference Grimshaw1979), the two levels are crucial since one cannot be deduced from the other. NP complements may play different roles in their relations with their selecting predicate (e.g., Theme, Benefactive, Attribute). Conversely, semantic roles can be realized by phrases of different syntactic categories. Moreover, syntactic selection is largely lexeme-specific, yet constrained by the realization class of the respective semantic role (cf. Grimshaw’s ask and wonder).
In order to illustrate the schema without going into HPSG-specific details, we use in this section a semi-formal notation to describe lexical entries. The description of each entry includes (1) the semantic arguments that the verb selects, (2) the realization class of each argument, (3) the realization frames (r-frames) that the verb licenses. The latter property is indicated by a string such as arg1/12/129, where a slash appears between frames to indicate disjunction, and in each frame the numbers indicate which arguments are realized.Footnote [20]
For example, the lexical entry in (27) describes a predicate that selects three semantic arguments: Arg1, Arg2, and Arg9. Arg1 and Arg9 are realized by NP and PP, respectively. Arg2 can be realized by either NP, CP, or PP. Furthermore, the predicate can be realized in three different frames: all arguments are realized (arg129), Arg2 is omitted (arg19), or both Arg2 and Arg9 are omitted (arg1).Footnote [21]
Although this information is stated explicitly in our semi-formal lexical entries, in the grammar itself lexemes are instances of lexical types which are cross-classified according to the different argument slots and realization specifications. This type of architecture is similar in spirit to the one proposed by Davis & Koenig (Reference Davis and Koenig2000), where more specific types multi-inherit from a number of more general types. Thus, similarly to Davis & Koenig’s act-und-rel, which inherits from act-rel and und-rel, lexemes such as the one illustrated in (27) are instances of a lexical type that inherits from arg1_n, arg2_ncp and arg9_p,Footnote [22] each contributing argument-specific information. In a type inheritance hierarchy, generalizations can be stated once for types at the appropriate level, and this information is further inherited by all subtypes of that type. With multiple-inheritance, constraints regarding syntax and semantics combine together to account for the argument structure of lexical types.
3.2 Benefits for linguistic analysis
The proposed approach involves significant changes to the way in which argument structure is viewed. The precise HPSG-specific aspects of this system are discussed in Section 4, which describes the way in which it is implemented. In this section, however, we adopt a framework-neutral perspective.
The system that we propose here distinguishes between semantic selection and syntactic selection, and provides a way of stating constraints regarding each level separately. More specifically, with this system it is possible to define constraints that target Theme arguments, regardless of their syntactic category, or constraints that distinguish between PPs that realize Recipients and PPs that realize Goals. This extends the expressive power of the HPSG approach and provides a way to account for phenomena that are better characterized in semantic terms. The following sections illustrate some of these phenomena.
3.2.1 Multiple subcategorization
Section 1 provided a short discussion of the phenomena that prompted Grimshaw (Reference Grimshaw1979) and Jackendoff (Reference Jackendoff1985) to consider the relationship between syntax and semantics in the domain of argument structure. In the two aforementioned studies the respective authors considered cases of multiple subcategorization and their implications for theories of argument structure.
The fact that predicates can appear in a number of different subcategorization frames is not difficult to capture. It is always possible to posit multiple lexical entries for a lexeme, each corresponding to one subcategorization frame. However, such a solution, we claim, overlooks important generalizations regarding the interchangeability of syntactic realizations of semantic arguments.
Consider as an example the three different subcategorization frames that are licensed by the verb raca (‘want’):
The complements of the verb denote whatever is wanted by the referent of the subject, regardless of whether the complement is a Thing, realized as an NP, or an Event, realized as a VP or CP. Thus, we posit that the three phrase types are members of one realization class which corresponds to the semantic role associated with the complement.
The same phrase types (i.e., NP, VP, CP) are also used as complements of the verb lamad (‘learn’), as is shown below.
The fact that these phrase types are found to be interchangeable as complements of various verbs suggests that they form a class. In our schema, this class is associated with the Theme argument role, annotated as Arg2. Under this type of analysis, the similar multiple subcategorization patterns observed for the two distinct verbs are not viewed as coincidental. Rather, they are captured by the schema, which assumes the correspondence between semantic roles and realization classes.Footnote [23] Consequently, the lexical entries of the two lexemes share the same argument structure specification.
3.2.2 Argument optionality and co-occurrence restrictions
It is well known that not all semantic arguments need to be realized syntactically. Argument optionality is usually indicated in the subcategorization notion shown in Table 1 above, and in (‘pen-and-paper’) phrase structure grammar by surrounding the phrase type with parentheses. This would suggest that argument optionality is a syntactic constraint. However, we argue that the actual level where constraints regarding the optionality of arguments should be stated is the semantic level. For example, the fact that the NP/CP/VP complement of the verb raca (‘want’) in sentences (28) above is obligatory need not be stated about each subcategorization frame separately. Rather, it is a general constraint about the Theme argument of this verb, regardless of how it is realized. Stating this separately for each subcategorization frame misses an important generalization.
Let us consider the following two example sentences of the verb amar (‘tell’), where optionality is indicated by parentheses.Footnote [24]
We claim that the fact that the NP in (31a) and CP in (31b) are both not optional is not coincidental; they both realize the same semantic role, Theme, and this semantic argument is obligatory in the case of amar (‘tell’). Conversely, Arg3, the Recipient, and Arg8, the Topic of Communication, are optional. This generalization can be simply stated in our system, as is evident from the argument structure specification of the lexical entry of amar (‘tell’) in (32); Arg2 is associated with the set containing both NP and CP, and all the disjoined frames in r-frames contain ‘2’.
An additional phenomenon that our system provides the means to capture involves cases where the realization of one argument blocks the realization of another, although they both play different semantic roles. Consider the following example:Footnote [25]
The complements of xašaš (‘fear’) in (33a) all denote negative outcomes of which the referent of the subject is afraid, and each is realized by a different syntactic phrase type (PP $_{\mathit{mi}}$ , PP $_{\mathit{le}}$ , VP $_{\mathit{inf}}$ , and CP, respectively). The complement of the verb in (33b), on the other hand, denotes the entity that can be badly affected by the negative outcome. Although the two arguments have clearly distinct semantic roles, and there does not seem to be a sense difference between the two xašaš (‘fear’) in (33), these arguments cannot co-occur in the same clause, as is shown in (33c).Footnote [26]
This co-occurrence restriction can be stated once, as a semantic constraint on the co-occurrence of Arg2 and Arg3. More specifically, the verb xašaš (‘fear’) is defined as semantically selecting three arguments (Arg1, Arg2, Arg3), yet its semantic realization frame specification is arg12/13, thus ruling out the co-occurrence of Arg2 and Arg3.
Here, too, the restriction is stated in terms of semantic selection, and is independent of the multiple syntactic realization options shown in (33a) for Arg2.
3.2.3 Multiple subcategorization versus polysemy
When syntactic and semantic constraints are conflated, multiple subcategorization and polysemy can be indistinguishable: both phenomena involve multiple lexical entries for a given lexeme. In our system we consider multiple subcategorization as a phenomenon where an identical semantic frame can be syntactically realized in different ways, either through argument optionality, or in terms of the use of different phrase types within the same realization class. As shown in the previous sections, this phenomenon can be defined once for one lexical entry. Polysemy, on the other hand, involves different senses, which may be associated with different semantic frames. In this case distinct lexical entries are defined, one per sense.
We will illustrate this point with the verb amar (‘tell’). The sense denoted by the verb in (31) belongs to FrameNet’s Telling frame, whereby ‘A Speaker addresses an Addressee with a Message, which may be indirectly referred to as a Topic.’ There is, however, a related sense of the verb, which is associated with a Request frame.Footnote [27] This is illustrated by the following sentence.
The semantic frame of this sense contains three obligatory arguments: a Speaker, an Addressee, and a Message (in FrameNet terms). In our system, the Addressee is categorized as the Arg3 argument and the Message as Arg2. These three semantic arguments are shared between the two senses. However, unlike the Telling sense, the Request sense is incompatible with a Topic (Arg8) argument. Consequently, this sense is associated with a single semantic frame, arg123, and a syntactic constraint which indicates that Arg2 should be realized as VP $_{\mathit{inf}}$ and Arg3 as PP. An additional characterization of this lexeme is the control pattern, whereby the unexpressed Agent (Arg1) of the VP complement is construed as (or controlled by) the referent of the object of amar (‘tell’), the Arg3 argument.Footnote [28]
Consequently, only two lexical entries are needed in order to account for the multiple subcategorization as well as the polysemy of amar (‘tell’); one entry per sense (36).
3.2.4 NPs and their semantic role
One aspect of the relationship between syntax and semantics in the domain of argument structure is expressed in the correlation between the morphosyntactic properties of NP complements, their syntactic function, and their semantic role. A case in point is their case marking patterns.
Noun phrases belong to four realization classes: Arg1, Arg2, Arg3, and Arg4. Following are examples of NPs in each of these roles, along with their semantic labels:
Accusative case in Hebrew is marked with the case marker et only on definite objects.Footnote [29] Thus, the indefinite object maraq (‘soup’) in (37a) is unmarked, while the definite NP in (37b) is obligatorily preceded by the accusative marker. Accusative case marking also appears on the Affectee/Recipient complement ha-yalda (‘the girl’) in (37c), which is associated with Arg3. However, when NPs function as predicates they are never marked with accusative case, regardless of their definiteness status. This is illustrated in the copular construction in (37d) and in (37e), where ha-more šeli (‘my teacher’) and are definite but unmarked for accusative case.
Although all NP complements in (37) are syntactic complements of the verb, their semantic role determines their morphosyntactic behavior. The correlation between the two domains is easily captured in our system, where NP complements are mapped to three distinct semantic roles. The grammar constrains definite Arg2 and Arg3 NP complements to be marked with accusative, and Arg4 NP complements to be unmarked. This generalization applies across the board, regardless of the construction in which NP complements appear.
3.2.5 Semantic PPs and argument-marking PPs
Similarly to noun phrases, prepositional phrases also belong to different realization classes. However, not all PPs are equal with respect to their semantic content. Sag et al. (Reference Sag, Wasow and Bender2003) distinguish between semantic prepositions and argument-marking prepositions, where the latter do not contribute to the semantics of the clause. Technically, argument-marking prepositions project the index of their NP complements to the PP level. Sag et al. (Reference Sag, Wasow and Bender2003) assume that this information is lexically specified for prepositions, and illustrate this with the lexical entry of to (p. 212). In doing so they can account for cases such as They talk to themselves, where the complement of the preposition to is a reflexive that is bound by the subject.
There are, however, some prepositions that can appear in both guises. Consider, for example, the following pair of sentences:
The preposition be (‘in’) (or in its definite form ba) is an argument marker in (38a), since it does not contribute any semantic content to the clause (as is evident from the gloss). In (38b), on the other hand, the preposition indicates a location, and can be replaced with other location denoting prepositions such as , leyad (‘next to’), or mul (‘opposite’), resulting, of course, in different meanings.
One possible way to account for this is to posit two different lexical entries for the preposition, one for each function. We, however, propose that the distinction between the two functions of the preposition is not lexically specified; there is only one lexical entry for be (‘in’), which is underspecified with respect to its semantic status. When heads combine with a PP, the linking of the argument depends on the semantic role of the PP. PPs that are associated with Arg1, Arg2, and Arg3 are considered to be argument markers, and, as such, when they combine with a head, their NP complement (or, more precisely, its index) is considered to be the semantic argument. Semantic PPs, on the other hand, which are associated with Arg4–Arg10, contribute their semantic content to the composition of the semantics of the clause. The technical specifics of this analysis are described in Section 4.4.1 (Figures 12 and 13).
Here, too, the association of complements with finer-grained semantic roles provides a way to express generalizations regarding the behavior of these complements. The proposed analysis attributes the property of being an argument marker or a semantic argument not to the PP itself, but rather to the relation between the head and the PP. This eliminates the necessity to posit two distinct lexemes in cases where a preposition serves both functions (with different verbs).
3.2.6 Complement order
An additional phenomenon that we find to be governed by semantic constraints is complement order. Unlike English, the order of complements in Modern Hebrew is fairly free. For example, the Arg2 and Arg3 complements of the verb natan (‘give’) in (39) can appear in any order.Footnote [30]
A closer examination of the data, however, reveals that not all complement orders are possible. One case that exhibits order constraints is the causative construction, exemplified below:
While some complement orders may sometimes ‘sound better’ than others, the sentences in (40c) and (40d), where the predicative Attribute (Arg4) precedes the Affectee object (Arg3), are unequivocally ungrammatical, regardless of the syntactic category of the Attribute.Footnote [31]
Consequently, similarly to the phenomena discussed in the previous sections, here too we suggest that the relevant level at which such constraints on complement order apply is the semantic level. More concretely, regardless of the type of syntactic phrase that realizes the predicative complement (Arg4), it cannot precede the Arg3 complement, provided that they are both required. Naturally, this particular constraint regarding complement ordering is specific to Modern Hebrew. Nevertheless, our approach provides a way of stating semantically based generalizations regarding complement order where applicable.Footnote [32]
3.2.7 Summary
We reviewed in this section a number of phenomena for which an account that is based on semantics is preferable to one that targets syntactic arguments. We maintain that our proposed list of semantic roles and their realization classes (Table 4), coupled with the lexical representation of verb arguments illustrated above, provides the correct granularity with which argument structure should be expressed. In particular, as we show above, it enables the specification of various generalizations, both semantic and syntactic, that would have been lost with existing approaches.
An additional benefit of our approach to the representation of argument structure is more practical. Computational grammars can be used for parsing, and the analyses provided by our implemented grammar of Hebrew can be used to drive downstream applications which will be able to utilize the semantic representations that the grammar produces. The next section discusses the actual implementation of the grammar that makes such applications possible.
4 Implementation in HPSG
The approach to argument structure that we have presented is implemented in HeGram, a grammar of Modern Hebrew implemented with the LKB (Copestake Reference Copestake1999, Reference Copestake2002b), a grammar development environment. The grammar is based on a starter grammar created with the LinGO Grammar Matrix customization system (Bender, Flickinger & Oepen Reference Bender, Flickinger and Oepen2002). Nevertheless, a number of major revisions were made to the ‘standard’ Matrix-based grammar in order to incorporate this approach:
-
∙ The valence feature structure is designed to distinguish between the ten different argument categories.
-
∙ Lexical types are cross-classified according to their semantic and syntactic selection.
-
∙ Linking between the argument slots of a predicate’s key relation and the indices of its arguments is done at the phrasal level.
The interplay of syntax and semantics with respect to argument structure is captured on different levels of generalization: in the lexicon, lexical type hierarchy, and phrasal type hierarchy. This section provides an overview of the essential components of the implemented grammar, from the lexical to the phrasal level.
4.1 The valence complex
The heart of the analysis lies in the valence feature, which in HeGram is split across xcomp, which accounts for complex predicates, ten dep features, each corresponding to a semantic role, and a number of additional features which will be described shortly. The distribution of arguments across several dep features, in contrast to the standard HPSG use of subj and comps, is inspired by Haugereid (Reference Haugereid2012). However, the specific deps, and in particular their association with semantic roles, are novel. An abbreviated description of valence is given in Figure 4.
R(ealization)-frames and s(emantic)-frame are two features that indicate the types of frames compatible with a particular lexical type. The value of the two features is of type link. R-frames, which was introduced in the description of lexical entries in Section 3.2, specifies the realization frames in which the verb can appear. For example, the r-frames value of the first sense of amar (‘tell’) is arg12/123/128/1238. This feature constrains the combination of a head with its subject or complements. s-frame specifies the verb’s semantic arguments. Thus, the s-frame of the aforementioned sense of amar (‘tell’) is arg1238. The two distinct features are needed in order to account for cases, such as raising or expletives, where a verb combines with a phrase that is not its semantic argument. For example, raised arguments appear in r-frames, but not in s-frame. This information is relevant at the phrasal level, where ‘real’ semantic arguments (i.e., those that appear in s-frame) are linked to semantic argument slots, while raised arguments are licensed as arguments but not linked.
Subj-arg is used to single out the dep that functions as the subject. This is explained in more detail in Section 4.4.2. The standard subj and comps list members are ‘spread’ across the ten dep features. True to the ‘one argument per label’ principle, the value of each dep feature is one entity of type dependent, rather than a list. The dependent complex has two features: dep, whose value is synsem, and real, whose value is of type link. Real is used to record the realization of arguments, and is described in more detail in Section 4.4.1.
Finally, the ppsort complex is used to restrict the types of PPs that can combine with the head for each dep. The value of each of its dep-p features is of type prep-p, which subsumes atomic preposition types, as well as disjunctive ones.Footnote [33]
4.2 Semantic selection: The $\text{link}$ hierarchy
The link type hierarchy defines the possible values of r-frames, s-frame, subj-arg, and real (Haugereid Reference Haugereid2012). It is an elaborate hierarchy, where each semantic frame combination is defined. A highly abbreviated hierarchy illustrating the intricate inheritance relations is given in Figure 5.
There are 20 immediate subtypes of link, each representing a positive or a negative value of one of the ten args. Each ‘leaf’ in the hierarchy is an (immediate or non-immediate) subtype of one of each pair: arg1+, arg1–; arg2+, arg2–; arg3+, arg3–, etc. For example, arg12 is a subtype of the positive arg1+, arg2+ and the negative arg3– through arg10– (not all of which are shown).
Between the general link type and the most specific leaf types are disjunctive types. For example, arg12/13 inherits from the positive arg1+ and the negative arg4– through arg10–. It does not inherit from arg2 and arg3, since they are the features where its two disjuncts vary. Arg12/13’s immediate subtypes, arg12 and arg13, each representing one of its disjuncts, fully specify their arg2 and arg3 values. Arg12 inherits from the positive arg2+ and negative arg3–, whereas arg13 inherits from arg2– and arg3+.
The r-frames feature of predicates determines the combination of a head with its dependents. In order for a predicate to combine with, for example, a dep2, two conditions must hold. First, its r-frames value must be unifiable with arg2+, thus indicating that it indeed selects for an Arg2. Second, the real value of its dep2 must be the negative arg2–. Predicates are lexically defined as having negative values in the real feature of each of their ten deps. Thus, for example, the value of real in the dependent value of dep1 is arg1–, and similarly in the rest of the deps. Once a dependent is realized, its real value is set to its respective positive value (arg1+ in the case of dep1).
The elaborate structure of the hierarchy is designed in order to support the definition of disjunctive semantic frames such as arg12/13. This type is unifiable with either arg2+ or arg3+, and consequently verbs with r-frames of this value are compatible with either Arg2 or Arg3 complements. However, once unification occurs, when the verb combines with the appropriate argument, the r-frames value ‘settles’ on either arg12 or arg13. This is the key to the mechanism that guides the combination of verbs with their arguments. Disjunctive frames notwithstanding, VPs are required to realize all of the arguments defined in one of the r-frames of their verbal head.Footnote [34]
4.3 Syntactic selection: The $\text{top}\text{-}\text{deps}\text{-}\text{lxm}$ hierarchy
The specification of the r-frames and s-frame values is essentially semantic selection. Syntactic selection involves the specification of constraints on the synsem values of the relevant dep features. Thus, for example, the verb xašaš (‘fear’) in (33) (repeated here in (41)) semantically selects an Arg2 and an Arg3. The realization class associated with Arg2 includes NP, VP, CP, and PP. Of them, xašaš (‘fear’) syntactically selects VP, PP, and CP. Arg3, on the other hand, can only be realized as a PP, a subset of its realization class $\{\text{NP},\text{PP}\}$ . This information is stated in the synsem value of dep3.Footnote [35]
The valence feature structure of xašaš (‘fear’) in (41) is presented in Figure 6. Note that disjunctive syntactic selection is represented as a disjunctive head value.Footnote [36] Adp is shorthand for adposition, an order-neutral term for what is usually referred to as ‘preposition’.
Naturally, information regarding the syntax–semantic correspondences in argument structure is not posited individually for each lexeme. Rather, the lexical type hierarchy in HeGram cross-classifies lexical types according to the different argument slots. Its most general type is top-deps-lxm, which dominates all lexical types whose instantiations serve as selecting heads. The immediate subtypes of top-deps-lxm are associated with each of the dep features, and specify a maximal disjoint head value (if applicable). The two highest levels of the hierarchy are illustrated in Figure 7.
To illustrate the structure of the hierarchy, we will focus on the sub-hierarchy associated with dep2. Following is the type definition of arg2_nvpc, the most general type which dominates types of Arg2-selecting predicates. Note that the type definition below, specified in the TDL formalism (Copestake Reference Copestake, Oepen, Flickinger, Tsujii and Uszkoreit2002a), determines the immediate supertype(s) of the particular type (top-deps-lxm in this case), as well as type-specific constraints.
Note the disjoint head value, nvpc, which corresponds to the realization class associated with this argument $\{\text{NP},\text{VP},\text{PP},\text{CP}\}$ . The semantic arg2 specification ensures that the semantic relation denoted by an Arg2-selecting predicate includes an arg2 slot, regardless of whether this argument is syntactically realized or not.
Below each of the dep-related general types (e.g., arg2_nvpc, arg3_np) is a full hierarchy which fleshes out all of the different possible head values. The hierarchy associated with arg2 is spelled out in Figure 8.
Maximal verbal types inherit from the appropriate dep-related types. Following are examples of definitions of two top-deps-lxm subtypes, each associated with a different sense of amar (‘tell’). The two definitions correspond to the lexical entries given in (36) (Section 3.2.3). Note that the type associated with the Request sense inherits from arg2_v, which defines an atomic head value in its dep2, while the type associated with the Telling sense inherits from arg2_nc, which imposes a disjoint head value (NP or CP). In addition, the Telling sense semantically selects an additional argument (the Topic of Communication, or Arg8), and consequently inherits from the type associated with it, namely arg8_p.
4.4 Semantic linking at the phrasal level
HeGram, similarly to grammars implemented within delph-in, produces MRS representations of input sentences, which include specified and underspecified scope information. The key distinction between the MRSs produced by the two approaches is in the argument roles. For example, the semantic relations produced for both ate and fly in (9a) and (9b), respectively, contain arg1 and arg2. HeGram produces more informative representations, where the relations’ arguments vary according to their semantic roles: arg1 and arg2 (Theme) for ate, and arg1 and arg6 (Goal) for flew.
An additional difference between the two approaches is the level at which argument slots in the semantic relation are linked to the arguments themselves. In standard HPSG this is part of the lexical definition, while in HeGram linking occurs at the phrasal level. This is a direct consequence of the approach which enables the definition of disjunctive argument frames (e.g,. Arg2 can be realized by either NP or CP). Essentially, arguments are linked to argument slots in the head’s key relation as they combine with it.
4.4.1 Head–Complement
A hierarchy of head–complement phrase types is defined in order to license the combination of heads with their complements and to define the appropriate semantic links between the semantic relation of the selecting head and its arguments. Semantic linking, as previously mentioned, is standardly assumed to be handled lexically, yet here it is done at the phrasal level.
Consider, for example, a verb such as gila (‘discover’), which may take either an NP or a CP as a complement.
In a Matrix-induced grammar, this verb would have had two lexical entries, each associated with a different lexical type.
Each lexical type handles the linking of the semantics of the complements with the semantic relation of the verb differently, according to the type of the complement.Footnote [37] When the verb combines with an NP complement (45), the arg2 of the semantic relation denoted by the verb is linked with the referential index of the NP. Conversely, when the complement is a CP (46), the content of arg2 in the semantic relation of the verb is a handle (ltop). In addition, a qeq relation is introduced in hcons, and it links that handle with the label of the relation denoted by the CP.
The HeGram semantic representation of a clause headed by such a verb is identical, yet the mechanism employed to form it is substantially different. The disjunctive nature of the argument frames of gila (‘discover’) is reflected in its lexical definition as an instance of the type arg12_nc, which inherits from arg2_nc (see Section 4.3). The semantic linking between arguments and argument slots in the semantic relation occurs at the phrasal level.
The basic head–comp phrase type is a very general type. Its main function is to percolate valence features which are not relevant to the combination of a head with its complement from the head daughter to the phrase. Its immediate subtype is basic-head-init-comp-phrase, which fixes the order to be head-initial. This type, in turn, immediately dominates nine head–comp types, each pertaining to one of the nine dep features: head-comp2-phrase, head-comp3-phrase, head-comp4-phrase, head-comp5-phrase, head-comp6-phrase, and so on.Footnote [38] In what follows we will focus on head-comp2-phrase and its subtypes in order to illustrate this aspect of the system. The same mechanism applies to all of its ‘sister’ phrase types.
As is shown in Figure 9, head-comp2-phrase dominates four subtypes, each pertaining to a different type of syntactic phrase within the realization class associated with Arg2 (i.e., NP, CP, VP, and PP). Head-np-comp2-phrase is the equivalent of transitive-lex-item in terms of the semantic linking it imposes between the NP complement and the semantic relation denoted by the selecting head. Likewise, for CP complements, the semantic links and the qeq relation in hcons, which are defined lexically in the Matrix (see clausal-second-arg-trans-lex-item in (46)), are the responsibility of hd-cp-comp2-ph. Similar phrase types are defined for the different head-comp phrases.Footnote [39]
Consider the type definition of head-comp2-phrase given in Figure 10. This definition states that a head can combine with a phrase in a head-comp2-ph configuration if and only if:
-
∙ The head subcategorizes for an Arg2.
-
∙ The non-head-dtr is unifiable with the dep2 requirement of the head.
-
∙ The Arg2 requirement of the head has not yet been fulfilled.
-
∙ Arg2 is not the head’s subject.
Subcategorization constraints are represented in r-frames. The value of the r-frames of the head daughter, tagged in Figure 10, is passed on, along with other valence features, to the mother. Moreover, it is structure-shared with a positive link-typed feature. The structure-sharing of the value of r-frames with arg2+ has two functions. First, it constrains the r-frames value of the head daughter to be compatible with arg2+. Thus, for example, a verb whose r-frames value is arg12/16 is compatible, while arg16 is not. Second, it can ‘force’ disjunctive r-frames values to be more restrictive. Thus, a verb with an r-frames of arg12/16 which combines with an NP complement creates a phrase whose r-frames value is the unification of arg12/16 with arg2+, namely arg12. Consequently, once such a verb combines with an NP complement it can no longer combine with a dep6 argument, since its r-frames value is not compatible with arg6+.
In addition, the combination of a head with a complement phrase requires that the non-head daughter be compatible (i.e., unifiable) with the complement requirement specified in the valence feature of the head daughter. The structure sharing between the synsem of the non-head daughter and the dep value of the head daughter’s dep2 value (both tagged ) ensures just that.
Unlike in the prevalent approach in HPSG, valence requirements are not canceled off as they are realized. Instead, the grammar keeps track of which arguments have been realized by way of the real(ized) feature, whose value is of type link. Lexical items are defined to have negative real values in each of their deps. Once a head combines with an argument, either a subject or a complement, the value of real in its respective dep is set to a positive value. This is shown in Figure 10, where the value of real of the dep2 of the head daughter is a negative arg2–, and in the mother it is a positive arg2+. Thus, the specific head–comp rule can only apply once.
The real feature is multi-purpose: it allows for free complement order, which is much more difficult to account for with a comps list; it enables us to control the complement order when necessary;Footnote [40] and it supports disjunctive frames in restricting the realization to only one of the disjuncts (see (33) and Figure 5).
Finally, the structure-sharing of a negative arg2– (tagged ) with thesubj-arg value of the head daughter determines that dep2 is not the head’s subject. The subj-arg features plays a key role in the head-subject-phrase, described in Section 4.4.2.
The type definition of head-comp2-phrase targets the syntactic aspects of the combination of a head with its complement. This is common to all Arg2 complements. The semantic linking between the arguments in the relation denoted by the verb and the complements is subject to phrase-specific constraints, which are defined for each subtype of head-comp2-phrase.
Figure 11 illustrates the type definition of head-cp-comp2-phrase, a subtype of head-comp2-phrase, which accounts for the combination of heads and Arg2 CP complements. As previously mentioned, this phrasal type mimics the semantic linking which is defined lexically in the Matrix (see clausal-second-arg-trans-lex-item in (46)). As this phrase type inherits from the more general head-comp2-ph, the constraints defined here are only those that are specific to CP complements.
The type-specific information includes all of the information defined in the Matrix for clausal-second-arg-trans-lex-item in (46) above. The creation of a qeq relation in hcons is defined in the c-cont feature (Copestake et al. Reference Copestake, Flickinger, Pollard and Sag2005), whose function is to introduce construction-specific (not lexically based) relations to rels. Note that the head-dtr’s toprel feature points to the key relations of the lexical item or the phrase. Its value is set at the lexical level, and it percolates upwards, along with the entire hook complex of the head daughter, for all types that inherit from head-compositional.Footnote [41]
Phrase-type-specific linking constraints are also responsible for capturing the distinction between argument-marking PPs and semantic PPs described in Section 3.2.5. Figure 12 illustrates head-pp-comp2-phrase, the type that licenses the combination of a head with an argument-marking PP, while in head-pp-comp7-phrase (Figure 13), the complement is a semantic PP. The two phrase types license sentences (38a) and (38b), respectively. The common property of the two is that they both require that the pred value of the preposition match the restrictions imposed by the head on their respective dependent. This is defined by the structure-sharing tagged . The two types diverge with respect to the semantic contribution of the complement. The arg2 argument of the head of head-pp-comp2-phrase is identified with the arg2 of the relation denoted by the PP (which in turn is identified with the index of its NP complement). This captures the fact that the head of the PP is only an argument marker. Conversely, with semantic PPs, the arg7 in the relation denoted by the head is structure-shared with the ltop (handle) of the semantic relation denoted by the PP.
4.4.2 Head–Subject
The ten dep features in valence are first and foremost associated with semantic arguments. Consequently, there is no one dep that represents the subject.
As an illustration of this point consider the following pair:
The two verbs are intransitive, yet there is a clear difference between the semantic role of their subject: the subject of the unergative qafac (‘jumped’) is an Actor (Arg1), and the subject of the unaccusative nišbera (‘broke’) is a Theme (Arg2). The difference between the two verbs is reflected in their valence features. Their respective subject requirements are defined in dep1 or dep2, depending on the verb type. Moreover, the subject dependent is identified by the subj-arg feature, which, too, is part of the valence feature complex. The role of subj-arg is to determine whether a combination of a dependent is licensed by head–subj or head–comp.
Similarly to the head–complement phrase types, head–subject phrases too are defined in a hierarchy, where specific phrase types target specific realizations. As an example, consider the type definition of head-subj1-phrase, a subtype of the more general head-subj-phrase, given in Figure 14. This phrase type licenses the combination of a VP with its Arg1 subject.
The subj-arg value of the head daughter is restricted to Arg1. Moreover, the structure-sharing of the real values of all deps except for dep1 (the subject) determines that the head daughter has realized all of its dependents, except dep1, whose real value is negative.
The synsem of the non-head daughter matches the dep value of dep1. Semantic linking is set by structure-sharing the index of the non-head daughter (i.e., the subject) with the value of arg1 in the toprel of the selecting head. In addition, subject–head agreement is reflected in the structure-sharing of the cncrd feature of the head daughter and the non-head daughter. Finally, the phrase rule sets the real feature in dep1 of the mother to arg1+, thus preventing it from combining with another dep1 again. The combination of Arg2 subjects with VPs is licensed by the ‘sister’ phrase type, head-subj2-phrase, which is identical to the phrase type in Figure 14, except that all specifications of dep1 and Arg1 are replaced with dep2 and Arg2.
5 Conclusions
We proposed an analysis that accounts for both the syntactic and the semantic aspects of argument structure. We maintain that argument structure phenomena can be best explained by distinguishing between syntactic selection and semantic selection, while accounting for the correspondences between the two levels.
The schema that we proposed here is fully implemented in HeGram, an HPSG grammar of Modern Hebrew. With this novel infrastructure in place, HeGram covers various constructions including raising, control, unbounded dependencies (wh-questions, topicalization), the Hebrew copula construction, including zero copula, complex predicates, and an inverted (V2) construction.
We are actively working on incorporating verbal multi-word expressions (MWEs) into the grammar (Herzig Sheinfux et al. Reference Herzig Sheinfux, Arad Greshler, Melnik and Wintner2015). We view MWEs as a challenging test case for the innovative architecture that we have implemented. In a way, MWEs constitute an extreme case of multiple subcategorization, so accounting for them is a natural next step in the development of the grammar. We expect that the finer semantic distinctions that our system allows, as well as the explicit differentiation between syntactic constraints and semantic constraints, will prove to be beneficial to the incorporation of MWEs into the grammar.
We regularly run [incr tsdb()] (Oepen Reference Oepen2001) on HeGram, whenever we implement changes in the grammar, and expand our test suite, which currently comprises 483 sentences, out of which 117 are ungrammatical. These sentences, for the most part, target the phenomena that we cover, and a few of them reflect phenomena that we have not covered yet. HeGram is freely available for download.Footnote [42]
One argument-structure phenomenon that is not yet accounted for by our grammar is valence alternation. Although this study addressed two phenomena that are often included under the term ‘valence alternation’, namely argument optionality and alternate syntactic realization, we view them as separate phenomena. The distinction is based on systematicity. The notion that the syntactic realization of semantic complements is mostly idiosyncratic (within the confines of realization classes) featured in the first example in this paper (i.e., ask and wonder in (1)). The idiosyncratic and lexeme-specific nature of argument optionality is often illustrated with the verbs dine, devour, and eat, which denote similar activities associated with the consumption of food, yet vary with respect to the optionality of their complement. By ‘valence alternations’ we refer to systematic correspondences between different realizations of argument structure for a given lexeme. Common cases are active–passive, causative, and applicative, which generally involve morpholexical operations, but also the English Spray/Load alternation, which is mentioned in Section 2.5. We intend to capture such systematic alternations using valence changing lexical rules, which apply to specific lexical types and consequently license additional lexical types with alternate valence specifications.
While the motivation for this representation stems from our attempt to engineer an HPSG grammar of Hebrew, we trust that the schema we propose will be easily adaptable to the specific details of other languages, and be usable in other linguistic theories. The applicability of the schema to other languages is currently being put to the test, as we are beginning to implement an HPSG grammar of Modern Standard Arabic. In this phase of our project we are developing the grammar of the two languages in parallel, sharing fragments of the code for areas where the two languages converge, and defining language-specific types for where they diverge. As the two languages are genetically related, we expect that significant parts of the grammar will be shared.