Spurious ambiguity is the phenomenon whereby distinct derivations in grammar may assign the same structural reading, resulting in redundancy in the parse search space and inefficiency in parsing. Understanding the problem depends on identifying the essential mathematical structure of derivations. This is trivial in the case of context free grammar, where the parse structures are ordered trees; in the case of type logical categorial grammar, the parse structures are proof nets. However, with respect to multiplicatives, intrinsic proof nets have not yet been given for displacement calculus, and proof nets for additives, which have applications to polymorphism, are not easy to characterize. In this context we approach here multiplicative-additive spurious ambiguity by means of the proof-theoretic technique of focalization.

In context free grammar (CFG), sequential rewriting derivations exhibit spurious ambiguity: Distinct rewriting derivations may correspond to the same parse structure (tree) and the same structural reading.1 In this case, it is transparent to develop parsing algorithms avoiding spurious ambiguity by reference to parse trees. In categorial grammar (CG), the problem is more subtle. The Cut-free Lambek sequent proof search space is finite, but involves a combinatorial explosion of spuriously ambiguous sequential proofs. This spurious ambiguity in CG can be understood, analogously to CFG, as involving inessential rule reorderings, which we parallelize in underlying geometric parse structures that are (planar) proof nets.

The planarity of Lambek proof nets reflects that the formalism is continuous or concatenative. But the challenge of natural grammar is discontinuity or apparent displacement, whereby there is syntactic/semantic mismatch, or elements appearing out of place. Hence the subsumption of Lambek calculus by displacement calculus D, including intercalation as well as concatenation (Morrill, Valentín, and Fadda 2011).

Proof nets for D must be partially non-planar; steps towards intrinsic correctness criteria for displacement proof nets are made in Fadda (2010) and Moot (2014, 2016). Additive proof nets are considered in Hughes and van Glabbeck (2005) and Abrusci and Maieli (2016). However, even in the case of Lambek calculus, it is not clear that in practice parsing by reference to intrinsic criteria (Morril 2011; Moot and Retoré 2012, Appendix B) is more efficient than parsing by reference to extrinsic criteria of uniform sequent calculus (Miller et al. 1991; Hendriks 1993). In its turn, on the other hand, uniform proof does not extend to product left rules and product unit left rules, nor to additives. The focalization of Andreoli (1992) is a methodology midway between proof nets and uniform proof. Here, we apply the focusing discipline to the parsing as deduction of D with additives.

In Chaudhuri, Miller, and Saurin (2008), multifocusing is defined for unit-free multiplicative-additive linear logic, providing canonical sequent proofs; an eventual goal would be to formulate multifocusing for multiplicative-additive categorial logic and for categorial logic generally. In this respect the present article represents an intermediate step (and includes units, which have linguistic use). Note that Simmons (2012) develops focusing for Lambek calculus with additives, but not for displacement logic, for which we show completeness of focusing here.

The article is structured as follows. In Sections 1.1 and 1.2 we describe spurious ambiguity in context-free grammar and Lambek calculus. In Section 2 we recall the displacement calculus with additives. In Section 3 we contextualize the problem of spurious ambiguity in computational linguistics. In Section 4 we discuss focalization. In Section 5 we present focalization for the displacement calculus with additives. In Section 6 we prove the completeness of focalization for displacement calculus with additives. In Section 7 we exemplify focalization and evaluate it compared with uniform proof. We conclude in Section 8. In Appendix A we prove the auxiliary technical result of Cut-elimination for weak focalization.

### 1.2. Spurious Ambiguity in CG

Lambek calculus L (Lambek 1958) is a logic of strings with the operation + of concatenation. Recall the definitions of types, configurations, and sequents of L in terms of a set of primitive types;2 in Backus Naur form (BNF) notation where is the set of types, the set of configurations, and Seq(L) the set of sequents:Lambek calculus types have the following interpretation in semigroups or monoids:Where A, B, C, D are types and Γ, Δ are configurations, the logical rules of Lambek calculus are as follows:There is completeness when the types are interpreted in free semigroups or monoids Pentus (1993, 1998).
Even among Cut-free proofs there is spurious ambiguity; consider, for example, the sequential derivations of Figure 1. These have the same parallelized parse structure (proof net), given in Figure 2: The structures that represent Lambek proofs without spurious ambiguity are planar graphs that must satisfy certain global and local properties, and are called proof nets; for a survey see Lamarche and Retoré (1996). Proof nets provide a geometric perspective on derivational equivalence. Alternatively, we may identify the same algebraic parse structure (Curry-Howard term):But Lambek calculus is continuous (planarity). A major issue in grammar is discontinuity, namely, syntax/semantics mismatch (e.g., the fact that quantifier phrases occur in situ but take sentential scope; or “gapping” coordination), hence the displacement calculus. This provides a general accommodation of discontinuity in grammar, by contrast with, for example, Combinatory Categorial Grammar (Steedman, 2000), which seeks minimal case by case augmentations of the formalism to deal with quantification, gapping, and so on.3

The basic categorial grammar of Ajdukiewicz and Bar-Hillel is concatenative/continuous/projective, and this feature is reflected in the fact that it is context-free equivalent in weak generative power (Bar-Hillel, Gaifman, and Shamir 1960). The same is true of the logical categorial grammar of Lambek (1958), which is still context free in generative power (Pentus 1992). The main challenge in natural grammar comes from syntax/semantics mismatch or displacement; such non-concatenativity/discontinuity/non-projectivity is treated in mainstream linguistics by overt movement (e.g., the verb-raising of cross-serial dependencies) and covert movement (e.g., the quantifier-raising of quantification). The displacement calculus is a response to this challenge, which preserves all the good design features of Lambek calculus while extending the generative power and capturing “movement” phenomena such as cross-serial dependencies and quantifer-raising.4

In this section we present displacement calculus D, and a displacement logic DA comprising D with additives. Although D is indeed a conservative extension of the Lambek calculus allowing empty antecedents (L*), we think of it not just as an extension of Lambek calculus but as a generalization, because it involves a whole reformulation to deal with discontinuity while conserving L* as a special case.

Displacement calculus is a logic of discontinuous strings—strings punctuated by a separator 1 and subject to operations of append (+; concatenation) and plug (×k; intercalation at the kth separator, counting from the left); see Figure 3. Recall the definition of types and their sorts, configurations and their sorts, and sequents, for the displacement calculus with additives (i and j range over the naturals 0, 1, …):Where Λ is the metasyntactic empty string there is now the BNF definition of configurations:For example, there is the configuration (S1N) ↑2N{N, 1: S1N{N}, S}, 1, N, 1
The figure $A→$ of a type A is defined by:Where Γ is a configuration of sort i and Δ1, …, Δi are configurations, the fold Γ ⊗ 〈Δ1: …: Δi〉 is the result of replacing the successive 1’s in Γ by Δ1, …, Δi respectively. For example ((S1N) ↑2N{N, 1: S1N{N}, S}, 1, N, 1) ⊗ 〈1: N, NS: Λ〉 is (S1N) ↑2N{N, 1: S1N{N}, S}, N, NS, N.
The continuous multiplicatives {∖, •, /, I} of Lambek (1958, 1988), which are defined in relation to concatenation, are the basic means of categorial (sub)categorization. The directional divisions over, /, and under, ∖, are exemplified by assignments such as the: N/CN for the man: N and sings: NS for John sings: S, and loves: (NS)/N for John loves Mary: S. Hence, for the man:And for John sings and John loves Mary:
The discontinuous multiplicatives {↓, ⊙, ↑, J}, the displacement connectives, of Morrill and Valentín (2010) and Morrill, Valentín, and Fadda (2011) are defined in relation to intercalation. When the value of the k subscript is 1 it may be omitted (i.e., it defaults to one). Circumfixation, or extraction, ↑, is exemplified by a discontinuous particle-verb assignment calls + 1 + up: (NS) ↑ N for Mary calls the man up: S:

This section elaborates the bibliographic context of so-called spurious ambiguity, a problem frequently arising in varieties of parsing of different formalisms. The spurious ambiguity that has been discussed in the literature falls in two broad classes: that for categorial parsing and that for dependency parsing.

The literature on spurious ambiguity for categorial parsing is represented by the following.

• •

Hepple (1990) provides an analysis of normal form theorem proving for Cut-free Lambek calculus without product. Two systems are considered: first, a notion of normalization for this implicative Lambek calculus, and second, a constructive calculus generating all and only the normal forms of the first. The latter consists of applying right rules as much as possible and then switching to left rules, which revert to the right phase in the minor premise and which conserve the value active type in the major premise. Hepple shows that the system is sound and complete and that it delivers a unique proof in each Cut-free semantic equivalence class. This amounts to uniform proof (Miller et al. 1991), but was developed independently. Retrospectively, it can be seen as focusing for the implicative fragment of Lambek calculus. It is straightforwardly extendible to product right (Hendriks 1993), but not product left, which requires the deeper understanding of the focusing method.

• •

Eisner (1996) provides a normal form framework for Combinatory Categorial Grammar (CCG) with generalized binary composition rules. CCG is a version of categorial grammar with a small number of combinatory schemata (or a version of CFG with an infinite number of non-terminals) in which the basic directional categorial cancellation rules are extended with phrase structure schemata corresponding to combinators of combinatory logic. Eisner, following Hepple and Morrill (1989), defines a notion of normal form by a restriction on which rule mothers can serve as which daughters of subsequent rules in bottom–up parsing. Eisner notes that his marking of rules has a rough resemblance to the rule regulation of Hendriks (1993) (which is focusing). But whereas the former is Cut-based rule normalization, the latter is Cut-free metarule normalization. Eisner’s method applies to a wide range of harmonic and non-harmonic composition rules, but not to type-lifting rules. It has been suggested by G. Penn (personal communication) that it is not clear whether CCG is actually categorial grammar; at least, it is not logical categorial grammar; in any case, the focusing methodology we use here is still normalization, but represents a much more general discipline than that used by Eisner.

• •

Moortgat and Moot (2011) study versions of proof nets and focusing for so-called Lambek-Grishin calculus. Lambek-Grishin calculus is like Lambek calculus but includes a disjunctive multiplicative connective family as well as the conjunctive multiplicative connective family of the Lambek calculus L, and it is thus multiple-conclusioned. For the case LG considered, suitable structural linear distributivity postulates facilitate the capacity to describe non-context free patterns. By contrast with displacement calculus D, which is single-conclusioned (it is intuitionistic), and which absorbs the structural properties in the sequent syntax (it has no structural postulates), LG has a non-classical term regime that can assign multiple readings to a proof net, and has semantically non-neutral focusing and defocusing rules. Thus it is quite different from the system considered here in several respects. The multiplicative focusing for LG is a straightforward adaptation of that for linear logic and additives are not addressed; importantly, the backward-chaining focused LG search space still requires the linear distributivity postulates (which leave no trace in the terms assigned), whereas the backward chaining D search space has no structural postulates.

The literature on spurious ambiguity for dependency parsing is represented by the following.

• •

Huang, Vogel, and Chen (2011) address the problem of word alignment between pairs of corresponding sentences in (statistical) machine translation. Because this task may be very complex (in fact the search space may grow exponentially), a technique called synchronous parsing is used to constrain the search space. However, this approach exhibits the problem of spurious ambiguity, which that paper elaborates in depth. We do not know at this moment whether this work can bring us useful techniques to deal with spurious ambiguity in the field of logic theorem-proving.

• •

Goldberg and Nivre (2012) and Cohen, Gómez-Rodríguez, and Satta (2012) focus on the problem of spurious ambiguity of a general technique for dependency parsing called transition-based dependency parsing. The spurious ambiguity investigated in that paper arises in transition systems where different sequences of transitions yield the same dependency tree. The framework of dependency grammars is non-logical in its nature and the spurious ambiguity referred here is completely different from the one addressed in our article, and unfortunately not useful to our work.

• •

Hayashi, Kondo, and Matsumoto (2013), which is situated in the field of dependency parsing, proposes sophisticated parsing techniques that may have spurious ambiguity. A method is presented in order to deal with it based on normalization and canonicity of sequences of actions of appropriate transition systems.

All these approaches tackle by a kind of normalization the general phenomena of spurious ambiguity, whereby different but equivalent alternative applications of operations result in the same output. Focalization in particular is applicable to logical grammar, in which parsing is deduction. This turns out to be quite a deep methodology, revealing, for example, not only invertible (“reversible”) rules, which were known to the early proof-theorists, but also dual focusing (“irreversible”) rules, providing a general discipline for logical normalization, which we now elaborate.

In previous sections we have shown the proof machinery of DA. A further sequent rule is missing, the so-called Cut rule, which incorporates in the sequent calculus the notion of (contextualised) transitivity:Note that the Cut rule does not have the property that every type in the premises is a (sub)type of the conclusion (the subformula property), hence it is an obstacle to proof search. Any standard logic should have the Cut rule because this encodes transitivity of the entailment relation, but at the same time a major result of logics presented as sequent calculi is the Cut-elimination theorem, or as it is usually known after Gentzen, Haupsatz. This very important theorem gives as a byproduct the subformula property and decidability in the case of DA.5

### 4.1. Properties of Cut-Free Proof Search

“Reversible” rules are rules in which the conclusions are provable if and only if the premises are provable. The reversible rules of displacement calculus with additives are: /R, ∖R, •L, IL, ↑kR, ↓kR, ⊙kL, JL, &R, and ⊕L. By way of example, consider /R:We can safely reverse sequent (25a) into sequent (25b) because both provability and non-provability are preserved—that is, Γ ⇒ C/B is provable iff Γ, BC is provable. We say that the type-occurrence of C/B in the succedent of (25a) is reversible.6 This means that in the face of alternative reversible rule options a choice can be made arbitrarily and the other possibilities forgotten (don’t care nondeterminism).
Dually, “irreversible” rules are rules in which it is not the case that the conclusions are provable if and only if the premises are provable. The irreversible rules of displacement calculus with additives are: /L, ∖L, •R, IR, ↑kL, ↓kL, ⊙kR, JR, &L and ⊕R. By way of example, consider the /L rule:We cannot assume safely that there exist configurations Γ and Δ(C) that match the antecedent of the end-sequent, and that preserve provability: We do not have that Δ(C/B, Γ) ⇒ D is provable iff Γ ⇒ B and Δ(C) ⇒ D are provable. In this case, we say that the distinguished type-occurrence C/B in the above sequent is irreversible.7 In the face of alternative irreversible rule options, the choice matters and different possibilities must be tried (don’t know nondeterminism).

#### 4.1.1. On Reversible Rules.

Consider the following sequents where two distinguished type-occurrences have been underlined:There are the following two commutative derivation fragments:
• •

Applying the two rules in either order, both (proof) derivations have the same syntactic structure (and the same semantic lambda-term labeling, which we introduce later when we present focused calculus).

• •

The reversible rules •L and ∖Rcommute in the order of application in the proof search. These commutations contribute to spurious ambiguity.

• •

Reversible rules can be applied don’t care nondeterministically. This result was already known by proof-theorists in the 1930s (Gentzen, and others such as Kleene). These rules were called invertible rules.

#### 4.1.2. On Irreversible Rules.

Consider the following:
• •

Dually, both (proof) derivations have the same syntactic structure, (i.e., proof net), as well as the same semantic term labeling codifying the structure of derivations.

• •

The irreversible rules for /(S/(NS))/CN and /(NS)/Ncommute in the order of application in the proof search. These commutations also contribute to spurious ambiguity.

• •

Contrary to the behavior of reversible types, the choice of the active type (the type decomposed) is critical when it is irreversible.

### 4.2. Consequences

• •

A combinatorial explosion in the finite Cut-free proof search space due to the aforementioned rule commutativity.

• •

The solution: proof nets. But proof nets for categorial logic in general are not fully understood, for example, in the case of units and additive connectives.

• •

This problem is exacerbated in displacement calculus because there are more connectives and hence more rules giving rise to spurious ambiguities.

• •

A good partial solution: the discipline of focalization.

### 4.3. Toward a (Considerable) Elimination of Spurious Ambiguity: Focalization

Reversible rules commute in their order of application and their key property is reversability. Dually, irreversible rules also commute and their key property (completely unknown to the early proof-theorists8) is so-called focalization, which was defined by Andreoli in Andreoli (1992).

### 4.4. The Discipline of Focalization

Given a sequent, reversible rules are applied in a don’t care nondeterministic fashion until no longer possible. When there are no occurrences of reversible types, one chooses an irreversible type don’t know nondeterministically as active type-occurrence; we say that this type occurrence is in focus, or that it is focalized; and one applies proof search to its subtypes while these remain irreversible. When one finds a reversible type, reversible rules are applied in a don’t care nondeterministic fashion again until no longer possible, when another irreversible type is chosen, and so on.

#### 4.4.1. A Non-focalized Proof: A “Gymkhana in the Proof-search”.

Let us consider the following provable sequent with types Q = S/(NS), and TV = (NS)/N.Sequent (27) has the algebraic Curry-Howard term:Consider the proof fragment:This fragment of proof concludes into a correct proof derivation. But crucially, the discipline of focalization is not applied resulting—in the words of the father of linear logic—in a gymkhana! (In the sense of jumping back and forth.) The foci on irreversible types are not preserved and alternating irreversible types are chosen. Concretely, the underlines show how the active irreversible type in the right premise is not a subtype of the active irreversible type in the endsequent. The discipline of Andreoli’s focalization consists of maintaining the focus on the irreversible subtypes, hence reducing the search space and reducing spurious ambiguity. This rests on the focusing property: that if an irreversible rule instance leads to a proof, it leads to a proof in which the subtypes of the active type are the active types of the premises, if they are irreversible. This is the remarkable proof-theoretic discovery of Andreoli, which fortunately can be applied in the present case as this article explains. The discipline of focalization reduces dramatically the combinatorial explosion of categorial proof-search.

### 4.5. A Last Ingredient in the Focalization Discipline

• •

What happens with literal types? Can they be considered reversible or irreversible? What happens then in the focalized proof-search paradigm?

• •

The answers: Literals can be assigned what is called in the literature a reversible or irreversible bias in any manner according to which they belong to the set of reversible or irreversible types. (As we state later we leave open the question of which biases may be favorable.)

Depending on the context, we will say compound reversible or irreversible types, or atomic reversible or irreversible types.

In focalization, situated (in the antecedent of a sequent, input: ; in the succedent of a sequent, output: °) compound types are classified as of reversible or irreversible polarity according to whether their associated rule is reversible or not; situated atoms obviously do not have associated rules, but we can extend the concept of polarity to them, overloading the terms reversible and irreversible, applying the previously mentioned concept of bias. We define in Example (31) a BNF grammar for arbitrary types (observe that atomic types are classified as either At or At°):P and Q in Example (31) denote reversible types (including atomic types) occurring in input and output position, respectively. Dually, if P and Q occur in the output and input position, respectively, then they are said to occur with irreversible polarity. In the atomic case, the same terms (reversible and irreversible) are also used. The table in Example (32) summarizes the notational convention on arbitrary (atomic or compound) polarized formulas P and Q:Notice that in Example (32) arbitrary polarized types exhibit a kind of De Morgan duality: If reversible output types Q occur in input position, then they are irreversible; whereas if reversible output types P occur in input position, then they are irreversible.

In order to prove that DA is complete with respect to focalization, we define a logic DAFoc with the following features: (a) the set of configurations is extended to the set box, (b) the set of sequents Seq(DA) is extended to the set Seq(DAFoc), (c) a new set of logical rules. The set of configurations box contains , and in addition it contains boxed configurations, by which we understand configurations where a unique irreversible type-occurrence is decorated with a box, thus: . The set of sequents Seq(DAFoc) includes DA sequents with possibly a box in the sequent. We have then Seq(DA) ⫋ Seq(DAFoc). Sequents of Seq(DAFoc) can contain at most one boxed type-occurrence. The meaning of such a box is to mark in the proof-search an irreversible (possibly atomic) type-occurrence either in the antecedent or in the succedent of a sequent. We will say that such a sequent is focalized.

We will use judgements foc and rev on DAFoc-sequents. Where Seq(DAFoc), foc means that contains a boxed type occurrence, and rev means that there is a complex reversible type occurrence. It follows that Constraint (33) can be judged as foc ∧ ¬rev. The judgment ¬ foc means that is not focalized (and so may contain reversible type-occurrences).

The rules for focused inference are given in Figures 412, where sequents are accompanied by judgments: focalized or not focalized and reversible or not reversible. The constraint (33) is key in determining the judgments; in the figures showing irreversible additive or multiplicative rules, when the premise is not focalized only the corresponding subtype can be reversible. This system, which is Cut-free, is what we call strongly focalized and induces maximal alternating phases of reversible and irreversible rule application. Pseudo-code for a recursive algorithm of strongly focalized proof search is as follows:

The top level call to determine whether a sequent S is provable is prove(S). The routine prove(S) calls the routine prove_rev_lst with actual parameter the unitary list [S]. The routine prove_rev_lst then applies reversible rules to its list of sequents Ls in a don’t care nondeterministic manner until none of the sequents contain any reversible type (i.e., it closes Ls under reversible rules). Then prove_irrev_lst is called on the list of sequents. This calls prove_irrev(S′) for focusings S′ of each sequent, and if some focusing of each sequent is provable the result true is returned, otherwise false is returned. The procedure prove_irrev applies focusing rules and recurses back on prove_rev_lst and prove_irrev_lst to determine provability for the given focusing.

In order to prove completeness of (strong) focalization we invoke also an intermediate weakly focalized system. In all we shall be dealing with three systems: the displacement calculus with additives DA with sequents notated Δ ⇒ A, the weakly focalized displacement calculus with additives DAFoc with sequents notated Δ ⇒wA, and the strongly focalized displacement calculus with additives DAFoc with sequents notated Δ⇒A. Sequents of both DAfoc and DAFoc may contain at most one focalized formula. When a DAfoc sequent is notated Δ ⇒wAfocalized, it means that the sequent possibly contains a (unique) focalized formula. Otherwise, Δ ⇒wA means that the sequent does not contain a focus. In DAfocConstraint (33) is not imposed. Thus, whereas strong focalization imposes maximal alternating phases of reversible and irreversible rule application, weak focalization does not impose this maximality. In this section we prove the strong focalization property for the displacement calculus with additives DA, that is, strong focalization is complete.

The focalization property for Linear Logic was defined by Andreoli (1992). In this article we follow the proof idea from Laurent (2004), which we adapt to the intuitionistic non-commutative case DA with twin multiplicative modes of combination, the continuous (concatenation), and the discontinuous (intercalation) products. The proof relies heavily on the Cut-elimination property for weakly focalized DA, which is proved in Appendix A. In our presentation of focalization we have avoided the react rules of Andreoli (1992) and Chaudhuri (2006), and use instead our simpler box notation suitable for non-commutativity.

DAFoc is a subsystem of DAfoc. DAfoc has the focusing rules foc and Cut rules p-Cut1, p-Cut2, n-Cut1, and n-Cut29 shown in Equation (34), and the reversible and irreversible rules displayed in the figures, which are read as allowing in irreversible rules the occurrence of reversible types, and in reversible rules as allowing arbitrary sequents possibly with a focalized type; when ◊focalized appears in both premise and conclusion, it means that they are either both focused or both unfocused:
With respect to bias assignment, consider the provable sequent n, ns, sqq. The biases nAt° and q, and sAt mean that we have the identity axioms n , q, and s.

### 6.1. Embedding of DA into DAfoc

The identity rule Id, which applies not just to atomic types (like Id), but to all types is easy to prove in both DA and DAfoc, but the same is not the case for DAFoc. This is the reason to consider what we have called weak focalization, which helps us to prove smoothly this crucial property for the proof of strong focalization.

Theorem 1 (Embedding of DA into DAfoc)

For any configuration Δ and type A, we have that if ⊢ Δ ⇒ A then ⊢ Δ ⇒wA.

Proof. We proceed by induction on the length of the derivation of DA proofs. In the following lines, we apply the induction hypothesis (i.h.) for each premise of DA rules (with the exception of the identity rule and the right rules of units):

Left unit rules apply as in the case of DA.

### 6.2. Embedding of DAfoc into DAFoc

Theorem 2
Let be a DAFoc sequent. The following holds:

Proof. We proceed by induction on the size of DAfoc-provable sequents, namely on | |.10 Since DAFoc is Cut-free (see Appendix A) we consider Cut-free DAfoc proofs of .

• –

Case | | = 0. This corresponds to the axiom case, which is the same for both calculi DAfoc and DAFoc — see Equation (35).

• –

Suppose | | > 0. Because | | > 0, does not correspond to an axiom. Hence is derivable with at least one logical rule. Otherwise, the size of would be equal to 0. The last rule ⋆ can be either a logical rule or a foc rule (keep in mind we are considering Cut-free DAfoc proofs!). We have two cases:

• a)

If ⋆ is logical, because is supposed to belong to Seq(DAFoc), it follows that its premises (possibly only one premise) belong also to Seq(DAFoc). The premises have size strictly less than . Therefore, we can safely apply i.h., whence we conclude.

• b)

Observe that | ′| = | |, because the size |⋅| function counts connectives, and not boxes of focalized type-occurrences. Because | ′| > 0, it follows that in the proof of ′ there must be at least one logical rule. ′ can have the following judgments:We consider the two cases:
• – foc ∧ ¬ rev. Because is focalized and there is no reversible type occurrence, the last rule of ′ must correspond either to a multiplicative or additive irreversible rule. The premises (possibly only one premise) have size strictly less than . We can then safely apply the induction hypothesis (i.h.), which gives us DAFoc provable premises to which we can apply the ⋆ rule. Whence DAfoc ′, and hence DAFoc .

• – focrev.

The size of ′ equals that of the end-sequent , and moreover the premise is focrev, which does not belong to Seq(DAfoc). Clearly, we cannot apply i.h.. What can we do?

Thanks to the flexibility of DAfoc we can overcome this difficulty. ′ contains at least one reversible formula. We see three cases that are illustrative of the situation:We consider these in turn:
• a)
We have by Id that . We apply to this sequent the reversible ⊙k left rule, whence . In case (37a), we have the following proof in DAfoc:To this DAfoc proof we apply Cut-elimination and we get the Cut-free DAfoc end-sequent . We have < . We can apply then i.h. and we derive the provable DAFoc sequent , to which we can apply the left ⊙k rule. We have obtained .
• b)
In the same way, we have that . Thus, in case (37b), we have the following proof in DAfoc:As before, we apply Cut-elimination to the above proof. We get the Cut-free DAfoc end-sequent . It has size less than . We can apply i.h. and we obtain the DAFoc provable sequent , to which we apply the ↑k right rule.
• c)
We haveby applying the foc rule and the invertibility of &R we obtain the provable DAfoc sequents and . These sequents have smaller size than . The aforementioned sequents have a Cut-free proof in DAfoc. We apply i.h. and we obtain and . We apply the & right rule in DAFoc, and we obtain .
Let ProvΣ(C) be the class of Σ sequents provable in calculus C.

Theorem 3

Strong focalization is complete.

Proof. Observe that in particular ProvSeq(DA)(DAFoc) = ProvSeq(DA)(DAfoc). Because, by Theorem 1 ProvSeq(DA)(DAFoc) = Prov(DA), we have that ProvSeq(DA)(DAFoc) = Prov(DA).

CatLog version f1.2, CatLog1 (Morrill 2012), is a parser/theorem-prover using uniform proof and count-invariance for multiplicatives. CatLog version k2, CatLog3 (Morrill 2017), is a parser/theorem-prover using focusing, as expounded here, and count-invariance for multiplicatives, additives, brackets, and exponentials (Kuznetsov, Morrill, and Valentín 2017). To evaluate the performance of uniform proof and focusing, we created a system version clock3f1.2 made by substituting the theorem-proving engine of CatLog1 into the theorem-proving environment of CatLog3 so that count-invariance and other factors were kept constant while the uniform and focusing theorem-proving engines were compared.

We performed the Montague test (Morrill and Valentín 2016) on the two systems, that is, the task of providing a computational grammar of the Montague fragment. In particular, we ran exhaustive parsing of the minicorpus given in Figure 13. The Montague lexicon was as in Figure 14. The tests were executed in XGP Prolog on a MacBook Air. The running times in seconds for exhaustive parsing were as follows:The tests for the minicorpus show that the focusing parsing/proof-search runs in half the time of the uniform parsing/proof-search.

We could have also made comparison with proof net parsing/theorem proving (Moot 2016), but our proposal includes additives for which proof nets are still an open question, and not just the displacement calculus multiplicatives, and the point of focalization is that it is a general methodology extendible to, for example, exponentials and other modalities also, for which proof nets are also still under investigation. That is, our focalization approach is scaleable in a way that proof nets currently are not, and in this sense comparison with proof nets is not quite appropriate.

We have claimed that just as the parse structures of context free grammar are ordered trees, the parse structures of categorial grammar are proof nets. Thus, just as we think of context free algorithms as finding ordered trees, bottom–up, top–down, and so forth, we can think of categorial algorithms as finding proof nets. The complication is that proof nets are much more sublime objects than ordered trees: They embody not only syntactic coherence but also semantic coherence. Focalization is a step on the way to eliminating spurious ambiguity by building such proof nets systematically. A further step on the way, eliminating all spurious ambiguity, would be multifocusing. This remains a topic for future study. Another topic for further study in focusing is the question of which assignments of bias are favorable for the processing of given lexica/corpora.

Alternatively, for context free grammar one can perform chart parsing, or tabularization, and at least for the basic case of Lambek calculus suitable notions of proof net also support tabularization (Morrill, 1996; de Groote, 1999; Pentus, 2010; Kanovich et al., 2017). This also remains a topic for future study.

For the time being, however, we hope to have motivated the relevance of focalization to categorial parsing as deduction in relation to the DA categorial logic fragment, which leads naturally to the program of focalization of extensions of this fragment with connectives such as exponentials.

We thank anonymous Computational Linguistics referees for comments and suggestions that have improved this article. This research was supported by grant TIN2017-89244-R from MINECO (Ministerio de Economia, Industria y Competitividad).

1

This paper is a revised and expanded version of Morrill and Valentín (2015).

2

The original Lambek calculus did not include the product unit and had a non-empty antecedent condition (“Lambek’s restriction”). The displacement calculus used in the present article conservatively extends the Lambek calculus without Lambek’s restriction and with product units.

3

Indeed, Sorokin (2013) and Wijnholds (2014) show that Displacement calculus recognizes the class of well-nested multiple context free languages (Kanazawa 2009); but Combinatory Categorial Grammar recognizes only the class of tree adjoining languages (Joshi, Vijay-Shanker, and Weir 1991).

4

It is known that displacement calculus (without additives) generates a well-known class of mildly context free languages: the well-nested multiple context free languages (Sorokin 2013; Wijnholds 2014). At the time of writing, only this and other lower bounds are known; tight upper bounds on the weak generative capacity of displacement calculus constitute an open question.

5

This is because in Cut-free backward-chaining proof search for a given goal sequent a finite number of rules can be applied backwards in only a finite number of ways to generate subgoals at each step, and these subgoals have lower complexity (fewer connectives) than the goal matched; hence the proof search space is finite.

6

Other terms found in the literature are invertible, asynchronous, or negative.

7

Other terms found in the literature are non-invertible, synchronous, or positive.

8

Unknown even to the inventor of Linear Logic, J.-Y. Girard.

9

If it is convenient, we may drop the subscripts.

10

For a given type A, the size of A, |A|, is the number of connectives in A. By recursion on configurations we have: . Moreover, we have: .

11

The size of |A| is the number of connectives appearing in A.

Abrusci
,
Vito Michele
and
Roberto
Maieli
.
2016
.
Cyclic multiplicative-additive proof nets of linear logic with an application to language parsing
. In
Proceedings of Formal Grammar: 20th and 21st International Conferences, FG 2015
, pages
43
59
,
Bozen
.
Andreoli
,
J. M.
1992
.
Logic programming with focusing in linear logic
.
Journal of Logic and Computation
,
2
(
3
):
297
347
.
Bar-Hillel
,
Y.
,
C.
Gaifman
, and
E.
Shamir
.
1960
.
On categorial and phrase structure grammars
.
Bulletin of the Research Council of Israel
,
9F
:
1
16
.
Bayer
,
Samuel
.
1996
.
The coordination of unlike categories
.
Language
,
72
(
3
):
579
616
.
Chaudhuri
,
Kaustuv
.
2006
.
The Focused Inverse Method for Linear Logic
.
Ph.D. thesis
,
Carnegie Mellon University
,
Pittsburgh, PA, USA
.
AAI3248489
.
Chaudhuri
,
Kaustuv
,
Dale
Miller
, and
Alexis
Saurin
.
2008
.
Canonical sequent proofs via multi-focusing
. In
Fifth IFIP International Conference on Theoretical Computer Science - TCS 2008, IFIP 20th World Computer Congress
, pages
383
396
,
Milano
.
Cohen
,
Shay B.
,
Carlos
Gómez-Rodríguez
, and
Giorgio
Satta
.
2012
.
Elimination of spurious ambiguity in transition-based dependency parsing
.
CoRR
,
abs/1206.6735
.
Eisner
,
Jason
.
1996
.
Efficient normal-form parsing for combinatory categorial grammar
. In
Proceedings of the 34th Annual Meeting of the Association for Computational Linguistics
, pages
79
86
,
Stroudsburg, PA
.
,
Mario
.
2010
.
Geometry of Grammar: Exercises in Lambek Style
.
Ph.D. thesis
,
Universitat Politècnica de Catalunya
,
Barcelona
.
Goldberg
,
Yoav
and
Joakim
Nivre
.
2012
.
A dynamic oracle for arc-eager dependency parsing
. In
COLING 2012, 24th International Conference on Computational Linguistics, Proceedings of the Conference: Technical Papers
, pages
959
976
,
Mumbai
.
de Groote
Philippe
.
1999
.
A dynamic programming approach to categorial deduction
. In
Proceedings of the 16th International Conference on Automated Deduction (CADE)
, pages
1
15
,
Trento
.
Hayashi
,
Katsuhiko
,
Shuhei
Kondo
, and
Yuji
Matsumoto
.
2013
.
Efficient stacked dependency parsing by forest reranking
.
TACL
,
1
:
139
150
.
Hendriks
,
H.
1993
.
Studied Flexibility. Categories and Types in Syntax and Semantics
.
Ph.D. thesis
,
Universiteit van Amsterdam, ILLC
,
Amsterdam
.
Hepple
,
Mark
.
1990
.
Normal form theorem proving for the Lambek calculus
. In
Proceedings of COLING
, pages
173
178
,
Stockholm
.
Hepple
,
Mark
and
Glyn
Morrill
.
1989
.
Parsing and derivational equivalence
. In
Proceedings of the Fourth Conference of the European Chapter of the Association for Computational Linguistics
, pages
10
18
,
Stroudsburg, PA
.
Huang
,
Shujian
,
Stephan
Vogel
, and
Jiajun
Chen
.
2011
.
Dealing with spurious ambiguity in learning ITG-based word alignment
. In
Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: Short Papers - Volume 2
, pages
379
383
,
Stroudsburg, PA
.
Hughes
,
Dominic J. D.
and
Rob J.
van Glabbeck
.
2005
.
Proof nets for unit-free multiplicative-additive linear logic
.
ACM Transactions on Computational Logic (TOCL)
,
6
(
4
):
784
842
.
Joshi
,
A. K.
,
K.
Vijay-Shanker
, and
D.
Weir
.
1991
.
The convergence of mildly context-sensitive grammar formalisms
. In
P.
Sells
,
S. M.
Shieber
, and
T.
Wasow
, editors,
Foundational Issues in Natural Language Processing
,
The MIT Press
,
Cambridge, MA
, pages
31
81
.
Kanazawa
,
M.
1992
.
The Lambek calculus enriched with additional connectives
.
Journal of Logic, Language and Information
,
1
:
141
171
.
Kanazawa
,
Makoto
.
2009
.
The pumping lemma for well-nested multiple context-free languages
. In
Developments in Language Theory: 13th International Conference, DLT 2009
, pages
312
325
,
Stuttgart
.
Kanovich
,
Max
,
Stepan
Kuznetsov
,
Glyn
Morrill
, and
Andreu
Scedrov
.
2017
.
A polynomial-time algorithm for the Lambek calculus with brackets of bounded order
. In
Proceedings of the 2nd International Conference on Formal Structures for Computation and Deduction (FSCD 2017)
, pages
22.1
22.17
,
Leibniz
.
Kuznetsov
,
Stepan
,
Glyn
Morrill
, and
Oriol
Valentín
.
2017
.
Count-invariance including exponentials
. In
15th Meeting on Mathematics of Language
, pages
128
139
,
London
.
Lamarche
,
F.
and
C.
Retoré
.
1996
.
Proof nets for the Lambek calculus - an overview
. In
Proofs and Linguistic Categories - Applications of Logic to the Analysis and Implementation of Natural Language
, pages
241
626
,
Bologna
.
Lambek
,
Joachim
.
1958
.
The mathematics of sentence structure
.
American Mathematical Monthly
,
65
:
154
170
.
Lambek
,
Joachim
.
1961
.
On the calculus of syntactic types
. In
Roman
Jakobson
, editor,
Structure of Language and its Mathematical Aspects, Proceedings of the Symposia in Applied Mathematics XII
, pages
166
178
,
Providence, RI
.
Lambek
,
Joachim
.
1988
.
Categorial and categorical grammars
. In
Richard T.
Oehrle
,
Emmon
Bach
, and
Deidre
Wheeler
, editors,
Categorial Grammars and Natural Language Structures
,
volume 32
of
Studies in Linguistics and Philosophy
,
D. Reidel
,
Dordrecht
, pages
297
317
.
Laurent
,
Olivier
.
2004
.
A proof of the focalization property of linear logic
.
Unpublished manuscript
,
CNRS - Université Paris VII
.
Miller
,
D.
,
G.
,
F.
Pfenning
, and
A.
Scedrov
.
1991
.
Uniform proofs as a foundation for logic programming
.
Annals of Pure and Applied Logic
,
51
:
125
157
.
Moortgat
,
Michael
and
Richard
Moot
.
2011
.
Proof nets for the Lambek-Grishin calculus
.
CoRR
,
abs/1112.6384
.
Moot
,
Richard
.
2014
.
Extended Lambek calculi and first-order linear logic
. In
,
Claudia
,
Bob
Coeke
,
Michael
Moortgat
, and
Philip
Scott
, editors,
Categories and Types in Logic, Language and Physics: Essays Dedicated to Jim Lambek on the Occasion of His 90th Birthday
,
volume 8222
of
LNCS, FoLLI Publications in Logic, Language and Information
,
Springer
,
Berlin
, pages
297
330
.
Moot
,
Richard
.
2016
.
Proof nets for the displacement calculus
. In
Annie
Foret
,
Glyn
Morrill
,
Reinhard
Muskens
,
Rainer
Osswald
, and
Sylvain
Pogodalla
, editors,
Formal Grammar: 20th and 21st International Conferences, FG 2015
, pages
273
289
,
Bozen
.
Moot
,
Richard
and
Christian
Retoré
.
2012
.
The Logic of Categorial Grammars: A Deductive Account of Natural Language Syntax and Semantics
.
Springer
,
Heidelberg
.
Morrill
,
Glyn
.
1990
.
Grammar and Logical Types
. In
Martin
,
Stockhof
, and
Leen
Torenvliet
, editors,
Proceedings of the Seventh Colloquium
, pages
429
450
,
Amsterdam
.
Morrill
,
Glyn
.
1996
.
Memoisation of categorial proof nets: Parallelism in categorial processing
. In
Proofs and Linguistic Categories, Proceedings 1996 Roma Workshop
, pages
157
169
,
Bologna
.
Morrill
,
Glyn
.
2000
.
Incremental Processing and Acceptability
.
Computational Linguistics
,
26
(3)
:
319
338
.
Morrill
,
Glyn
.
2012
.
CatLog: A categorial parser/theorem-prover
. In
LACL 2012 System Demonstrations
, pages
13
16
,
Nantes
.
Morrill
,
Glyn
.
2017
.
Parsing logical grammar: Catlog3
. In
Proceedings of the Workshop on Logic and Algorithms in Computational Linguistics 2017
, pages
107
131
,
Stockholm University
.
Morrill
,
Glyn
and
Oriol
Valentín
.
2010
.
Displacement calculus
.
Linguistic Analysis
,
36
(
1–4
):
167
192
.
Special issue Festschrift for Joachim Lambek
.
Morrill
,
Glyn
and
Oriol
Valentín
.
2015
.
Multiplicative-additive focusing for parsing as deduction
. In
First International Workshop on Focusing, Workshop Affiliated with LPAR 2015
, pages
29
54
,
Suva
.
Morrill
,
Glyn
and
Oriol
Valentín
.
2016
.
Computational coverage of type logical grammar: The Montague test
. In
C.
Piñón
, editor,
Empirical Issues in Syntax and Semantics
,
volume 11
.
Colloque de Syntaxe et Sémantique Ĺ Paris (CSSP)
,
Paris
, pages
141
170
.
Morrill
,
Glyn
,
Oriol
Valentín
, and
Mario
.
2011
.
The displacement calculus
.
Journal of Logic, Language and Information
,
20
(
1
):
1
48
.
Morrill
,
Glyn V.
2011
.
Categorial Grammar: Logical Syntax, Semantics, and Processing
.
Oxford University Press
.
Pentus
,
M.
1992
.
Lambek grammars are context-free
.
Technical report
,
Department of Mathematics and Logic, Steklov Math. Institute
,
Moskow
.
Also published as ILLC Report, University of Amsterdam, 1993, and in Proceedings Eighth Annual IEEE Symposium on Logic in Computer Science. Montreal, 1993
.
Pentus
,
M.
1998
.
Free monoid completeness of the Lambek calculus allowing empty premises
. In
J. M.
Larrazabal
,
D.
Lascar
, and
G.
Mints
, editors,
Logic Colloquium ’96
,
volume 12
.
Springer
, pages
171
209
.
Pentus
,
Mati
.
1993
.
Lambek calculus is L-complete, ILLC Report, University of Amsterdam
.
Shortened version published as Language completeness of the Lambek calculus
,
Proceedings of the Ninth Annual IEEE Symposium on Logic in Computer Science
, pages
487
496
,
Paris
.
Pentus
,
Mati
.
2010
.
A polynomial-time algorithm for Lambek grammars of bounded order
.
Linguistic Analysis
,
36
(
1–4
):
441
472
. Special issue Festschrift for Joachim Lambek.
Simmons
,
Robert J.
2012
.
Substructural Logical Specifications
.
Ph.D. thesis
,
Carnegie Mellon University
,
Pittsburgh
.
Sorokin
,
Alexey
.
2013
.
Normal forms for multiple context-free languages and displacement Lambek grammars
. In
Sergei
Artemov
and
Anil
Nerode
, editors,
Logical Foundations of Computer Science: International Symposium
, pages
319
334
,
San Diego, CA
.
Steedman
,
Mark
.
2000
.
The Syntactic Process
,
MIT Press
,
Cambridge, MA
.
Valentín
,
Oriol
.
2016
.
Models for the displacement calculus
. In
Formal Grammar: 20th and 21st International Conferences
, pages
147
163
,
Bozen
.
Wijnholds
,
Gijs Jasper
.
2014
.
Conversions between D and MCFG: Logical characterizations of the mildly context-sensitive languages
.
Computational Linguistics in the Netherlands Journal
,
4
:
137
148
.
This is an open-access article distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits you to copy and redistribute in any medium or format, for non-commercial use only, provided that the original work is not remixed, transformed, or built upon, and that appropriate credit to the original source is given. For a full description of the license, please visit https://creativecommons.org/licenses/by-nc-nd/4.0/legalcode.