1. Introduction
1.1. Conversational Implicatures and the Generalized/Particularized Distinction
The term conversational implicature was proposed by
Grice (
1975) to explain the meaning gap between logical meaning (i.e., the syntactically and semantically encoded meaning of a sentence) and the pragmatic meaning (i.e., the speaker meaning that is inferred in a particular communicative situation). For example, in (1), the sentence meaning of the answer could be paraphrased as (1a) and it could have one of the pragmatic meanings of (1b).
1. — John, are you ok?
— I’m freezing, Mary.
a. John is very cold.
b. John wants Mary to turn down the air conditioner or turn up the heat.
From a theoretical perspective, Grice recognized two types of conversational implicatures. On the one hand, (1) is a case of a particularized implicature (PCI) because contextual assumptions have a crucial role in sentence comprehension. In order to infer (1b), Mary needs to know if the air conditioner is high or the heat is low or if it is winter or summer. It could also be the case that they were walking in the street and Mary had an extra sweater and John wants Mary to give it to him, or that the temperature is not the problem, but John is sick. Thus, knowing facts about the situation is decisive to derive the actual meaning of the answer in (1).
On the other hand, there are implicatures where specific lexical or syntactic structures work as triggers for the derivation of an implicature (e.g., quantifiers, numbers, conjunctions). Generalized implicatures (GCI), such as the one in (2), are relatively independent of context and can be generally derived without recurring to it.
2. — Did the packages arrive?
— Some of them did.
a. At least one package arrived (and it could be all of them).
b. Some, but not all packages arrived.
Example (2) is a subtype of GCI called scalar implicatures (
Horn 1984). In this example,
some functions as an implicature trigger and takes the pragmatic meaning of
some but not all based on the assumption that the speaker is cooperating and, therefore, if she had more information (i.e., that
all the packages arrived), she would provide it. Quantifiers such as
some and
all can be considered to form a scale, in which
all is more informative or stronger than
some. If it is assumed that the conversational maxim of quantity is being held (that is to say, that the speaker is being as informative as possible), when a member of the scale is asserted, the higher members of the scale are negated because they are more informative, despite the fact that the logical meaning of the trigger does not codify that negation (
some is logically implied by
all, it does not negate it). Implicature triggers can also appear in contexts where the GCI derived meaning is not relevant. For instance, in (3), the
only some meaning of (3b) is not the most relevant for this context.
3. — What happened?
— Some packages arrived.
a. At least one package arrived (equivalent to ‘packages arrived’).
b. Some, but not all packages arrived.
In this case, the idea that
an indefinite number of packages has arrived (the logical meaning paraphrased in (3a)) is more relevant for this context. These cases are called lower-bound contexts in opposition to contexts like (2), which are labelled as upper-bound (
Horn 1984;
Breheny et al. 2006). Thus, in lower-bound contexts, the logical meaning of the GCI trigger will be the preferred one, whereas in upper-bound contexts, the pragmatic one will be the most relevant. It is noteworthy that how and why a context acts as upper or lower bound is still an open question (
Dupuy et al. 2016).
The idea that GCIs are derived in the presence of specific triggers and relatively independently to context led to the hypothesis that this type of implicature could be derived by default (
Horn 1984;
Levinson 2000). Thus, the defaultness approach to GCIs states that in the presence of one of these triggers, the implicated meaning is derived rapidly. If the trigger appears in a lower-bound context, the implicated meaning is also derived, but it is cancelled in a later stage, when the context is integrated to the interpretation process in order to retrieve the logical meaning. Under this approach, GCIs and PCIs differ in their derivation mechanisms; the first undergoes a default process that is triggered by specific words or structures, whereas the latter relies on a non-default process highly constrained by contextual assumptions. There is also another approach that states that the context is integrated early to the derivation (
Sperber and Wilson 1995;
Carston 1998), and that there are no substantial differences between the derivation of GCIs and PCIs. In the presence of any kind of linguistic stimuli, the context is integrated early and any access to implicatures is driven by the contextual assumptions. In this case, scalar implicatures will only be derived if the context is upper-bound, whereas in lower-bound contexts, they will not. Nevertheless, it is important to notice that not only the time course of the context integration is important for the distinction between PCIs and GCIs, but also the role that context has in the derivation of each kind of implicature. PCIs are characterized as being a case of vagueness, because the set of possible meanings is, at least initially, indefinite. Conversely, GCIs tend to have one closed set of possible meanings available, and could be treated as a case of ambiguity resolution.
1.2. Experimental Studies of Conversational Implicatures
There have been several experimental approaches to conversational implicatures and, more specifically, to scalar implicatures (cfr.
Noveck and Reboul 2008 or
Breheny 2019 for a review). The majority of these studies focused on the nature of GCI processing, leaving aside the direct distinction with PCIs, and the main objective of these experiments was to find indicators of the processing cost of GCI triggers in the two contexts presented above.
Two hypotheses linked with the models previously outlined were proposed:
1. If GCIs are processed by default, the processing cost of GCI triggers in upper-bound contexts should be lower than in lower-bound contexts, because the default approach presumes that contextual assumptions are computed later in the derivation process, and that if the default meaning is the preferred one (upper-bound contexts), these assumptions will not add extra effort to the derivation. On the contrary, if the preferred meaning is the logical one (lower-bound contexts), there will be an extra step of implicature cancellation in which contextual assumptions are used. Thus, in this case, the processing cost will be higher.
2. If GCIs are not processed by default, the processing cost of GCI triggers in upper-bound contexts should be higher than in lower-bound contexts, as the guided-by-context approach states that the derivation of the GCI is effortful and it is only made if the context supports it. Because there is no derivation of implicatures in lower-bound contexts and contextual assumptions are not used, there is no extra processing cost.
In the particular case of the scalar implicature with
some, there are two studies with controversial findings. On the one side, in
Breheny et al. (
2006), the scalar implicature was tested by assessing the reading times for the quantifier
some in the two contexts mentioned before (lower-bound and upper-bound). They used reported speech stimuli, as can be seen in (4):
4. Upper-bound: Mary asked John whether he intended to host all of his relatives in his tiny apartment. John replied that he intended to host some of his relatives. The rest would stay in a nearby hotel
Lower-bound: Mary was surprised to see John cleaning his apartment and she asked the reason why. John told her that he intended to host some of his relatives. The rest would stay in a nearby hotel.
Their results showed higher reading times at the some segment in upper-bound contexts, in line with the guided-by-context approach. Furthermore, they found that the reading times at the rest segment were lower in the upper-bound contexts, and used this as a way to assess whether or not the implicature was derived.
On the other side, the replication of this study by
Politzer-Ahles and Fiorentino (
2013) showed that the issue is still open to debate. They presented the same task with changes in the stimuli, arguing that the repetition of the noun that is being quantified could be the cause of the delay that
Breheny et al. (
2006) observed in their experiment. Thus, the repetition of relatives in the example above, where a pronoun would be expected, could be affecting reading times. In their study, the only difference between the two conditions was one word:
any in lower-bound and
all in upper-bound contexts:
5. Upper-bound: Mary was preparing to throw a party for John’s relatives. She asked John whether all of them were staying in his apartment. John said that some of them were. He added that the rest would be staying in a hotel.
Lower-bound: Mary was preparing to throw a party for John’s relatives. She asked John whether any of them were staying in his apartment. John said that some of them were. He added that the rest would be staying in a hotel.
With these stimuli, their results showed no significant differences in the reading times for the segments with
some. They used the same method as
Breheny et al. (
2006) for assessing the realization of the implicature (i.e., measuring reading times for
the rest segment) and they obtained the same results, with a higher latency in lower-bound contexts. Therefore, their data suggest a context sensitive, but effortless derivation of scalar implicatures.
However, there are two issues that were not addressed. Firstly, there are not sufficient experimental data to give a clear definition about the nature of the distinction between lower and upper bound contexts. Both
Breheny et al. (
2006) and
Politzer-Ahles and Fiorentino (
2013) tested that distinction indirectly with the reading time of
the rest segment of the stimuli, and linked the higher latencies to the fact that in lower-bound contexts, the reference to
the rest of the quantified noun phrase is not available. Secondly, there is no clear explanation of which factors of the context are actually constraining the reading times for the GCI trigger. As we describe in the next section, the use of an indirect speech structure could be imposing restrictions on the implicature processing that have not been tested yet.
1.3. The Present Study
Our study had three goals. The first goal was to test if the upper-bound contexts of our experiment make the reading of the GCI trigger some as only some more relevant and if the lower-bound contexts have the opposite effect. To do so, we designed a judgment task to contrast the acceptability of the combination of both contexts with some and only some. We hypothesize that upper-bound contexts will be acceptable with some and only some, whereas lower-bound contexts will only be acceptable with some and not with only some, because in the latter, the only some reading will be less relevant.
The second goal was to rule out the effects that indirect speech structures might impose. For that purpose, we used a self-paced reading task, as in the studies of
Breheny et al. (
2006) and
Politzer-Ahles and Fiorentino (
2013), but with some changes in the stimuli. Both of these studies used a text in which the conversation involved was presented in reported speech, as shown in (4) and (5). This kind of structure could be interfering with other processes that can be co-occurrent with implicature processing. In our experiment, we present a dialogical structure with a context, a question, and an answer, using the notion of adjacency pair proposed by
Schegloff and Sacks (
1973). An adjacency pair is a sequence of two utterances produced by different speakers with a functional relationship between the two parts (e.g., question–answer, offer–acceptance/rejection). As
Levinson (
1983) noted, adjacency pairs have conditional relevance, that is, given the first utterance, the following utterance is relevant and expectable. Thus, this unit, which works as the fundamental unit of conversation, simplifies the indirect speech structure and provides a reduced, but richer context for the stimuli.
The third goal was to assess how the implicature with algunos ‘some’ behaves in Spanish. It is worth mentioning that the semantics of the quantifier algunos ‘some’ overlaps with the meaning of the indefinite plural article unos/unas in lower-bound contexts (
Gutiérrez-Rexach 2001). For example, in Spanish, the dialog in (3) could also be formulated with
unos instead of
algunos ‘some’.
6. — ¿Qué pasó?
— Llegaron unos/algunos paquetes.
Nevertheless, the main distinction among the two is that only
algunos ‘some’ can work, as it does in English, as an implicature trigger (
Vargas-Tokuda et al. 2009). This raises the question about whether it is possible to access to the logical meaning of
algunos ‘some’ in lower-bound contexts. If this is the case, we would expect the acceptability of the stimuli with lower-bound contexts +
algunos ‘unos’ to not show significant differences when compared with upper-bound contexts. To avoid possible confounds, we did not include
unos as a variable in our experiments.
4. General Discussion
In our study, we conducted two experiments that explored the processing of scalar implicatures with
algunos (Spanish equivalent to ‘some’) in adjacency pair contexts. The novelty in our approach was twofold. On the one side, we tested the implicature in another language, Spanish. Although there is previous research that explores the topic in Spanish (
Vargas-Tokuda et al. 2009;
Miller et al. 2016), it focused in acquisition in monolingual Spanish-speaking children and comprehension in L2 learners. To our knowledge, our study is the first in this language to approach implicature processing in adult native speakers and to support, with experimental evidence, the idea that the word
algunos ‘some’ triggers the same kind of scalar implicature that it does in other languages like English, Greek, and French. Furthermore, using an acceptability judgment task, we tested if the contexts of our stimuli were constraining the interpretation of the quantifiers. Our results showed that dialogues that combined lower-bound contexts with the quantifier
only some were less acceptable, suggesting that lower-bound contexts made less relevant the meaning of
some as
only some under that condition.
On the other side, we proposed a simplification of the stimuli in order to rule out other cues that can interact, to either facilitate or inhibit, with the inference process. The results of the self-paced reading task suggested that with an adjacency pair structure, the derivation of scalar implicatures is context sensitive, rapid, and effortless, whereas the retrieving of the logical meaning of some is effortful.
In general terms, our results are in line with the predictions of default models and support the idea that there is not a ‘literal first’ interpretation (
Breheny 2019) of the quantifier
algunos ‘some’ in Spanish.
Our findings could be explained in three different ways.
Degen and Tanenhaus (
2015) have proposed, in line with context-based approaches, a unique mechanism for the derivation of implicatures, sensitive to different cues that can make inferences appear context-based (effortful and slow) or default (effortless and rapid). This could explain our results in terms of the manipulation of the structure of our stimuli: adjacency pairs rule out the complexity of the reported speech structure and make the contextual cues that influence the processing of the dialog more readily available. The adjacency pair structure, along with the upper-bound context, supports the expectation of a certain answer (that is, the quantified element) and this expectation results in a lower reading time for the answer segment with
some. In this sense, we argue that the adjacency pair structure that we use in our stimuli offered a more natural, conversational-like input to process, with a higher and more accessible contextual support. In line with this,
Huang and Snedeker (
2011) proposed a two-route derivation (bottom-up and top-down). Scalar implicatures could be derived by a top-down route when the information structure of the context causes a specific meaning to be expected, even before the appearance of the GCI trigger. In a less supportive context (LB), a bottom-up mechanism is necessary to arrive at the actual meaning of the trigger.
Another alternative is to consider scalar inferences as a case of lexical ambiguity. From this perspective,
some would have two meanings specified in the lexical entry, and instead of the implicated meaning being derived by a global-inference mechanism, it would be a result of ambiguity resolution. Seminal research in lexical ambiguity (
Simpson 1994) has proposed two models for processing ambiguity, which differ in the role of the disambiguating context in the initial phase of access to meaning. These models may be linked to the implicature models proposed in the introduction. One model, the selective access model, states that context highly constrains the access process, and that only the appropriate meaning of an ambiguous word is accessed. This model is in line with the guided-by-context approach, although it will predict no differences in reading times for quantifiers (such as the study of
Politzer-Ahles and Fiorentino 2013). For the other model, the autonomous access model, prior context does not affect the access phase because all of the meanings are accessed exhaustively. However, after the access phase, there is an integration phase where the appropriate meaning of the ambiguous word is selected. For a biased ambiguous word (a word that has one dominant meaning), the context will help in the selection between the options listed in the lexicon. The autonomous model can be considered to be in line with the default approach if it is assumed that the implicated meaning of scalar terms is the dominant one. If it is not the case (i.e., in lower-bound contexts), pragmatic processes would be involved in the cancellation of the dominant meaning and a further pragmatic derivation would be necessary to access the subordinate meaning. Nevertheless, it is not clear to what extent the implicated meaning is the dominant one. One possibility is that most frequent meanings are more dominant in terms of lexical access. The meaning of
some could be biased towards the
only some interpretations, if this is the most frequent use. This could be the case in Spanish where, as we have mentioned, the indefinite plural article (
unos/unas) alternates with
algunos ‘some’. A possible objection to the dominance of the so-called derived meaning of
some is the logical contradiction that arises if
at least one is considered as a subordinated meaning. However, it could be the case that the derivation of strengthened meanings from scalar terms is not actually a restriction from a looser meaning to a more restrictive one, but a case of lexical broadening (
Wilson and Cartson 2007), where the derivation starts at a restrictive meaning and extends its meaning to other loose uses. A recent study (
Barbet and Thierry 2018) showed that the
not all meaning of
some played a role in a stroop-like task even when it was not necessary for the task itself, suggesting that the lexical item
some could be ambiguous in the mental lexicon and that the
not all interpretation is the preferred one in the absence of special circumstances.
A final explanation for our data could be modeled in terms of an extended sentence gestalt (SG) comprehension model (
McClelland et al. 1989;
St. John and McClelland 1999;
Rabovsky et al. 2018). Under this explanation, which is related to the top-down/bottom-up approach, UB and LB contexts and the adjacency pair structure create a framework that acts as a representation of an event that is occurring and that includes expectations or predictions about what is coming next. For UB contexts, the expectation of encountering the meaning
only some is higher than the expectations for
at least one, whereas for LB contexts, it is the opposite.
Assuming that in Spanish, the dominant meaning of
some is the restricted one (
some but not all), the probability of encountering the word
algunos ‘some’ is lower in LB contexts than in UB contexts, or than the probability of encountering the word
unos. Therefore, a bigger effort is required to merge the word to the ongoing representation of the sentence gestalt (in a broader sense of what
McClelland et al. (
1989) meant by SG and that goes beyond sentences and acts over adjacency pairs as language units) when it appears, as broadening its meaning is necessary to make it compatible. This will be reflected in higher reading times for
some in LB contexts. The same model can explain the higher reading times for
the rest segment in LB contexts + some. In this case,
the rest segment requires a bigger effort to be adjusted to the SG representation, as it is not expected and not in line with the SG representation, which treated
some as
at least one and it could be all.
In contrast to default explanations, this model does not claim that there is a backtracking or cancellation of meaning. Instead, there is only an adjustment of an SG representation that is being built online. In these terms, the cancellation of an implicature does not involve ‘going back’ and reanalyzing previous representations, but an effort to adjust the evidence provided by the new input to the current representation (and not the other way around).
This approach to implicature processing could also explain the theoretical distinction between PCIs and GCIs without appealing to two differentiated mechanisms and, what is more, contribute to a better understanding of how context modulates the relevance of newer inputs.
Under a model that combines relevance (
Sperber and Wilson 1995) and a maximal incremental account of communication (
Altmann and Mirkovic 2009), context yields not only assumptions (factual propositions), but also expectations about the following inputs (predicted information/structure). Thus, not only is the context merged into the processing after the new stimuli are presented, but it creates in advance a set of expected characteristics in the utterances that will follow. These expectations could be syntactic/structural (e.g., after a question an answer is expected) or semantic (the answer will be semantically related to the question, e.g., if we ask
what happened, an unknown event must be described). When a new utterance appears, the processor assumes that this new input needs to fit into the expectations (communicative principle of relevance), and to do so, the process will follow the optimal path to meet those expectations (cognitive principle of relevance). This explains why GCIs are rapid and effortless when the context provides clear expectations. For example, in scalar implicatures, if what is expected is
only some and the input is
some, the predicted meaning of
some is
only some, but if the context does not create that expectation or creates expectations of a different kind, the meaning of
some will need to be defined a posteriori, resulting in the extra effort reflected in the reading time latency of our experiment. Nevertheless, this proposal still gives rise to some unanswered questions, such as when these expectations are created and what aspects of the context provide the cues to create them, so further studies to analyze the role of prediction in the processing of conversational implicatures are needed.
To conclude, in our two experiments, we explored the processing of scalar implicatures in adjacency pair contexts with some. As in previous studies, our results showed that the access to the meaning of some as only some is context sensitive. However, our results suggested that the adjacency pair structure that we proposed helped to make that meaning more rapidly available.