Publications

Degaetano-Ortlieb, Stefania; Teich, Elke

Toward an optimal code for communication: the case of scientific English Journal Article

Corpus Linguistics and Linguistic Theory, 18, pp. 1-33, 2019.

We present a model of the linguistic development of scientific English from the mid-seventeenth to the late-nineteenth century, a period that witnessed significant political and social changes, including the evolution of modern science. There is a wealth of descriptive accounts of scientific English, both from a synchronic and a diachronic perspective, but only few attempts at a unified explanation of its evolution. The explanation we offer here is a communicative one: while external pressures (specialization, diversification) push for an increase in expressivity, communicative concerns pull toward convergence on particular options (conventionalization). What emerges over time is a code which is optimized for written, specialist communication, relying on specific linguistic means to modulate information content. As we show, this is achieved by the systematic interplay between lexis and grammar. The corpora we employ are the Royal Society Corpus (RSC) and for comparative purposes, the Corpus of Late Modern English (CLMET). We build various diachronic, computational n-gram language models of these corpora and then apply formal measures of information content (here: relative entropy and surprisal) to detect the linguistic features significantly contributing to diachronic change, estimate the (changing) level of information of features and capture the time course of change.

 

@article{Degaetano-Ortlieb2019b,
title = {Toward an optimal code for communication: the case of scientific English},
author = {Stefania Degaetano-Ortlieb and Elke Teich},
url = {https://www.degruyter.com/document/doi/10.1515/cllt-2018-0088/html?lang=en},
doi = {https://doi.org/10.1515/cllt-2018-0088},
year = {2019},
date = {2019},
journal = {Corpus Linguistics and Linguistic Theory},
pages = {1-33},
volume = {18},
number = {1},
abstract = {We present a model of the linguistic development of scientific English from the mid-seventeenth to the late-nineteenth century, a period that witnessed significant political and social changes, including the evolution of modern science. There is a wealth of descriptive accounts of scientific English, both from a synchronic and a diachronic perspective, but only few attempts at a unified explanation of its evolution. The explanation we offer here is a communicative one: while external pressures (specialization, diversification) push for an increase in expressivity, communicative concerns pull toward convergence on particular options (conventionalization). What emerges over time is a code which is optimized for written, specialist communication, relying on specific linguistic means to modulate information content. As we show, this is achieved by the systematic interplay between lexis and grammar. The corpora we employ are the Royal Society Corpus (RSC) and for comparative purposes, the Corpus of Late Modern English (CLMET). We build various diachronic, computational n-gram language models of these corpora and then apply formal measures of information content (here: relative entropy and surprisal) to detect the linguistic features significantly contributing to diachronic change, estimate the (changing) level of information of features and capture the time course of change.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   B1

Krielke, Marie-Pauline; Degaetano-Ortlieb, Stefania; Menzel, Katrin; Teich, Elke

Paradigmatic change and redistribution of functional load: The case of relative clauses in scientific English Miscellaneous

Symposium on Corpus Approaches to Lexicogrammar (Book of Abstracts), Edge Hill University, 2019.

@miscellaneous{Krielke2019,
title = {Paradigmatic change and redistribution of functional load: The case of relative clauses in scientific English},
author = {Marie-Pauline Krielke and Stefania Degaetano-Ortlieb and Katrin Menzel and Elke Teich},
year = {2019},
date = {2019},
booktitle = {Symposium on Corpus Approaches to Lexicogrammar (Book of Abstracts)},
address = {Edge Hill University},
pubstate = {published},
type = {miscellaneous}
}

Copy BibTeX to Clipboard

Project:   B1

Menzel, Katrin; Teich, Elke

Medical discourse across 300 years: insights from the Royal Society Corpus Inproceedings

2nd International Conference on Historical Medical Discourse (CHIMED-2), 2019.

@inproceedings{Menzel2019b,
title = {Medical discourse across 300 years: insights from the Royal Society Corpus},
author = {Katrin Menzel and Elke Teich},
year = {2019},
date = {2019},
booktitle = {2nd International Conference on Historical Medical Discourse (CHIMED-2)},
pubstate = {published},
type = {inproceedings}
}

Copy BibTeX to Clipboard

Project:   B1

Degaetano-Ortlieb, Stefania; Teich, Elke; Khamis, Ashraf; Kermes, Hannah

An Information-Theoretic Approach to Modeling Diachronic Change in Scientific English Book Chapter

Suhr, Carla; Nevalainen, Terttu; Taavitsainen, Irma (Ed.): From Data to Evidence in English Language Research, Brill, pp. 258-281, Leiden, 2019.

We present an information-theoretic approach to investigate diachronic change in scientific English. Our main assumption is that over time scientific English has become increasingly dense, i.e. linguistic constructions allowing dense packing of information are progressively used. So far, diachronic change in scientific writing has been investigated by means of frequency-based approaches (see e.g. Halliday (1988); Atkinson (1998); Biber (2006b, c); Biber and Gray (2016); Banks (2008); Taavitsainen and Pahta (2010)). We use information-theoretic measures (entropy, surprisal; Shannon (1949)) to assess features previously stated to change over time and to discover new, latent features from the data itself that are involved in diachronic change. For this, we use the Royal Society Corpus (rsc) (Kermes et al. (2016)), which spans over the time period 1665 to 1869. We present three kinds of analyses: nominal compounding (typical of academic writing), modal verbs (shown to have changed in frequency over time), and an analysis based on part-of-speech trigrams to detect new features that change diachronically. We show how information-theoretic measures help to investigate, evaluate and detect features involved in diachronic change.

@inbook{Degaetano-Ortlieb2019,
title = {An Information-Theoretic Approach to Modeling Diachronic Change in Scientific English},
author = {Stefania Degaetano-Ortlieb and Elke Teich and Ashraf Khamis and Hannah Kermes},
editor = {Carla Suhr and Terttu Nevalainen and Irma Taavitsainen},
url = {https://brill.com/display/book/edcoll/9789004390652/BP000014.xml},
doi = {https://doi.org/10.1163/9789004390652},
year = {2019},
date = {2019},
booktitle = {From Data to Evidence in English Language Research},
pages = {258-281},
publisher = {Brill},
address = {Leiden},
abstract = {We present an information-theoretic approach to investigate diachronic change in scientific English. Our main assumption is that over time scientific English has become increasingly dense, i.e. linguistic constructions allowing dense packing of information are progressively used. So far, diachronic change in scientific writing has been investigated by means of frequency-based approaches (see e.g. Halliday (1988); Atkinson (1998); Biber (2006b, c); Biber and Gray (2016); Banks (2008); Taavitsainen and Pahta (2010)). We use information-theoretic measures (entropy, surprisal; Shannon (1949)) to assess features previously stated to change over time and to discover new, latent features from the data itself that are involved in diachronic change. For this, we use the Royal Society Corpus (rsc) (Kermes et al. (2016)), which spans over the time period 1665 to 1869. We present three kinds of analyses: nominal compounding (typical of academic writing), modal verbs (shown to have changed in frequency over time), and an analysis based on part-of-speech trigrams to detect new features that change diachronically. We show how information-theoretic measures help to investigate, evaluate and detect features involved in diachronic change.},
pubstate = {published},
type = {inbook}
}

Copy BibTeX to Clipboard

Project:   B1

Wichlacz, Julia; Torralba, Álvaro; Hoffmann, Jörg

Construction-Planning Models in Minecraft Inproceedings

Proceedings of the 2nd Workshop on Hierarchical Planning at ICAPS 2019, pp. 1-5, 2019.

Minecraft is a videogame that offers many interesting challenges for AI systems. In this paper, we focus in construction scenarios where an agent must build a complex structure made of individual blocks. As higher-level objects are formed of lower-level objects, the construction can naturally be modelled as a hierarchical task network. We model a house-construction scenario in classical and HTN planning and compare the advantages and disadvantages of both kinds of models.

@inproceedings{Wichlacz2019,
title = {Construction-Planning Models in Minecraft},
author = {Julia Wichlacz and {\'A}lvaro Torralba and J{\"o}rg Hoffmann},
url = {https://www.semanticscholar.org/paper/Construction-Planning-Models-in-Minecraft-Wichlacz-Torralba/d2ffb1c4b815f1b245f248d436baf9a3c28cc148},
year = {2019},
date = {2019},
booktitle = {Proceedings of the 2nd Workshop on Hierarchical Planning at ICAPS 2019},
pages = {1-5},
abstract = {Minecraft is a videogame that offers many interesting challenges for AI systems. In this paper, we focus in construction scenarios where an agent must build a complex structure made of individual blocks. As higher-level objects are formed of lower-level objects, the construction can naturally be modelled as a hierarchical task network. We model a house-construction scenario in classical and HTN planning and compare the advantages and disadvantages of both kinds of models.},
pubstate = {published},
type = {inproceedings}
}

Copy BibTeX to Clipboard

Project:   A7

Köhn, Arne; Koller, Alexander

Talking about what is not there: Generating indefinite referring expressions in Minecraft Inproceedings

Proceedings of the 12th International Conference on Natural Language Generation, Association for Computational Linguistics, pp. 1-10, Tokyo, Japan, 2019.

When generating technical instructions, it is often necessary to describe an object that does not exist yet. For example, an NLG system which explains how to build a house needs to generate sentences like “build a wall of height five to your left” and “now build a wall on the other side.” Generating (indefinite) referring expressions to objects that do not exist yet is fundamentally different from generating the usual definite referring expressions, because the new object must be distinguished from an infinite set of possible alternatives. We formalize this problem and present an algorithm for generating such expressions, in the context of generating building instructions within the Minecraft video game.

@inproceedings{Köhn2019,
title = {Talking about what is not there: Generating indefinite referring expressions in Minecraft},
author = {Arne K{\"o}hn and Alexander Koller},
url = {https://www.aclweb.org/anthology/W19-8601},
doi = {https://doi.org/10.18653/v1/W19-8601},
year = {2019},
date = {2019},
booktitle = {Proceedings of the 12th International Conference on Natural Language Generation},
pages = {1-10},
publisher = {Association for Computational Linguistics},
address = {Tokyo, Japan},
abstract = {When generating technical instructions, it is often necessary to describe an object that does not exist yet. For example, an NLG system which explains how to build a house needs to generate sentences like “build a wall of height five to your left” and “now build a wall on the other side.” Generating (indefinite) referring expressions to objects that do not exist yet is fundamentally different from generating the usual definite referring expressions, because the new object must be distinguished from an infinite set of possible alternatives. We formalize this problem and present an algorithm for generating such expressions, in the context of generating building instructions within the Minecraft video game.},
pubstate = {published},
type = {inproceedings}
}

Copy BibTeX to Clipboard

Project:   A7

Höltje, Gerrit; Lubahn, Bente; Mecklinger, Axel

The congruent, the incongruent, and the unexpected: Event-related potentials unveil the processes involved in schematic encoding Journal Article

Neuropsychologia, 131, pp. 285-293, 2019.

Learning is most effective when new information can be related to a preexisting knowledge structure or schema. In the present study, event-related potentials (ERPs) were used to investigate the temporal dynamics of the processes by which activated schemata support the encoding of schema-congruent information. Participants learned category exemplar words that were either semantically congruent or incongruent with a preceding category cue phrase. Congruent words were composed of expected (high typicality, HT) and unexpected (low typicality, LT) category exemplars. On the next day, recognition memory for the exemplars and the category cues they were presented with was tested. Semantically related lures were used in order to ascertain that memory judgements were based on episodic memory for specific category exemplars. Generally, congruent (HT and LT) exemplars were remembered better than incongruent exemplars. ERPs recorded during the encoding of the exemplar words were compared for subsequently remembered and forgotten items. Subsequent memory effects (SME) emerged in the N400 time window at frontal electrodes and did not differ between congruent and incongruent exemplars. In the same epoch, an SME with a parietal distribution was specific for congruent exemplars, suggesting that activated schemata strengthened memory for congruent exemplars by supporting the encoding of item-specific details. Subsequently remembered LT exemplars were associated with a late frontal positivity that is assumed to reflect expectancy mismatch-related processing such as the contextual integration of an unexpected word by the suppression of strongly expected words. A correlation analysis revealed that the greater the involvement of the processes reflected by the frontal positivity, the lower the level of false positive memory responses in the test phase one day later. These results suggest that the contextual integration of schema-congruent but unexpected events involves a weakening of the representations of semantically related, but unstudied items in memory and by this benefits subsequent memory.

@article{Höltje2019,
title = {The congruent, the incongruent, and the unexpected: Event-related potentials unveil the processes involved in schematic encoding},
author = {Gerrit H{\"o}ltje and Bente Lubahn and Axel Mecklinger},
url = {https://www.sciencedirect.com/science/article/pii/S0028393219301228?via%3Dihub},
doi = {https://doi.org/10.1016/j.neuropsychologia.2019.05.013},
year = {2019},
date = {2019},
journal = {Neuropsychologia},
pages = {285-293},
volume = {131},
abstract = {Learning is most effective when new information can be related to a preexisting knowledge structure or schema. In the present study, event-related potentials (ERPs) were used to investigate the temporal dynamics of the processes by which activated schemata support the encoding of schema-congruent information. Participants learned category exemplar words that were either semantically congruent or incongruent with a preceding category cue phrase. Congruent words were composed of expected (high typicality, HT) and unexpected (low typicality, LT) category exemplars. On the next day, recognition memory for the exemplars and the category cues they were presented with was tested. Semantically related lures were used in order to ascertain that memory judgements were based on episodic memory for specific category exemplars. Generally, congruent (HT and LT) exemplars were remembered better than incongruent exemplars. ERPs recorded during the encoding of the exemplar words were compared for subsequently remembered and forgotten items. Subsequent memory effects (SME) emerged in the N400 time window at frontal electrodes and did not differ between congruent and incongruent exemplars. In the same epoch, an SME with a parietal distribution was specific for congruent exemplars, suggesting that activated schemata strengthened memory for congruent exemplars by supporting the encoding of item-specific details. Subsequently remembered LT exemplars were associated with a late frontal positivity that is assumed to reflect expectancy mismatch-related processing such as the contextual integration of an unexpected word by the suppression of strongly expected words. A correlation analysis revealed that the greater the involvement of the processes reflected by the frontal positivity, the lower the level of false positive memory responses in the test phase one day later. These results suggest that the contextual integration of schema-congruent but unexpected events involves a weakening of the representations of semantically related, but unstudied items in memory and by this benefits subsequent memory.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A6

Ankener, Christine

The influence of visual information on word predictability and processing effort PhD Thesis

Saarland University, Saarbruecken, Germany, 2019.

A word’s predictability or surprisal in linguistic context, as determined by cloze probabilities or languagemodels (e.g. Frank, 2013a) is related to processing effort, in that less expected words take more effort to process (e.g. Hale, 2001). This shows how, in purely linguistic contexts, rational approaches have been proven valid to predict and formalise results from language processing studies. However, the surprisal (or predictability) of a word may also be influenced by extra-linguistic factors, such as visual context information, as given in situated language processing. While, in the case of linguistic contexts, it is known that the incrementally processed information affects the mental model (e.g. Zwaan and Radvansky, 1998) at each word in a probabilistic way, no such observations have been made so far in the case of visual context information. Although it has been shown that in the visual world paradigm (VWP), anticipatory eye movements suggest that listeners exploit the scene to predict what will be mentioned next (Altmann and Kamide, 1999), it is so far unclear how visual information actually affects expectations for and processing effort of target words. If visual context effects on word processing effort can be observed, we hypothesise that rational concepts can be extended in order to formalise these effects, hereby making them statistically accessible for language models. In a line of experiments, I hence observe how visual information – which is inherently different from linguistic context, for instance in its non-incremental-at once-accessibility– affects target words. Our findings are a clear and robust demonstration that the non-linguistic context can immediately influence both lexical expectations, and surprisal-based processing effort as assessed by two different on-line measures of effort (a pupillary and an EEG one). Finally, I use surprisal to formalise the measured results and propose an extended formula to take visual information into account.

@phdthesis{Ankener_diss_2019,
title = {The influence of visual information on word predictability and processing effort},
author = {Christine Ankener},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:291--ds-284518},
doi = {https://doi.org/http://dx.doi.org/10.22028/D291-28451},
year = {2019},
date = {2019},
school = {Saarland University},
address = {Saarbruecken, Germany},
abstract = {A word’s predictability or surprisal in linguistic context, as determined by cloze probabilities or languagemodels (e.g. Frank, 2013a) is related to processing effort, in that less expected words take more effort to process (e.g. Hale, 2001). This shows how, in purely linguistic contexts, rational approaches have been proven valid to predict and formalise results from language processing studies. However, the surprisal (or predictability) of a word may also be influenced by extra-linguistic factors, such as visual context information, as given in situated language processing. While, in the case of linguistic contexts, it is known that the incrementally processed information affects the mental model (e.g. Zwaan and Radvansky, 1998) at each word in a probabilistic way, no such observations have been made so far in the case of visual context information. Although it has been shown that in the visual world paradigm (VWP), anticipatory eye movements suggest that listeners exploit the scene to predict what will be mentioned next (Altmann and Kamide, 1999), it is so far unclear how visual information actually affects expectations for and processing effort of target words. If visual context effects on word processing effort can be observed, we hypothesise that rational concepts can be extended in order to formalise these effects, hereby making them statistically accessible for language models. In a line of experiments, I hence observe how visual information – which is inherently different from linguistic context, for instance in its non-incremental-at once-accessibility– affects target words. Our findings are a clear and robust demonstration that the non-linguistic context can immediately influence both lexical expectations, and surprisal-based processing effort as assessed by two different on-line measures of effort (a pupillary and an EEG one). Finally, I use surprisal to formalise the measured results and propose an extended formula to take visual information into account.},
pubstate = {published},
type = {phdthesis}
}

Copy BibTeX to Clipboard

Project:   A5

Delogu, Francesca; Jachmann, Torsten; Staudte, Maria; Vespignani, Francesco; Molinaro, Nicola

Discourse Expectations Are Sensitive to the Question Under Discussion: Evidence From ERPs Journal Article

Discourse Processes, pp. 1-19, 2019.

Questions under Discussion (QUDs) have been suggested to influence the integration of individual utterances into a discourse-level representation. Previous work has shown that processing ungrammatical ellipses is facilitated when the elided material addresses an implicit QUD raised through a nonactuality implicature (NAIs). It is not clear, however, if QUDs influence discourse coherence during comprehension of fully acceptable discourse. We present two ERP studies examining the effects of QUDs introduced by NAIs using two-sentence discourses. Experiment 1 showed that processing definite NPs with inaccessible antecedents is facilitated when their content is relevant to the QUD. Using acceptable discourses, Experiment 2 showed that definite NPs failing to address a QUD elicit increased processing cost. Overall, our results indicate that QUDs raise the expectation that the following discourse will address them, providing unambiguous evidence that their influence is not limited to the processing of ungrammatical input.

@article{Delogu2019b,
title = {Discourse Expectations Are Sensitive to the Question Under Discussion: Evidence From ERPs},
author = {Francesca Delogu and Torsten Jachmann and Maria Staudte and Francesco Vespignani and Nicola Molinaro},
url = {https://doi.org/10.1080/0163853X.2019.1575140},
doi = {https://doi.org/10.1080/0163853X.2019.1575140},
year = {2019},
date = {2019},
journal = {Discourse Processes},
pages = {1-19},
abstract = {Questions under Discussion (QUDs) have been suggested to influence the integration of individual utterances into a discourse-level representation. Previous work has shown that processing ungrammatical ellipses is facilitated when the elided material addresses an implicit QUD raised through a nonactuality implicature (NAIs). It is not clear, however, if QUDs influence discourse coherence during comprehension of fully acceptable discourse. We present two ERP studies examining the effects of QUDs introduced by NAIs using two-sentence discourses. Experiment 1 showed that processing definite NPs with inaccessible antecedents is facilitated when their content is relevant to the QUD. Using acceptable discourses, Experiment 2 showed that definite NPs failing to address a QUD elicit increased processing cost. Overall, our results indicate that QUDs raise the expectation that the following discourse will address them, providing unambiguous evidence that their influence is not limited to the processing of ungrammatical input.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A5

Sekicki, Mirjana

Exploiting referential gaze for uncertainty reduction in situated language processing: an information-theoretic approach PhD Thesis

Saarland University, Saarbrücken, 2019.

A large body of contemporary psycholinguistic research utilizes the information-theoretic notions related to the transmission of information in an attempt to better understand and formalize the regularities of language production and comprehension. The overarching hypothesis is that prediction is a core mechanism underlying language comprehension. Anticipating what is likely to be mentioned next based on the previous context is what is assumed to allow for smooth and effortless communication. The anticipation of linguistic units that fit the current context reduces the uncertainty about the upcoming material, which consequently facilitates the processing of that material, in a typically noisy channel. Situated language processing allows for the integration of not only linguistic but also non-linguistic visual information that contribute to establishing the context, and facilitate the creation of anticipations regarding the upcoming linguistic material. Moreover, noticing that our interlocutor is directing her attention to a certain object, inspires a shift in our visual attention towards the same entity. Since what is relevant for our interlocutor is highly likely to be relevant for us, too, whether simply conversationally, or more importantly, even existentially (Emery, 2000). Hence, following the speaker’s referential gaze cue towards an object relevant for the current conversation has been shown to benefit listeners’ language processing, measured by shorter reaction times on subsequent tasks (e.g., Staudte & Crocker, 2011; Staudte, Crocker, Heloir, & Kipp, 2014; Knoeferle & Kreysa, 2012; Macdonald & Tatler, 2013, 2014). The present thesis aimed to provide an insight into the mechanisms behind this facilitation. We examined the dynamics of combining visual and linguistic information in creating anticipation for a specific object to be mentioned, and the effect this has on language processing. To this end we used a pupillary measure of cognitive load that is robust enough to allow for free eye movements (the Index of Cognitive Activity; Marshall, 2000). This enabled us to measure not only the visual attention during language comprehension, but also the immediately induced cognitive load at various relevant points during the auditory presentation of the linguistic stimulus. Eight experiments were conducted towards addressing our research questions. The initial three experiments established the ICA measurement in the context of our linguistic manipulation. This series of experiments included reading, cognitive load during listening, and the examination of visual attention together with cognitive load in the visual world paradigm (VWP). Subsequently, we conducted five eye tracking experiments in the VWP where the linguistic context was further enriched by a referential gaze cue. All five experiments simultaneously assessed both visual attention and the immediate cognitive load induced at different stages of sentence processing. We manipulated the existence of the referential gaze cue (Exp. 4), the probability of mention of the cued object (Exp. 4, 5), the congruency of the gaze cue and the subsequent referring expression (Exp. 6), as well as the number of cued objects with equal probability of mention (Exp. 7, 8). Finally, we examined whether the gaze cue can take the role of fully disambiguating the target referent (Exp. 8). We quantified the importance of the visual context in language processing, and showed that if a certain object from the visual context has a higher likelihood of mention given the linguistic context, its processing is facilitated, in comparison to the processing of the same sentence without the visual context. Furthermore, our results support the previous findings that the referential gaze cue leads to a shift in visual attention towards the cued object, thereby facilitating language processing. We expanded these findings by showing that it is the processing of the linguistic reference, that is the referent noun, that is facilitated by gaze-following. Importantly, perceiving and following the gaze cue did not prove costly in terms of cognitive effort, unless the cued object did not fit the verb selectional preferences. This is true regardless of the number of objects cued, or the lower likelihood of mention of the cued object. We conclude that listeners strategically use visual information to reduce the referential uncertainty for upcoming nouns but that the visual cues, such as the referential gaze cue, do not underly the same kinds of expectations (and resulting cognitive costs) as linguistic references. We did not find evidence that the gaze cue is processed in a manner comparable to noun processing, rather, it is likely perceived as a relevant piece of information introduced in addition to the linguistic material, in order to aid language processing, but, importantly, not there to substitute it.

@phdthesis{Sekicki_diss_0919,
title = {Exploiting referential gaze for uncertainty reduction in situated language processing: an information-theoretic approach},
author = {Mirjana Sekicki},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:291--ds-285651},
doi = {https://doi.org/http://dx.doi.org/10.22028/D291-28565},
year = {2019},
date = {2019},
school = {Saarland University},
address = {Saarbr{\"u}cken},
abstract = {A large body of contemporary psycholinguistic research utilizes the information-theoretic notions related to the transmission of information in an attempt to better understand and formalize the regularities of language production and comprehension. The overarching hypothesis is that prediction is a core mechanism underlying language comprehension. Anticipating what is likely to be mentioned next based on the previous context is what is assumed to allow for smooth and effortless communication. The anticipation of linguistic units that fit the current context reduces the uncertainty about the upcoming material, which consequently facilitates the processing of that material, in a typically noisy channel. Situated language processing allows for the integration of not only linguistic but also non-linguistic visual information that contribute to establishing the context, and facilitate the creation of anticipations regarding the upcoming linguistic material. Moreover, noticing that our interlocutor is directing her attention to a certain object, inspires a shift in our visual attention towards the same entity. Since what is relevant for our interlocutor is highly likely to be relevant for us, too, whether simply conversationally, or more importantly, even existentially (Emery, 2000). Hence, following the speaker’s referential gaze cue towards an object relevant for the current conversation has been shown to benefit listeners’ language processing, measured by shorter reaction times on subsequent tasks (e.g., Staudte & Crocker, 2011; Staudte, Crocker, Heloir, & Kipp, 2014; Knoeferle & Kreysa, 2012; Macdonald & Tatler, 2013, 2014). The present thesis aimed to provide an insight into the mechanisms behind this facilitation. We examined the dynamics of combining visual and linguistic information in creating anticipation for a specific object to be mentioned, and the effect this has on language processing. To this end we used a pupillary measure of cognitive load that is robust enough to allow for free eye movements (the Index of Cognitive Activity; Marshall, 2000). This enabled us to measure not only the visual attention during language comprehension, but also the immediately induced cognitive load at various relevant points during the auditory presentation of the linguistic stimulus. Eight experiments were conducted towards addressing our research questions. The initial three experiments established the ICA measurement in the context of our linguistic manipulation. This series of experiments included reading, cognitive load during listening, and the examination of visual attention together with cognitive load in the visual world paradigm (VWP). Subsequently, we conducted five eye tracking experiments in the VWP where the linguistic context was further enriched by a referential gaze cue. All five experiments simultaneously assessed both visual attention and the immediate cognitive load induced at different stages of sentence processing. We manipulated the existence of the referential gaze cue (Exp. 4), the probability of mention of the cued object (Exp. 4, 5), the congruency of the gaze cue and the subsequent referring expression (Exp. 6), as well as the number of cued objects with equal probability of mention (Exp. 7, 8). Finally, we examined whether the gaze cue can take the role of fully disambiguating the target referent (Exp. 8). We quantified the importance of the visual context in language processing, and showed that if a certain object from the visual context has a higher likelihood of mention given the linguistic context, its processing is facilitated, in comparison to the processing of the same sentence without the visual context. Furthermore, our results support the previous findings that the referential gaze cue leads to a shift in visual attention towards the cued object, thereby facilitating language processing. We expanded these findings by showing that it is the processing of the linguistic reference, that is the referent noun, that is facilitated by gaze-following. Importantly, perceiving and following the gaze cue did not prove costly in terms of cognitive effort, unless the cued object did not fit the verb selectional preferences. This is true regardless of the number of objects cued, or the lower likelihood of mention of the cued object. We conclude that listeners strategically use visual information to reduce the referential uncertainty for upcoming nouns but that the visual cues, such as the referential gaze cue, do not underly the same kinds of expectations (and resulting cognitive costs) as linguistic references. We did not find evidence that the gaze cue is processed in a manner comparable to noun processing, rather, it is likely perceived as a relevant piece of information introduced in addition to the linguistic material, in order to aid language processing, but, importantly, not there to substitute it.},
pubstate = {published},
type = {phdthesis}
}

Copy BibTeX to Clipboard

Project:   A5

Häuser, Katja; Demberg, Vera; Kray, Jutta

Effects of aging and dual-task demands on the comprehension of less expected sentence continuations: Evidence from pupillometry Journal Article

Frontiers in Psychology, 10, 2019, ISSN 1664-1078.

Prior studies on language processing in aging have shown that older adults experience integration difficulties for contextually unpredictable target words (as indicated by low cloze probabilities in prior ratings), and that such comprehension difficulties are more likely to occur under more demanding processing conditions (e.g., dual-task situations). However, these effects have primarily been demonstrated for conditions when cloze probability of the linguistic stimuli was very low. The question we asked here was do dual-task demands also impair comprehension when target words provide a good, but not perfect, match with prior context? We used a dual-task design, consisting of a sentence comprehension and secondary motor tracking task. Critical target words were those which were not perfectly predictable based on context (words with a cloze probability of 0.7), as opposed to words that were near perfectly predictable based on context (cloze probabilities of 0.99). As a measure to index online processing difficulty for less expected target words, we took into account participants’ pupil size. Separate mixed effects models were fit for language comprehension, motor tracking, and pupil size, showing the following: (1) dual-task demands led to age-related comprehension difficulties when target words were less expected (as opposed to very highly expected), (2) integration difficulty in older adults was related to cognitive overload as less expected sentence continuations progressed over time, resulting in behavioral trade-offs between language comprehension and motor tracking, and (3) lower levels of working memory were predictive of whether or not older adults experienced cognitive overload when processing less expected words. In sum, more demanding processing conditions lead to comprehension impairments when words are highly unpredictable based on context, as many prior studies showed. Comprehension impairments among older individuals also occur for conditions when words provide a good, but not perfect, match with prior context. Higher working memory capacity can alleviate such impairments in older adults, thereby suggesting that only high-WM older adults have sufficient cognitive resources to pre-activate words that complete a sentence context plausibly, but not perfectly.

@article{Häuser2019,
title = {Effects of aging and dual-task demands on the comprehension of less expected sentence continuations: Evidence from pupillometry},
author = {Katja H{\"a}user and Vera Demberg and Jutta Kray},
url = {https://www.frontiersin.org/articles/10.3389/fpsyg.2019.00709/full},
doi = {https://doi.org/10.3389/fpsyg.2019.00709},
year = {2019},
date = {2019},
journal = {Frontiers in Psychology},
volume = {10},
abstract = {Prior studies on language processing in aging have shown that older adults experience integration difficulties for contextually unpredictable target words (as indicated by low cloze probabilities in prior ratings), and that such comprehension difficulties are more likely to occur under more demanding processing conditions (e.g., dual-task situations). However, these effects have primarily been demonstrated for conditions when cloze probability of the linguistic stimuli was very low. The question we asked here was do dual-task demands also impair comprehension when target words provide a good, but not perfect, match with prior context? We used a dual-task design, consisting of a sentence comprehension and secondary motor tracking task. Critical target words were those which were not perfectly predictable based on context (words with a cloze probability of 0.7), as opposed to words that were near perfectly predictable based on context (cloze probabilities of 0.99). As a measure to index online processing difficulty for less expected target words, we took into account participants’ pupil size. Separate mixed effects models were fit for language comprehension, motor tracking, and pupil size, showing the following: (1) dual-task demands led to age-related comprehension difficulties when target words were less expected (as opposed to very highly expected), (2) integration difficulty in older adults was related to cognitive overload as less expected sentence continuations progressed over time, resulting in behavioral trade-offs between language comprehension and motor tracking, and (3) lower levels of working memory were predictive of whether or not older adults experienced cognitive overload when processing less expected words. In sum, more demanding processing conditions lead to comprehension impairments when words are highly unpredictable based on context, as many prior studies showed. Comprehension impairments among older individuals also occur for conditions when words provide a good, but not perfect, match with prior context. Higher working memory capacity can alleviate such impairments in older adults, thereby suggesting that only high-WM older adults have sufficient cognitive resources to pre-activate words that complete a sentence context plausibly, but not perfectly.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A4

Tröger, Johannes; Linz, Nicklas; König, Alexandra; Robert, Philippe; Alexandersson, Jan; Peter, Jessica; Kray, Jutta

Exploitation vs. Exploration - Computational temporal and semantic analysis explains semantic verbal fluency impairment in Alzheimer’s disease Journal Article

Neuropsychologia, 131, pp. 53-61, 2019.

Impaired Semantic Verbal Fluency (SVF) in dementia due to Alzheimer’s Disease (AD) and its precursor Mild Cognitive Impairment (MCI) is well known. Yet, it remains open whether this impairment mirrors the breakdown of semantic memory retrieval processes or executive control processes. Therefore, qualitative analysis of the SVF has been proposed but is limited in terms of methodology and feasibility in clinical practice. Consequently, research draws no conclusive picture which of these afore-mentioned processes drives the SVF impairment in AD and MCI. This study uses a qualitative computational approach – combining temporal and semantic information – to investigate exploitation and exploration patterns as indicators for semantic memory retrieval and executive control processes. Audio SVF recordings of 20 controls (C, 66-81 years), 55 MCI (57-94 years) and 20 AD subjects (66-82 years) were assessed while groups were matched according to age and education. All groups produced, on average, the same amount of semantically related items in rapid succession within word clusters. Conversely, towards AD, there was a clear decline in semantic as well as temporal exploration patterns between clusters. Results strongly point towards preserved exploitation – semantic memory retrieval processes – and hampered exploration – executive control processes – in AD and potentially in MCI.

@article{Tröger2019,
title = {Exploitation vs. Exploration - Computational temporal and semantic analysis explains semantic verbal fluency impairment in Alzheimer’s disease},
author = {Johannes Tr{\"o}ger and Nicklas Linz and Alexandra K{\"o}nig and Philippe Robert and Jan Alexandersson and Jessica Peter and Jutta Kray},
url = {https://www.sciencedirect.com/science/article/abs/pii/S0028393218305116?via%3Dihub},
doi = {https://doi.org/10.1016/j.neuropsychologia.2019.05.007},
year = {2019},
date = {2019},
journal = {Neuropsychologia},
pages = {53-61},
volume = {131},
abstract = {Impaired Semantic Verbal Fluency (SVF) in dementia due to Alzheimer's Disease (AD) and its precursor Mild Cognitive Impairment (MCI) is well known. Yet, it remains open whether this impairment mirrors the breakdown of semantic memory retrieval processes or executive control processes. Therefore, qualitative analysis of the SVF has been proposed but is limited in terms of methodology and feasibility in clinical practice. Consequently, research draws no conclusive picture which of these afore-mentioned processes drives the SVF impairment in AD and MCI. This study uses a qualitative computational approach - combining temporal and semantic information - to investigate exploitation and exploration patterns as indicators for semantic memory retrieval and executive control processes. Audio SVF recordings of 20 controls (C, 66-81 years), 55 MCI (57-94 years) and 20 AD subjects (66-82 years) were assessed while groups were matched according to age and education. All groups produced, on average, the same amount of semantically related items in rapid succession within word clusters. Conversely, towards AD, there was a clear decline in semantic as well as temporal exploration patterns between clusters. Results strongly point towards preserved exploitation - semantic memory retrieval processes - and hampered exploration - executive control processes - in AD and potentially in MCI.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A4

Vogels, Jorrig; Howcroft, David M.; Tourtouri, Elli; Demberg, Vera

How speakers adapt object descriptions to listeners under load Journal Article

Language, Cognition and Neuroscience, 35, Routledge, pp. 78-92, 2019.

A controversial issue in psycholinguistics is the degree to which speakers employ audience design during language production. Hypothesising that a consideration of the listener’s needs is particularly relevant when the listener is under cognitive load, we had speakers describe objects for a listener performing an easy or a difficult simulated driving task. We predicted that speakers would introduce more redundancy in their descriptions in the difficult driving task, thereby accommodating the listener’s reduced cognitive capacity. The results showed that speakers did not adapt their descriptions to a change in the listener’s cognitive load. However, speakers who had experienced the driving task themselves before and who were presented with the difficult driving task first were more redundant than other speakers. These findings may suggest that speakers only consider the listener’s needs in the presence of strong enough cues, and do not update their beliefs about these needs during the task.

@article{Vogels2019,
title = {How speakers adapt object descriptions to listeners under load},
author = {Jorrig Vogels and David M. Howcroft and Elli Tourtouri and Vera Demberg},
url = {https://www.tandfonline.com/doi/full/10.1080/23273798.2019.1648839},
doi = {https://doi.org/10.1080/23273798.2019.1648839},
year = {2019},
date = {2019},
journal = {Language, Cognition and Neuroscience},
pages = {78-92},
publisher = {Routledge},
volume = {35},
number = {1},
abstract = {A controversial issue in psycholinguistics is the degree to which speakers employ audience design during language production. Hypothesising that a consideration of the listener’s needs is particularly relevant when the listener is under cognitive load, we had speakers describe objects for a listener performing an easy or a difficult simulated driving task. We predicted that speakers would introduce more redundancy in their descriptions in the difficult driving task, thereby accommodating the listener’s reduced cognitive capacity. The results showed that speakers did not adapt their descriptions to a change in the listener’s cognitive load. However, speakers who had experienced the driving task themselves before and who were presented with the difficult driving task first were more redundant than other speakers. These findings may suggest that speakers only consider the listener’s needs in the presence of strong enough cues, and do not update their beliefs about these needs during the task.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A4

Ostermann, Simon

Script knowledge for natural language understanding PhD Thesis

Saarland University, Saarbruecken, Germany, 2019.

While people process text, they make frequent use of information that is assumed to be common ground and left implicit in the text. One important type of such commonsense knowledge is script knowledge, which is the knowledge about the events and participants in everyday activities such as visiting a restaurant. Due to its implicitness, it is hard for machines to exploit such script knowledge for natural language processing (NLP). This dissertation addresses the role of script knowledge in a central field of NLP, natural language understanding (NLU). In the first part of this thesis, we address script parsing. The idea of script parsing is to align event and participant mentions in a text with an underlying script representation. This makes it possible for a system to leverage script knowledge for downstream tasks. We develop the first script parsing model for events that can be trained on a large scale on crowdsourced script data. The model is implemented as a linear-chain conditional random field and trained on sequences of short event descriptions, implicitly exploiting the inherent event ordering information. We show that this ordering information plays a crucial role for script parsing. Our model provides an important first step towards facilitating the use of script knowledge for NLU. In the second part of the thesis, we move our focus to an actual application in the area of NLU, i.e. machine comprehension. For the first time, we provide data sets for the systematic evaluation of the contribution of script knowledge for machine comprehension. We create MCScript, a corpus of narrations about everyday activities and questions on the texts. By collecting questions based on a scenario rather than a text, we aimed at creating challenging questions which require script knowledge for finding the correct answer. Based on the findings of a shared task carried out with the data set, which indicated that script knowledge is not relevant for good performance on our corpus, we revised the data collection process and created a second version of the data set.

@phdthesis{Ostermann2019,
title = {Script knowledge for natural language understanding},
author = {Simon Ostermann},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:291--ds-313016},
doi = {https://doi.org/10.22028/D291-31301},
year = {2019},
date = {2019},
school = {Saarland University},
address = {Saarbruecken, Germany},
abstract = {While people process text, they make frequent use of information that is assumed to be common ground and left implicit in the text. One important type of such commonsense knowledge is script knowledge, which is the knowledge about the events and participants in everyday activities such as visiting a restaurant. Due to its implicitness, it is hard for machines to exploit such script knowledge for natural language processing (NLP). This dissertation addresses the role of script knowledge in a central field of NLP, natural language understanding (NLU). In the first part of this thesis, we address script parsing. The idea of script parsing is to align event and participant mentions in a text with an underlying script representation. This makes it possible for a system to leverage script knowledge for downstream tasks. We develop the first script parsing model for events that can be trained on a large scale on crowdsourced script data. The model is implemented as a linear-chain conditional random field and trained on sequences of short event descriptions, implicitly exploiting the inherent event ordering information. We show that this ordering information plays a crucial role for script parsing. Our model provides an important first step towards facilitating the use of script knowledge for NLU. In the second part of the thesis, we move our focus to an actual application in the area of NLU, i.e. machine comprehension. For the first time, we provide data sets for the systematic evaluation of the contribution of script knowledge for machine comprehension. We create MCScript, a corpus of narrations about everyday activities and questions on the texts. By collecting questions based on a scenario rather than a text, we aimed at creating challenging questions which require script knowledge for finding the correct answer. Based on the findings of a shared task carried out with the data set, which indicated that script knowledge is not relevant for good performance on our corpus, we revised the data collection process and created a second version of the data set.},
pubstate = {published},
type = {phdthesis}
}

Copy BibTeX to Clipboard

Project:   A3

Wanzare, Lilian Diana Awuor

Script acquisition: a crowdsourcing and text mining approach PhD Thesis

Saarland University, Saarbruecken, Germany, 2019.

According to Grice’s (1975) theory of pragmatics, people tend to omit basic information when participating in a conversation (or writing a narrative) under the assumption that left out details are already known or can be inferred from commonsense knowledge by the hearer (or reader). Writing and understanding of texts makes particular use of a specific kind of common-sense knowledge, referred to as script knowledge. Schank and Abelson (1977) proposed Scripts as a model of human knowledge represented in memory that stores the frequent habitual activities, called scenarios, (e.g. eating in a fast food restaurant, etc.), and the different courses of action in those routines. This thesis addresses measures to provide a sound empirical basis for high-quality script models. We work on three key areas related to script modeling: script knowledge acquisition, script induction and script identification in text. We extend the existing repository of script knowledge bases in two different ways. First, we crowdsource a corpus of 40 scenarios with 100 event sequence descriptions (ESDs) each, thus going beyond the size of previous script collections. Second, the corpus is enriched with partial alignments of ESDs, done by human annotators. The crowdsourced partial alignments are used as prior knowledge to guide the semi-supervised script-induction algorithm proposed in this dissertation. We further present a semi-supervised clustering approach to induce script structure from crowdsourced descriptions of event sequences by grouping event descriptions into paraphrase sets and inducing their temporal order. The proposed semi-supervised clustering model better handles order variation in scripts and extends script representation formalism, Temporal Script graphs, by incorporating „arbitrary order“ equivalence classes in order to allow for the flexible event order inherent in scripts. In the third part of this dissertation, we introduce the task of scenario detection, in which we identify references to scripts in narrative texts. We curate a benchmark dataset of annotated narrative texts, with segments labeled according to the scripts they instantiate. The dataset is the first of its kind. The analysis of the annotation shows that one can identify scenario references in text with reasonable reliability. Subsequently, we proposes a benchmark model that automatically segments and identifies text fragments referring to given scenarios. The proposed model achieved promising results, and therefore opens up research on script parsing and wide coverage script acquisition.

@phdthesis{Wanzare2019,
title = {Script acquisition: a crowdsourcing and text mining approach},
author = {Lilian Diana Awuor Wanzare},
url = {http://nbn-resolving.de/urn:nbn:de:bsz:291--ds-301634},
doi = {https://doi.org/http://dx.doi.org/10.22028/D291-30163},
year = {2019},
date = {2019},
school = {Saarland University},
address = {Saarbruecken, Germany},
abstract = {According to Grice’s (1975) theory of pragmatics, people tend to omit basic information when participating in a conversation (or writing a narrative) under the assumption that left out details are already known or can be inferred from commonsense knowledge by the hearer (or reader). Writing and understanding of texts makes particular use of a specific kind of common-sense knowledge, referred to as script knowledge. Schank and Abelson (1977) proposed Scripts as a model of human knowledge represented in memory that stores the frequent habitual activities, called scenarios, (e.g. eating in a fast food restaurant, etc.), and the different courses of action in those routines. This thesis addresses measures to provide a sound empirical basis for high-quality script models. We work on three key areas related to script modeling: script knowledge acquisition, script induction and script identification in text. We extend the existing repository of script knowledge bases in two different ways. First, we crowdsource a corpus of 40 scenarios with 100 event sequence descriptions (ESDs) each, thus going beyond the size of previous script collections. Second, the corpus is enriched with partial alignments of ESDs, done by human annotators. The crowdsourced partial alignments are used as prior knowledge to guide the semi-supervised script-induction algorithm proposed in this dissertation. We further present a semi-supervised clustering approach to induce script structure from crowdsourced descriptions of event sequences by grouping event descriptions into paraphrase sets and inducing their temporal order. The proposed semi-supervised clustering model better handles order variation in scripts and extends script representation formalism, Temporal Script graphs, by incorporating "arbitrary order" equivalence classes in order to allow for the flexible event order inherent in scripts. In the third part of this dissertation, we introduce the task of scenario detection, in which we identify references to scripts in narrative texts. We curate a benchmark dataset of annotated narrative texts, with segments labeled according to the scripts they instantiate. The dataset is the first of its kind. The analysis of the annotation shows that one can identify scenario references in text with reasonable reliability. Subsequently, we proposes a benchmark model that automatically segments and identifies text fragments referring to given scenarios. The proposed model achieved promising results, and therefore opens up research on script parsing and wide coverage script acquisition.},
pubstate = {published},
type = {phdthesis}
}

Copy BibTeX to Clipboard

Project:   A3

Ostermann, Simon; Roth, Michael; Pinkal, Manfred

MCScript2.0: A Machine Comprehension Corpus Focused on Script Events and Participants Inproceedings

Proceedings of the Eighth Joint Conference on Lexical and Computational Semantics (* SEM 2019), pp. 103-117, 2019.

We introduce MCScript2.0, a machine comprehension corpus for the end-to-end evaluation of script knowledge. MCScript2.0 contains approx. 20,000 questions on approx. 3,500 texts, crowdsourced based on a new collection process that results in challenging questions. Half of the questions cannot be answered from the reading texts, but require the use of commonsense and, in particular, script knowledge. We give a thorough analysis of our corpus and show that while the task is not challenging to humans, existing machine comprehension models fail to perform well on the data, even if they make use of a commonsense knowledge base.

@inproceedings{ostermann2019mcscript2,
title = {MCScript2.0: A Machine Comprehension Corpus Focused on Script Events and Participants},
author = {Simon Ostermann and Michael Roth and Manfred Pinkal},
url = {https://www.aclweb.org/anthology/S19-1012},
year = {2019},
date = {2019-10-17},
booktitle = {Proceedings of the Eighth Joint Conference on Lexical and Computational Semantics (* SEM 2019)},
pages = {103-117},
abstract = {We introduce MCScript2.0, a machine comprehension corpus for the end-to-end evaluation of script knowledge. MCScript2.0 contains approx. 20,000 questions on approx. 3,500 texts, crowdsourced based on a new collection process that results in challenging questions. Half of the questions cannot be answered from the reading texts, but require the use of commonsense and, in particular, script knowledge. We give a thorough analysis of our corpus and show that while the task is not challenging to humans, existing machine comprehension models fail to perform well on the data, even if they make use of a commonsense knowledge base.},
pubstate = {published},
type = {inproceedings}
}

Copy BibTeX to Clipboard

Project:   A3

Zhai, Fangzhou; Demberg, Vera; Shkadzko, Pavel; Shi, Wei; Sayeed, Asad

A Hybrid Model for Globally Coherent Story Generation Inproceedings

Proceedings of the Second Workshop on Storytelling, Association for Computational Linguistics, pp. 34-45, Florence, Italy, 2019.

Automatically generating globally coherent stories is a challenging problem. Neural text generation models have been shown to perform well at generating fluent sentences from data, but they usually fail to keep track of the overall coherence of the story after a couple of sentences. Existing work that incorporates a text planning module succeeded in generating recipes and dialogues, but appears quite data-demanding. We propose a novel story generation approach that generates globally coherent stories from a fairly small corpus. The model exploits a symbolic text planning module to produce text plans, thus reducing the demand of data; a neural surface realization module then generates fluent text conditioned on the text plan. Human evaluation showed that our model outperforms various baselines by a wide margin and generates stories which are fluent as well as globally coherent.

@inproceedings{Fangzhou2019,
title = {A Hybrid Model for Globally Coherent Story Generation},
author = {Fangzhou Zhai and Vera Demberg and Pavel Shkadzko and Wei Shi and Asad Sayeed},
url = {https://aclanthology.org/W19-3404},
doi = {https://doi.org/10.18653/v1/W19-3404},
year = {2019},
date = {2019},
booktitle = {Proceedings of the Second Workshop on Storytelling},
pages = {34-45},
publisher = {Association for Computational Linguistics},
address = {Florence, Italy},
abstract = {Automatically generating globally coherent stories is a challenging problem. Neural text generation models have been shown to perform well at generating fluent sentences from data, but they usually fail to keep track of the overall coherence of the story after a couple of sentences. Existing work that incorporates a text planning module succeeded in generating recipes and dialogues, but appears quite data-demanding. We propose a novel story generation approach that generates globally coherent stories from a fairly small corpus. The model exploits a symbolic text planning module to produce text plans, thus reducing the demand of data; a neural surface realization module then generates fluent text conditioned on the text plan. Human evaluation showed that our model outperforms various baselines by a wide margin and generates stories which are fluent as well as globally coherent.},
pubstate = {published},
type = {inproceedings}
}

Copy BibTeX to Clipboard

Projects:   A3 B2

Venhuizen, Noortje; Crocker, Matthew W.; Brouwer, Harm

Expectation-based Comprehension: Modeling the interaction of world knowledge and linguistic experience Journal Article

Discourse Processes, 56, pp. 229-255, 2019.

The processing difficulty of each word we encounter in a sentence is affected by both our prior linguistic experience and our general knowledge about the world. Computational models of incremental language processing have, however, been limited in accounting for the influence of world knowledge.

We develop an incremental model of language comprehension that constructs – on a word-by-word basis – rich, probabilistic situation model representations. To quantify linguistic processing effort, we adopt Surprisal Theory, which asserts that the processing difficulty incurred by a word is inversely proportional to its expectancy (Hale, 2001; Levy, 2008). In contrast with typical language model implementations of surprisal, the proposed model instantiates a novel comprehension-centric metric of surprisal that reflects the likelihood of the unfolding utterance meaning as established after processing each word. Simulations are presented that demonstrate that linguistic experience and world knowledge are integrated in the model at the level of interpretation and combine in determining online expectations.

@article{Venhuizen2019,
title = {Expectation-based Comprehension: Modeling the interaction of world knowledge and linguistic experience},
author = {Noortje Venhuizen and Matthew W. Crocker and Harm Brouwer},
url = {https://www.tandfonline.com/doi/full/10.1080/0163853X.2018.1448677},
doi = {https://doi.org/10.1080/0163853X.2018.1448677},
year = {2019},
date = {2019},
journal = {Discourse Processes},
pages = {229-255},
volume = {56},
number = {3},
abstract = {The processing difficulty of each word we encounter in a sentence is affected by both our prior linguistic experience and our general knowledge about the world. Computational models of incremental language processing have, however, been limited in accounting for the influence of world knowledge. We develop an incremental model of language comprehension that constructs - on a word-by-word basis - rich, probabilistic situation model representations. To quantify linguistic processing effort, we adopt Surprisal Theory, which asserts that the processing difficulty incurred by a word is inversely proportional to its expectancy (Hale, 2001; Levy, 2008). In contrast with typical language model implementations of surprisal, the proposed model instantiates a novel comprehension-centric metric of surprisal that reflects the likelihood of the unfolding utterance meaning as established after processing each word. Simulations are presented that demonstrate that linguistic experience and world knowledge are integrated in the model at the level of interpretation and combine in determining online expectations.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Projects:   A1 C3

Delogu, Francesca; Brouwer, Harm; Crocker, Matthew W.

Event-related potentials index lexical retrieval (N400) and integration (P600) during language comprehension Journal Article

Brain and Cognition, 135, 2019.

The functional interpretation of two salient language-sensitive ERP components – the N400 and the P600 – remains a matter of debate. Prominent alternative accounts link the N400 to processes related to lexical retrieval, semantic integration, or both, while the P600 has been associated with syntactic reanalysis or, alternatively, to semantic integration. The often overlapping predictions of these competing accounts in extant experimental designs, however, has meant that previous findings have failed to clearly decide among them. Here, we present an experiment that directly tests the competing hypotheses using a design that clearly teases apart the retrieval versus integration view of the N400, while also dissociating a syntactic reanalysis/reprocessing account of the P600 from semantic integration. Our findings provide support for an integrated functional interpretation according to which the N400 reflects context-sensitive lexical retrieval – but not integration – processes. While the observed P600 effects were not predicted by any account, we argue that they can be reconciled with the integration view, if spatio-temporal overlap of ERP components is taken into consideration.

@article{delogu2019event,
title = {Event-related potentials index lexical retrieval (N400) and integration (P600) during language comprehension},
author = {Francesca Delogu and Harm Brouwer and Matthew W. Crocker},
url = {https://www.sciencedirect.com/science/article/pii/S0278262618304299},
doi = {https://doi.org/10.1016/j.bandc.2019.05.007},
year = {2019},
date = {2019},
journal = {Brain and Cognition},
volume = {135},
abstract = {The functional interpretation of two salient language-sensitive ERP components - the N400 and the P600 - remains a matter of debate. Prominent alternative accounts link the N400 to processes related to lexical retrieval, semantic integration, or both, while the P600 has been associated with syntactic reanalysis or, alternatively, to semantic integration. The often overlapping predictions of these competing accounts in extant experimental designs, however, has meant that previous findings have failed to clearly decide among them. Here, we present an experiment that directly tests the competing hypotheses using a design that clearly teases apart the retrieval versus integration view of the N400, while also dissociating a syntactic reanalysis/reprocessing account of the P600 from semantic integration. Our findings provide support for an integrated functional interpretation according to which the N400 reflects context-sensitive lexical retrieval - but not integration - processes. While the observed P600 effects were not predicted by any account, we argue that they can be reconciled with the integration view, if spatio-temporal overlap of ERP components is taken into consideration.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Project:   A1

Tourtouri, Elli; Delogu, Francesca; Sikos, Les; Crocker, Matthew W.

Rational over-specification in visually-situated comprehension and production Journal Article

Journal of Cultural Cognitive Science, 3, pp. 175-202, 2019.

Contrary to the Gricean maxims of quantity (Grice, in: Cole, Morgan (eds) Syntax and semantics: speech acts, vol III, pp 41-58, Academic Press, New York, 1975), it has been repeatedly shown that speakers often include redundant information in their utterances (over-specifications). Previous research on referential communication has long debated whether this redundancy is the result of speaker-internal or addressee-oriented processes, while it is also unclear whether referential redundancy hinders or facilitates comprehension.

We present an information-theoretic explanation for the use of over-specification in visually-situated communication, which quantifies the amount of uncertainty regarding the referent as entropy (Shannon in Bell Syst Tech J 5:10, https://doi.org/10.1002/j.1538-7305.1948.tb01338.x, 1948). Examining both the comprehension and production of over-specifications, we present evidence that (a) listeners’ processing is facilitated by the use of redundancy as well as by a greater reduction of uncertainty early on in the utterance, and (b) that at least for some speakers, listeners’ processing concerns influence their encoding of over-specifications: Speakers were more likely to use redundant adjectives when these adjectives reduced entropy to a higher degree than adjectives necessary for target identification.

@article{Tourtouri2019,
title = {Rational over-specification in visually-situated comprehension and production},
author = {Elli Tourtouri and Francesca Delogu and Les Sikos and Matthew W. Crocker},
url = {https://link.springer.com/article/10.1007%2Fs41809-019-00032-6},
doi = {https://doi.org/10.1007/s41809-019-00032-6},
year = {2019},
date = {2019},
journal = {Journal of Cultural Cognitive Science},
pages = {175-202},
volume = {3},
number = {2},
abstract = {Contrary to the Gricean maxims of quantity (Grice, in: Cole, Morgan (eds) Syntax and semantics: speech acts, vol III, pp 41-58, Academic Press, New York, 1975), it has been repeatedly shown that speakers often include redundant information in their utterances (over-specifications). Previous research on referential communication has long debated whether this redundancy is the result of speaker-internal or addressee-oriented processes, while it is also unclear whether referential redundancy hinders or facilitates comprehension. We present an information-theoretic explanation for the use of over-specification in visually-situated communication, which quantifies the amount of uncertainty regarding the referent as entropy (Shannon in Bell Syst Tech J 5:10, https://doi.org/10.1002/j.1538-7305.1948.tb01338.x, 1948). Examining both the comprehension and production of over-specifications, we present evidence that (a) listeners’ processing is facilitated by the use of redundancy as well as by a greater reduction of uncertainty early on in the utterance, and (b) that at least for some speakers, listeners’ processing concerns influence their encoding of over-specifications: Speakers were more likely to use redundant adjectives when these adjectives reduced entropy to a higher degree than adjectives necessary for target identification.},
pubstate = {published},
type = {article}
}

Copy BibTeX to Clipboard

Projects:   A1 C3

Successfully