Skip to main content
Skip to main content

Research

Research at our top-ranked department spans syntax, semantics, phonology, language acquisition, computational linguistics, psycholinguistics and neurolinguistics. 

Connections between our core competencies are strong, with theoretical, experimental and computational work typically pursued in tandem.

A network of collaboration at all levels sustains a research climate that is both vigorous and friendly. Here new ideas develop in conversation, stimulated by the steady activity of our labs and research groups, frequent student meetings with faculty, regular talks by local and invited scholars and collaborations with the broader University of Maryland language science community, the largest and most integrated language science research community in North America.

Show activities matching...

filter by...

All Focus is Contrastive: On Polarity (Verum) Focus, Answer Focus, Contrastive Focus and Givenness

A general theory of focus and givenness.

Linguistics

Contributor(s): Daniel Goodhue
Dates:

I develop a general theory of focus and givenness that can account for truly contrastive focus, and for polarity focus, including data that are sometimes set apart under the label “verum focus”. I show that polarity focus creates challenges for classic theories of focus (e.g. Rooth 1992, a.o.) that can be dealt with by requiring that all focus marking is truly contrastive, and that givenness deaccenting imposes its own distinct requirement on prominence shifts. To enforce true contrast, I employ innocent exclusion (Fox 2007), which I suggest may impose a general filter on what counts as a valid alternative. A key, novel feature of my account is that focal targets are split into two kinds, those that are contextually supported and those that are constructed ad hoc, and that the presence of a contextually supported target can block the ability to construct an ad hoc target. This enables a novel explanation of the data motivating true contrast, and enables polarity focus to be brought into the fold of a unified and truly contrastive theory of focus. I then compare the account to theories of verum focus that make use of non-focus-based VERUM operators, and make the argument that the focus account is more parsimonious and has better empirical coverage.

Read More about All Focus is Contrastive: On Polarity (Verum) Focus, Answer Focus, Contrastive Focus and Givenness

The Power of Ignoring: Filtering Input for Argument Structure Acquisition

How to avoid learning from misleading data by identifying a filter without knowing what to filter.

Linguistics

Contributor(s): Naomi Feldman, Jeffrey Lidz
Non-ARHU Contributor(s):

Laurel Perkins *19 (UCLA)

Dates:

Learning in any domain depends on how the data for learning are represented. In the domain of language acquisition, children’s representations of the speech they hear determine what generalizations they can draw about their target grammar. But these input representations change over development asa function of children’s developing linguistic knowledge, and may be incomplete or inaccurate when children lack the knowledge to parse their input veridically. How does learning succeed in the face of potentially misleading data? We address this issue using the case study of “non-basic” clauses inverb learning. A young infant hearing What did Amy fix? might not recognize that what stands in for the direct object of fix, and might think that fix is occurring without a direct object. We follow a previous proposal that children might filter nonbasic clauses out of the data for learning verb argument structure, but offer a new approach. Instead of assuming that children identify the data to filter ina dvance, we demonstrate computationally that it is possible for learners to infer a filter on their input without knowing which clauses are nonbasic. We instantiate a learner that considers the possibility that it misparses some of the sentences it hears, and learns to filter out those parsing errors in order to correctly infer transitivity for the majority of 50 frequent verbs in child-directed speech. Our learner offers a novel solution to the problem of learning from immature input representations: Learners maybe able to avoid drawing faulty inferences from misleading data by identifying a filter on their input,without knowing in advance what needs to be filtered.

Read More about The Power of Ignoring: Filtering Input for Argument Structure Acquisition

Parallel processing in speech perception with local and global representations of linguistic context

MEG evidence for parallel representation of local and global context in speech processing.

Linguistics

Contributor(s): Ellen Lau, Philip Resnik, Shohini Bhattasali
Non-ARHU Contributor(s):

Christian Brodbeck, Aura Cruz Heredia, Jonathan Simon

Dates:

Speech processing is highly incremental. It is widely accepted that human listeners continuously use the linguistic context to anticipate upcoming concepts, words, and phonemes. However, previous evidence supports two seemingly contradictory models of how a predictive context is integrated with the bottom-up sensory input: Classic psycholinguistic paradigms suggest a two-stage process, in which acoustic input initially leads to local, context-independent representations, which are then quickly integrated with contextual constraints. This contrasts with the view that the brain constructs a single coherent, unified interpretation of the input, which fully integrates available information across representational hierarchies, and thus uses contextual constraints to modulate even the earliest sensory representations. To distinguish these hypotheses, we tested magnetoencephalography responses to continuous narrative speech for signatures of local and unified predictive models. Results provide evidence that listeners employ both types of models in parallel. Two local context models uniquely predict some part of early neural responses, one based on sublexical phoneme sequences, and one based on the phonemes in the current word alone; at the same time, even early responses to phonemes also reflect a unified model that incorporates sentence level constraints to predict upcoming phonemes. Neural source localization places the anatomical origins of the different predictive models in non-identical parts of the superior temporal lobes bilaterally, with the right hemisphere showing a relative preference for more local models. These results suggest that speech processing recruits both local and unified predictive models in parallel, reconciling previous disparate findings. Parallel models might make the perceptual system more robust, facilitate processing of unexpected inputs, and serve a function in language acquisition.

Read More about Parallel processing in speech perception with local and global representations of linguistic context

On the Acquisition of Attitude Verbs

On the acquisition of attitude verbs.

Linguistics

Contributor(s): Jeffrey Lidz, Valentine Hacquard
Dates:

Attitude verbs, such as think, want, and know, describe internal mental states that leave few cues as to their meanings in the physical world. Consequently, their acquisition requires learners to draw from indirect evidence stemming from the linguistic and conversational contexts in which they occur. This provides us a unique opportunity to probe the linguistic and cognitive abilities that children deploy in acquiring these words. Through a few case studies, we show how children make use of syntactic and pragmatic cues to figure out attitude verb meanings and how their successes, and even their mistakes, reveal remarkable conceptual, linguistic, and pragmatic sophistication.

Using surprisal and fMRI to map the neural bases of broad and local contextual prediction during natural language comprehension

Modeling the influence of local and topical context on processing via an analysis of fMRI time courses during naturalistic listening.

Linguistics

Contributor(s): Philip Resnik, Shohini Bhattasali
Dates:

Context guides comprehenders’ expectations during language processing, and information theoretic surprisal is commonly used as an index of cognitive processing effort. However, prior work using surprisal has considered only within-sentence context, using n-grams, neural language models, or syntactic structure as conditioning context. In this paper, we extend the surprisal approach to use broader topical context, investigating the influence of local and topical context on processing via an analysis of fMRI time courses collected during naturalistic listening. Lexical surprisal calculated from ngram and LSTM language models is used to capture effects of local context; to capture the effects of broader context a new metric based on topic models, topical surprisal, is introduced. We identify distinct patterns of neural activation for lexical surprisal and topical surprisal. These differing neuro-anatomical correlates suggest that local and broad contextual cues during sentence processing recruit different brain regions and that those regions of the language network functionally contribute to processing different dimensions of contextual information during comprehension. More generally, our approach adds to a growing literature using methods from computational linguistics to operationalize and test hypotheses about neuro-cognitive mechanisms in sentence processing.

Read More about Using surprisal and fMRI to map the neural bases of broad and local contextual prediction during natural language comprehension

Eighteen-month-old infants represent nonlocal syntactic dependencies

Evidence that 18-month olds already represent filler-gap dependencies.

Linguistics

Contributor(s): Jeffrey Lidz
Non-ARHU Contributor(s):

Laurel Perkins *19 (UCLA)

Dates:

The human ability to produce and understand an indefinite number of sentences is driven by syntax, a cognitive system that can combine a finite number of primitive linguistic elements to build arbitrarily complex expressions. The expressive power of syntax comes in part from its ability to encode potentially unbounded dependencies over abstract structural configurations. How does such a system develop in human minds? We show that 18-mo-old infants are capable of representing abstract nonlocal dependencies, suggesting that a core property of syntax emerges early in development. Our test case is English wh-questions, in which a fronted wh-phrase can act as the argument of a verb at a distance (e.g., What did the chef burn?). Whereas prior work has focused on infants’ interpretations of these questions, we introduce a test to probe their underlying syntactic representations, independent of meaning. We ask when infants know that an object wh-phrase and a local object of a verb cannot co-occur because they both express the same argument relation (e.g., *What did the chef burn the pizza). We find that 1) 18 mo olds demonstrate awareness of this complementary distribution pattern and thus represent the nonlocal grammatical dependency between the wh-phrase and the verb, but 2) younger infants do not. These results suggest that the second year of life is a period of active syntactic development, during which the computational capacities for representing nonlocal syntactic dependencies become evident.

Read More about Eighteen-month-old infants represent nonlocal syntactic dependencies

The mental representation of universal quantifers

On the psychological representations that give the meanings of "every" and "each".

Linguistics

Contributor(s): Jeffrey Lidz, Paul Pietroski
Non-ARHU Contributor(s):

Tyler Knowlton *21, Justin Halberda (Hopkins)

Dates:
Publisher: Springer
PhD student Tyler Knowlton smiling at the camera, surrounded by six members of his PhD committee, one joining remotely through an iPad

A sentence like every circle is blue might be understood in terms of individuals and their properties (e.g., for each thing that is a circle, it is blue) or in terms of a relation between groups (e.g., the blue things include the circles). Relatedly, theorists can specify the contents of universally quantified sentences in first-order or second-order terms. We offer new evidence that this logical first-order vs. second-order distinction corresponds to a psychologically robust individual vs. group distinction that has behavioral repercussions. Participants were shown displays of dots and asked to evaluate sentences with eachevery, or all combined with a predicate (e.g., big dot). We find that participants are better at estimating how many things the predicate applied to after evaluating sentences in which universal quantification is indicated with every or all, as opposed to each. We argue that every and all are understood in second-order terms that encourage group representation, while each is understood in first-order terms that encourage individual representation. Since the sentences that participants evaluate are truth-conditionally equivalent, our results also bear on questions concerning how meanings are related to truth-conditions.

Read More about The mental representation of universal quantifers

Comment on “Nonadjacent dependency processing in monkeys, apes, and humans”

Auditory pattern recognition in nonhuman animals shares important characteristics with human phonology, but not human syntax.

Linguistics

Contributor(s): William Idsardi
Non-ARHU Contributor(s):

Jonathan Rawsi (Stony Brook), Jeffrey Heinz (Stony Brook)

Dates:
Publisher: American Academy for the Advancement of Sciences

We comment on the technical interpretation of the study of Watson et al. and caution against their conclusion that the behavioral evidence in their experiments points to nonhuman animals’ ability to learn syntactic dependencies, because their results are also consistent with the learning of phonological dependencies in human languages.

Read More about Comment on “Nonadjacent dependency processing in monkeys, apes, and humans”

Automated Topic Model Evaluation Broken? The Incoherence of Coherence

Questioning automatic coherence evaluations for neural topic models.

Linguistics

Contributor(s): Philip Resnik
Non-ARHU Contributor(s):

Alexander Hoyle, Pranav Goel, Denis Peskov, Andrew Hian-Cheong, Jordan Boyd-Graber

Dates:

Topic model evaluation, like evaluation of other unsupervised methods, can be contentious. However, the field has coalesced around automated estimates of topic coherence, which rely on the frequency of word co-occurrences in a reference corpus. Recent models relying on neural components surpass classical topic models according to these metrics. At the same time, unlike classical models, the practice of neural topic model evaluation suffers from a validation gap: automatic coherence for neural models has not been validated using human experimentation. In addition, as we show via a meta-analysis of topic modeling literature, there is a substantial standardization gap in the use of automated topic modeling benchmarks. We address both the standardization gap and the validation gap. Using two of the most widely used topic model evaluation datasets, we assess a dominant classical model and two state-of-the-art neural models in a systematic, clearly documented, reproducible way. We use automatic coherence along with the two most widely accepted human judgment tasks, namely, topic rating and word intrusion. Automated evaluation will declare one model significantly different from another when corresponding human evaluations do not, calling into question the validity of fully automatic evaluations independent of human judgments.

Read More about Automated Topic Model Evaluation Broken? The Incoherence of Coherence

Informativity, topicality, and speech cost: comparing models of speakers’ choices of referring expressions

Is use of a pronoun motivated by topicality or efficiency?

Linguistics

Contributor(s): Naomi Feldman
Non-ARHU Contributor(s):

Naho Orita *15 (Tokyo University of Science)

Dates:

This study formalizes and compares two major hypotheses in speakers’ choices of referring expressions: the topicality model that chooses a form based on the topicality of the referent, and the rational model that chooses a form based on the informativity of the form and its speech cost. Simulations suggest that both the topicality of the referent and the informativity of the word are important to consider in speakers’ choices of reference forms, while a speech cost metric that prefers shorter forms may not be.

Read More about Informativity, topicality, and speech cost: comparing models of speakers’ choices of referring expressions