MENU
April 17th (Thu), 9:40-10:40
Liina Pylkkänen is a Professor of Linguistics and Psychology at New York University. She studies the combinatory potential of human language. Pylkkanen's research aims to characterize the representational and processing properties of this combinatory system and investigate the question what are its basic computational units and how do they structurally combine. After completing a substantial body of theoretical work addressing the syntax-semantics interface for a particular subdomain of grammar (the verb phrase), she turned her research focus to characterizing the brain mechanisms responsible for the semantic combinatorics of language . The operations by which our brains build complex meanings from simpler pieces are intimately intertwined with computations building complex syntactic structures.Recently, she combines semantic theories and computational work with magnetoencephalography and functional magnetic resonance imaging.
Website
Language can be expressed and received through various modalities, such as sound, sight, and touch. Each modality has its own temporal dynamics, including different degrees of serialism and parallelism. In the neurobiology of language, we are limited by the fact that for each study, we must choose some specific modality. This hinders our ability to discern whether the observed results stem from properties of language or modality-specific dynamics. Comparing modalities directly is a slow way to uncover any inherent modality independent organization within the language system. In this talk, I discuss new work in which we approach this question from a different angle and ask: How does the language system organize itself when the input lacks temporal sequencing, allowing the brain to order computations in whatever way is natural? That organization should be a useful window into the brain’s inherent way to travel from form to meaning. We present short written sentences all at once, quickly enough to eliminate eye movements, in so-called rapid parallel visual presentation (RPVP). Psycholinguistic work has shown that full sentences can elicit syntactic and semantic effects in RPVP even if flashed for only 200ms (Snell & Grainger, 2017). We use magnetoencephalography (MEG) to reveal the neural organization of syntactic and semantic computations for parallel visual language input. Our initial results show that in a simple matching task, grammatical sentences in RPVP drive increased neural activity in left fronto-temporal cortex starting at ~130ms after sentence onset as compared to word lists. By varying the linguistic properties of the sentences, we aim to characterize the spatiotemporal organization of syntactic and semantic computations for a stimulus that does not itself impose any order. Further, since a parallel stimulus does not unfold over time, it gives us a measure of how input is mapped to knowledge in the absence of predictions from a temporally preceding context.
<< Back to "Speakers"