File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Mapping between fMRI responses to movies and their natural language annotations

TitleMapping between fMRI responses to movies and their natural language annotations
Authors
KeywordsFMRI
Multi-modal model
Natural language processing
Natural movie stimulus
Shared response model
Text annotations
Issue Date2018
Citation
NeuroImage, 2018, v. 180, p. 223-231 How to Cite?
AbstractSeveral research groups have shown how to map fMRI responses to the meanings of presented stimuli. This paper presents new methods for doing so when only a natural language annotation is available as the description of the stimulus. We study fMRI data gathered from subjects watching an episode of BBCs Sherlock (Chen et al., 2017), and learn bidirectional mappings between fMRI responses and natural language representations. By leveraging data from multiple subjects watching the same movie, we were able to perform scene classification with 72% accuracy (random guessing would give 4%) and scene ranking with average rank in the top 4% (random guessing would give 50%). The key ingredients underlying this high level of performance are (a) the use of the Shared Response Model (SRM) and its variant SRM-ICA (Chen et al., 2015; Zhang et al., 2016) to aggregate fMRI data from multiple subjects, both of which are shown to be superior to standard PCA in producing low-dimensional representations for the tasks in this paper; (b) a sentence embedding technique adapted from the natural language processing (NLP) literature (Arora et al., 2017) that produces semantic vector representation of the annotations; (c) using previous timestep information in the featurization of the predictor data. These optimizations in how we featurize the fMRI data and text annotations provide a substantial improvement in classification performance, relative to standard approaches.
Persistent Identifierhttp://hdl.handle.net/10722/341211
ISSN
2023 Impact Factor: 4.7
2023 SCImago Journal Rankings: 2.436
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorVodrahalli, Kiran-
dc.contributor.authorChen, Po Hsuan-
dc.contributor.authorLiang, Yingyu-
dc.contributor.authorBaldassano, Christopher-
dc.contributor.authorChen, Janice-
dc.contributor.authorYong, Esther-
dc.contributor.authorHoney, Christopher-
dc.contributor.authorHasson, Uri-
dc.contributor.authorRamadge, Peter-
dc.contributor.authorNorman, Kenneth A.-
dc.contributor.authorArora, Sanjeev-
dc.date.accessioned2024-03-13T08:41:02Z-
dc.date.available2024-03-13T08:41:02Z-
dc.date.issued2018-
dc.identifier.citationNeuroImage, 2018, v. 180, p. 223-231-
dc.identifier.issn1053-8119-
dc.identifier.urihttp://hdl.handle.net/10722/341211-
dc.description.abstractSeveral research groups have shown how to map fMRI responses to the meanings of presented stimuli. This paper presents new methods for doing so when only a natural language annotation is available as the description of the stimulus. We study fMRI data gathered from subjects watching an episode of BBCs Sherlock (Chen et al., 2017), and learn bidirectional mappings between fMRI responses and natural language representations. By leveraging data from multiple subjects watching the same movie, we were able to perform scene classification with 72% accuracy (random guessing would give 4%) and scene ranking with average rank in the top 4% (random guessing would give 50%). The key ingredients underlying this high level of performance are (a) the use of the Shared Response Model (SRM) and its variant SRM-ICA (Chen et al., 2015; Zhang et al., 2016) to aggregate fMRI data from multiple subjects, both of which are shown to be superior to standard PCA in producing low-dimensional representations for the tasks in this paper; (b) a sentence embedding technique adapted from the natural language processing (NLP) literature (Arora et al., 2017) that produces semantic vector representation of the annotations; (c) using previous timestep information in the featurization of the predictor data. These optimizations in how we featurize the fMRI data and text annotations provide a substantial improvement in classification performance, relative to standard approaches.-
dc.languageeng-
dc.relation.ispartofNeuroImage-
dc.subjectFMRI-
dc.subjectMulti-modal model-
dc.subjectNatural language processing-
dc.subjectNatural movie stimulus-
dc.subjectShared response model-
dc.subjectText annotations-
dc.titleMapping between fMRI responses to movies and their natural language annotations-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1016/j.neuroimage.2017.06.042-
dc.identifier.pmid28648889-
dc.identifier.scopuseid_2-s2.0-85023200931-
dc.identifier.volume180-
dc.identifier.spage223-
dc.identifier.epage231-
dc.identifier.eissn1095-9572-
dc.identifier.isiWOS:000443268900022-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats