<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1d3 20150301//EN" "http://jats.nlm.nih.gov/publishing/1.1d3/JATS-journalpublishing1.dtd">
<article article-type="research-article" dtd-version="1.1d3" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">PLoS Comput Biol</journal-id>
<journal-id journal-id-type="publisher-id">plos</journal-id>
<journal-id journal-id-type="pmc">ploscomp</journal-id>
<journal-title-group>
<journal-title>PLOS Computational Biology</journal-title>
</journal-title-group>
<issn pub-type="ppub">1553-734X</issn>
<issn pub-type="epub">1553-7358</issn>
<publisher>
<publisher-name>Public Library of Science</publisher-name>
<publisher-loc>San Francisco, CA USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138</article-id>
<article-id pub-id-type="publisher-id">PCOMPBIOL-D-20-01475</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Research Article</subject>
</subj-group>
<subj-group subj-group-type="Discipline-v3">
<subject>Social sciences</subject><subj-group><subject>Linguistics</subject><subj-group><subject>Semantics</subject></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Physical sciences</subject><subj-group><subject>Mathematics</subject><subj-group><subject>Algebra</subject><subj-group><subject>Linear algebra</subject><subj-group><subject>Vector spaces</subject></subj-group></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Social sciences</subject><subj-group><subject>Linguistics</subject><subj-group><subject>Semantics</subject><subj-group><subject>Conceptual semantics</subject></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Brain mapping</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Medicine and health sciences</subject><subj-group><subject>Diagnostic medicine</subject><subj-group><subject>Diagnostic radiology</subject><subj-group><subject>Magnetic resonance imaging</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Research and analysis methods</subject><subj-group><subject>Imaging techniques</subject><subj-group><subject>Diagnostic radiology</subject><subj-group><subject>Magnetic resonance imaging</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Medicine and health sciences</subject><subj-group><subject>Radiology and imaging</subject><subj-group><subject>Diagnostic radiology</subject><subj-group><subject>Magnetic resonance imaging</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Research and analysis methods</subject><subj-group><subject>Imaging techniques</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Functional magnetic resonance imaging</subject></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Biology and life sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Experimental psychology</subject></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Social sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Experimental psychology</subject></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Physical sciences</subject><subj-group><subject>Mathematics</subject><subj-group><subject>Discrete mathematics</subject><subj-group><subject>Combinatorics</subject><subj-group><subject>Permutation</subject></subj-group></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Engineering and technology</subject><subj-group><subject>Structural engineering</subject><subj-group><subject>Built structures</subject></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Research and analysis methods</subject><subj-group><subject>Mathematical and statistical techniques</subject><subj-group><subject>Statistical methods</subject><subj-group><subject>Forecasting</subject></subj-group></subj-group></subj-group></subj-group><subj-group subj-group-type="Discipline-v3">
<subject>Physical sciences</subject><subj-group><subject>Mathematics</subject><subj-group><subject>Statistics</subject><subj-group><subject>Statistical methods</subject><subj-group><subject>Forecasting</subject></subj-group></subj-group></subj-group></subj-group></subj-group></article-categories>
<title-group>
<article-title>Behavioral correlates of cortical semantic representations modeled by word vectors</article-title>
<alt-title alt-title-type="running-head">Behavioral correlates of modeled cortical semantic representations</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes" xlink:type="simple">
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0003-0555-518X</contrib-id>
<name name-style="western">
<surname>Nishida</surname>
<given-names>Satoshi</given-names>
</name>
<role content-type="https://casrai.org/credit/">Conceptualization</role>
<role content-type="https://casrai.org/credit/">Data curation</role>
<role content-type="https://casrai.org/credit/">Formal analysis</role>
<role content-type="https://casrai.org/credit/">Funding acquisition</role>
<role content-type="https://casrai.org/credit/">Investigation</role>
<role content-type="https://casrai.org/credit/">Methodology</role>
<role content-type="https://casrai.org/credit/">Project administration</role>
<role content-type="https://casrai.org/credit/">Resources</role>
<role content-type="https://casrai.org/credit/">Software</role>
<role content-type="https://casrai.org/credit/">Supervision</role>
<role content-type="https://casrai.org/credit/">Validation</role>
<role content-type="https://casrai.org/credit/">Visualization</role>
<role content-type="https://casrai.org/credit/">Writing – original draft</role>
<role content-type="https://casrai.org/credit/">Writing – review &amp; editing</role>
<xref ref-type="aff" rid="aff001"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff002"><sup>2</sup></xref>
<xref ref-type="corresp" rid="cor001">*</xref>
</contrib>
<contrib contrib-type="author" xlink:type="simple">
<name name-style="western">
<surname>Blanc</surname>
<given-names>Antoine</given-names>
</name>
<role content-type="https://casrai.org/credit/">Data curation</role>
<role content-type="https://casrai.org/credit/">Formal analysis</role>
<role content-type="https://casrai.org/credit/">Methodology</role>
<role content-type="https://casrai.org/credit/">Software</role>
<role content-type="https://casrai.org/credit/">Writing – review &amp; editing</role>
<xref ref-type="aff" rid="aff001"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author" xlink:type="simple">
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0001-9491-4433</contrib-id>
<name name-style="western">
<surname>Maeda</surname>
<given-names>Naoya</given-names>
</name>
<role content-type="https://casrai.org/credit/">Data curation</role>
<role content-type="https://casrai.org/credit/">Resources</role>
<role content-type="https://casrai.org/credit/">Writing – review &amp; editing</role>
<xref ref-type="aff" rid="aff003"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author" xlink:type="simple">
<name name-style="western">
<surname>Kado</surname>
<given-names>Masataka</given-names>
</name>
<role content-type="https://casrai.org/credit/">Data curation</role>
<role content-type="https://casrai.org/credit/">Resources</role>
<role content-type="https://casrai.org/credit/">Writing – review &amp; editing</role>
<xref ref-type="aff" rid="aff003"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author" xlink:type="simple">
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0001-8015-340X</contrib-id>
<name name-style="western">
<surname>Nishimoto</surname>
<given-names>Shinji</given-names>
</name>
<role content-type="https://casrai.org/credit/">Funding acquisition</role>
<role content-type="https://casrai.org/credit/">Supervision</role>
<role content-type="https://casrai.org/credit/">Writing – review &amp; editing</role>
<xref ref-type="aff" rid="aff001"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff002"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff004"><sup>4</sup></xref>
</contrib>
</contrib-group>
<aff id="aff001"><label>1</label> <addr-line>Center for Information and Neural Networks (CiNet), National Institute of Information and Communications Technology (NICT), Suita, Osaka, Japan</addr-line></aff>
<aff id="aff002"><label>2</label> <addr-line>Graduate School of Frontier Biosciences, Osaka University, Suita, Osaka, Japan</addr-line></aff>
<aff id="aff003"><label>3</label> <addr-line>NTT DATA Corporation, Tokyo, Japan</addr-line></aff>
<aff id="aff004"><label>4</label> <addr-line>Graduate School of Medicine, Osaka University, Suita, Osaka, Japan</addr-line></aff>
<contrib-group>
<contrib contrib-type="editor" xlink:type="simple">
<name name-style="western">
<surname>Cai</surname>
<given-names>Ming Bo</given-names>
</name>
<role>Editor</role>
<xref ref-type="aff" rid="edit1"/>
</contrib>
</contrib-group>
<aff id="edit1"><addr-line>Tokyo Daigaku, JAPAN</addr-line></aff>
<author-notes>
<fn fn-type="conflict" id="coi001">
<p>I have read the journal’s policy and the authors of this manuscript have the following competing interests: This study was funded by NTT Data Corp. NM and MK are employees of NTT Data Corp.</p>
</fn>
<corresp id="cor001">* E-mail: <email xlink:type="simple">s-nishida@nict.go.jp</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>23</day>
<month>6</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<month>6</month>
<year>2021</year>
</pub-date>
<volume>17</volume>
<issue>6</issue>
<elocation-id>e1009138</elocation-id>
<history>
<date date-type="received">
<day>17</day>
<month>8</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>1</day>
<month>6</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-year>2021</copyright-year>
<copyright-holder>Nishida et al</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="info:doi/10.1371/journal.pcbi.1009138"/>
<abstract>
<p>The quantitative modeling of semantic representations in the brain plays a key role in understanding the neural basis of semantic processing. Previous studies have demonstrated that word vectors, which were originally developed for use in the field of natural language processing, provide a powerful tool for such quantitative modeling. However, whether semantic representations in the brain revealed by the word vector-based models actually capture our perception of semantic information remains unclear, as there has been no study explicitly examining the behavioral correlates of the modeled brain semantic representations. To address this issue, we compared the semantic structure of nouns and adjectives in the brain estimated from word vector-based brain models with that evaluated from human behavior. The brain models were constructed using voxelwise modeling to predict the functional magnetic resonance imaging (fMRI) response to natural movies from semantic contents in each movie scene through a word vector space. The semantic dissimilarity of brain word representations was then evaluated using the brain models. Meanwhile, data on human behavior reflecting the perception of semantic dissimilarity between words were collected in psychological experiments. We found a significant correlation between brain model- and behavior-derived semantic dissimilarities of words. This finding suggests that semantic representations in the brain modeled via word vectors appropriately capture our perception of word meanings.</p>
</abstract>
<abstract abstract-type="summary">
<title>Author summary</title>
<p>Word vectors, which have been originally developed in the field of engineering (natural language processing), have been extensively leveraged in neuroscience studies to model semantic representations in the human brain. These studies have attempted to model brain semantic representations by associating them with the meanings of thousands of words via a word vector space. However, there has been no study explicitly examining whether the brain semantic representations modeled by word vectors actually capture our perception of semantic information. To address this issue, we compared the semantic representational structure of words in the brain estimated from word vector-based brain models with that evaluated from behavioral data in psychological experiments. The results revealed a significant correlation between these model- and behavior-derived semantic representational structures of words. This indicates that the brain semantic representations modeled using word vectors actually reflect the human perception of word meanings. Our findings contribute to the establishment of word vector-based brain modeling as a useful tool in studying human semantic processing.</p>
</abstract>
<funding-group>
<award-group id="award001">
<funding-source>
<institution-wrap>
<institution-id institution-id-type="funder-id">http://dx.doi.org/10.13039/501100001691</institution-id>
<institution>Japan Society for the Promotion of Science</institution>
</institution-wrap>
</funding-source>
<award-id>18K18141</award-id>
<principal-award-recipient>
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0003-0555-518X</contrib-id>
<name name-style="western">
<surname>Nishida</surname>
<given-names>Satoshi</given-names>
</name>
</principal-award-recipient>
</award-group>
<award-group id="award002">
<funding-source>
<institution-wrap>
<institution-id institution-id-type="funder-id">http://dx.doi.org/10.13039/501100001691</institution-id>
<institution>Japan Society for the Promotion of Science</institution>
</institution-wrap>
</funding-source>
<award-id>15H05311</award-id>
<principal-award-recipient>
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0001-8015-340X</contrib-id>
<name name-style="western">
<surname>Nishimoto</surname>
<given-names>Shinji</given-names>
</name>
</principal-award-recipient>
</award-group>
<award-group id="award003">
<funding-source>
<institution>Precursory Research for Embryonic Science and Technology (JP)</institution>
</funding-source>
<award-id>JPMJPR20C6</award-id>
<principal-award-recipient>
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0003-0555-518X</contrib-id>
<name name-style="western">
<surname>Nishida</surname>
<given-names>Satoshi</given-names>
</name>
</principal-award-recipient>
</award-group>
<award-group id="award004">
<funding-source>
<institution-wrap>
<institution-id institution-id-type="funder-id">http://dx.doi.org/10.13039/501100009024</institution-id>
<institution>Exploratory Research for Advanced Technology</institution>
</institution-wrap>
</funding-source>
<award-id>JPMJER1801</award-id>
<principal-award-recipient>
<contrib-id authenticated="true" contrib-id-type="orcid">https://orcid.org/0000-0001-8015-340X</contrib-id>
<name name-style="western">
<surname>Nishimoto</surname>
<given-names>Shinji</given-names>
</name>
</principal-award-recipient>
</award-group>
<funding-statement>The work was supported by Japan Society for the Promotion of Science (<ext-link ext-link-type="uri" xlink:href="https://www.jsps.go.jp/english/" xlink:type="simple">https://www.jsps.go.jp/english/</ext-link>) KAKENHI Grant-in-Aid for Early-Career Scientists (18K18141) to S.N. and for Young Scientists A (15H05311) to S.N., and Japan Science Technology agency (<ext-link ext-link-type="uri" xlink:href="https://www.jst.go.jp/EN/" xlink:type="simple">https://www.jst.go.jp/EN/</ext-link>) PRESTO (JPMJPR20C6) to S.N. and ERATO (JPMJER1801) to S.N.. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.</funding-statement>
</funding-group>
<counts>
<fig-count count="10"/>
<table-count count="3"/>
<page-count count="35"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>PLOS Publication Stage</meta-name>
<meta-value>vor-update-to-uncorrected-proof</meta-value>
</custom-meta>
<custom-meta>
<meta-name>Publication Update</meta-name>
<meta-value>2021-07-06</meta-value>
</custom-meta>
<custom-meta id="data-availability">
<meta-name>Data Availability</meta-name>
<meta-value>The data underlying the results presented in the study are available from GitHub (<ext-link ext-link-type="uri" xlink:href="https://github.com/s-nishida/behav_corr_sem_model" xlink:type="simple">https://github.com/s-nishida/behav_corr_sem_model</ext-link>). The raw behavioral data collected in the word-arrangement task are available from Open Science Framework (<ext-link ext-link-type="uri" xlink:href="https://osf.io/um3qg/" xlink:type="simple">https://osf.io/um3qg/</ext-link>).</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="sec001" sec-type="intro">
<title>Introduction</title>
<p>Natural language processing is a branch of machine learning that aims to develop machines that understand the meanings of words. In the field of natural language processing, a number of algorithms have been developed to capture the semantic representations of words from word statistics in large-scale text data as word vectors [<xref ref-type="bibr" rid="pcbi.1009138.ref001">1</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref005">5</xref>]. The word vectors obtained using these algorithms have effectively captured the latent semantic structure of words and further performed various types of natural language tasks, such as word similarity judgment [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref006">6</xref>], sentiment analysis [<xref ref-type="bibr" rid="pcbi.1009138.ref007">7</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref008">8</xref>], and question answering [<xref ref-type="bibr" rid="pcbi.1009138.ref009">9</xref>].</p>
<p>Furthermore, word vectors can be also used in neuroimaging studies to model semantic representations in the brain [<xref ref-type="bibr" rid="pcbi.1009138.ref010">10</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref017">17</xref>]. These studies have reported that word vector-based models have the ability to predict the brain response evoked by semantic perceptual experiences [<xref ref-type="bibr" rid="pcbi.1009138.ref010">10</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref012">12</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref016">16</xref>]. These models are also able to recover perceived semantic contents from brain response [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref017">17</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref018">18</xref>]. These findings suggest that word vectors capture at least some aspects of the semantic representations in the brain. However, whether the brain semantic representations modeled by word vectors accurately reflect the semantic perception of humans is yet to be determined. In other words, no study has yet identified the behavioral correlates of the modeled brain semantic representations. This clarification is important in order to establish the brain modeling with word vectors as an accurate methodology for investigating human semantic processing.</p>
<p>To examine the behavioral correlates of the brain semantic representations modeled by word vectors, we compared the semantic representational structure estimated by word vector-based brain models with that evaluated from human behavior. The estimation of the semantic representational structure in the brain was performed using voxelwise modeling with a word vector space of fastText [<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>], GloVe [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>], or word2vec [<xref ref-type="bibr" rid="pcbi.1009138.ref002">2</xref>] (<xref ref-type="fig" rid="pcbi.1009138.g001">Fig 1A</xref>). For this purpose, we conducted two sets of functional magnetic resonance imaging (fMRI) experiments in which 52 participants were asked to view natural movies in the scanner; among them, 36 participated in one or the other of these experiments, and 16 participated in both. The word vector-based voxelwise models predicted movie-evoked fMRI signals from the semantic contents in individual movie scenes in the word vector space. We then transformed original word vectors into brain representations using the model weights, and a brain-derived word dissimilarity matrix from these brain representations was obtained (<xref ref-type="fig" rid="pcbi.1009138.g001">Fig 1B</xref>).</p>
<fig id="pcbi.1009138.g001" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g001</object-id>
<label>Fig 1</label>
<caption>
<title>Voxelwise modeling and brain-derived word dissimilarity.</title>
<p><bold>A</bold>) Voxelwise modeling based on a word vector space. The model predicts fMRI signals evoked by natural movie scenes via a weighted linear summation of the vector representations of semantic descriptions of each scene (scene vectors). Scene vectors were obtained by transforming manual descriptions of each movie scene through a word vector space pretrained using statistical learning (fastText, GloVe, or word2vec) from a text corpus. The weights of the linear prediction model were trained using the corresponding time series of movies and fMRI signals of each brain. <bold>B</bold>) Estimation of brain-derived word dissimilarity. Word representations in the vector space were transformed into word representations in the modeled brain space by multiplying original word vectors by the model weights. Then, the correlation distance of the modeled word representations between all possible pairs of words was calculated, producing a brain-derived word dissimilarity matrix.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g001" xlink:type="simple"/>
</fig>
<p>Meanwhile, the behavior-derived semantic representational structure was obtained from behavioral data in a psychological task, in which participants separately arranged tens of words (nouns and adjectives) in a two-dimensional space as per their semantic relationship (<xref ref-type="fig" rid="pcbi.1009138.g002">Fig 2</xref>). This task was a modified version of a psychological task introduced previously [<xref ref-type="bibr" rid="pcbi.1009138.ref019">19</xref>]. A behavior-derived word dissimilarity matrix was then estimated using these behavioral data. Finally, we examined the correlation between the brain- and behavior-derived word dissimilarity matrices separately for both nouns and adjectives.</p>
<fig id="pcbi.1009138.g002" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g002</object-id>
<label>Fig 2</label>
<caption>
<title>Word-arrangement task and behavior-derived word dissimilarity.</title>
<p>To evaluate the word dissimilarity structure derived from human behavior, we conducted psychological experiments in which each participant performed a word-arrangement task. On each trial of this task, participants were required to arrange multiple (≤60) words in a two-dimensional space according to the semantic relationship of those words. After each participant completed ≤1 h of this task, a behavior-derived word dissimilarity matrix was established using inverse multidimensional scaling (see <xref ref-type="sec" rid="sec007">Methods</xref> for more details).</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g002" xlink:type="simple"/>
</fig>
</sec>
<sec id="sec002" sec-type="results">
<title>Results</title>
<sec id="sec003">
<title>Performance of voxelwise models based on word vectors</title>
<p>We first determined whether the voxelwise models based on a word vector space appropriately predict movie-evoked brain responses. The performance of brain-response prediction was evaluated using the following two measures: (1) prediction accuracy calculated as the correlation coefficients between predicted and measured brain responses in the test dataset, and (2) the fraction of significant voxels, among all cortical voxels, for which the prediction accuracy reached a significance threshold (p &lt; 0.05 after the correction for multiple comparisons using the false discovery rate [FDR]; for more details, see <xref ref-type="sec" rid="sec007">Methods</xref>). The prediction accuracy averaged across all cortical voxels and across participants for the fastText, GloVe, and word2vec vector spaces was 0.0928, 0.0955, and 0.0967, respectively, for movie set 1 (28 participants) and 0.0865, 0.0884, and 0.0891, respectively, for movie set 2 (40 participants); no participant showed mean prediction accuracy less than 0 (<xref ref-type="fig" rid="pcbi.1009138.g003">Fig 3A, 3C, 3E, 3G, 3I and 3K</xref>). This accuracy was sufficiently high because it was averaged over all cortical voxels, and previous studies on voxelwise modeling have reported a similar tendency [<xref ref-type="bibr" rid="pcbi.1009138.ref020">20</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref022">22</xref>]. The fraction of significant voxels averaged across participants for the fastText, GloVe, and word2vec vector spaces was 0.575, 0.589, and 0.594, respectively, for movie set 1 and 0.526, 0.541, and 0.543, respectively, for movie set 2 (<xref ref-type="fig" rid="pcbi.1009138.g003">Fig 3B, 3D, 3F, 3H, 3J and 3I</xref>). These results indicate that the voxelwise models trained showed sufficient performance in the modeling of semantic representations consistently across different stimulus sets.</p>
<fig id="pcbi.1009138.g003" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g003</object-id>
<label>Fig 3</label>
<caption>
<title>Performance of voxelwise models in brain-response prediction.</title>
<p>The performance of voxelwise models (vector dimension = 1000) was evaluated in terms of predicting brain responses in the test dataset. For this purpose, prediction accuracy and the fraction of significant voxels for each brain were calculated for each type of word vectors (<bold>A</bold>–<bold>D</bold>, fastText; <bold>E</bold>–<bold>H</bold>, GloVe; <bold>I</bold>–<bold>L</bold>, word2vec). The distribution of prediction accuracy (<bold>A</bold>, <bold>C</bold>, <bold>E</bold>, <bold>G</bold>, <bold>I</bold>, and <bold>K</bold>) and that of the fraction of significant voxels (<bold>B</bold>, <bold>D</bold>, <bold>F</bold>, <bold>H</bold>, <bold>J</bold>, and <bold>L</bold>) were separately shown for movie sets 1 (<bold>A</bold>–<bold>B</bold>, <bold>E</bold>–<bold>F</bold>, and <bold>I</bold>–<bold>J</bold>) and 2 (<bold>C</bold>–<bold>D</bold>, <bold>G</bold>–<bold>H</bold>, and <bold>K</bold>–<bold>L</bold>). The vertical dashed line in each panel indicates the prediction accuracy averaged across participants.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g003" xlink:type="simple"/>
</fig>
<p>In the comparison of prediction performance between different types of word vectors, word2vec vectors showed significantly higher accuracy than the other vectors for both movie sets (Wilcoxon test, p &lt; 0.05, FDR corrected) while GloVe vectors showed significantly higher accuracy than fastText vectors for both movie sets (p &lt; 0.0001, FDR corrected). In the comparison of the fraction of significant voxels, word2vec and GloVe vectors showed significantly higher fractions than fastText vectors for both movie sets (p &lt; 0.0001, FDR corrected) while the differences between word2vec and GloVe vectors were not significant for both movie sets (p &gt; 0.06, FDR corrected). Nonetheless, the differences of the prediction performance between these vectors were totally small (&lt;0.004 for mean prediction accuracy and &lt;0.02 for the mean fraction of significant voxels).</p>
<p>Although the vector dimensionality of the word vector spaces had little effect on the model performance, there was no clear tendency toward vector-dimensionality dependency. The change of model performance was not monotonic against the change of vector dimensions (Kendall rank correlation, τ = −0.071–0.708, p &gt; 0.17) and was inconsistent across movie sets and different vector types (<xref ref-type="supplementary-material" rid="pcbi.1009138.s001">S1</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s003">S3</xref> Figs).</p>
<p>Previous studies performed voxelwise modeling using simpler, discrete word features, such as binary labels of objects and actions occurring in movie scenes (e.g., [<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]). However, because such discrete word features have been examined separately from word vectors in voxelwise modeling, it is still unclear whether word vectors are more effective to model semantic representations in the brain than discrete word features. To address this, we compared these two types of semantic features, first of all, in terms of the performance of brain-response prediction. For this purpose, we constructed a voxelwise model using the binary labeling of semantic contents in each movie scene (binary-labeling model). Although this model was similar to one used in a previous study [<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>], we obtained binary labels of semantic contents by extracting words from the manual scene descriptions rather than manual word labeling performed in the previous study (for more details, see <xref ref-type="sec" rid="sec007">Methods</xref>). In this model, a semantic feature in a given scene was represented by a binary vector with 100–2000 dimension that corresponded to the occurrence of 100–2000 specific words in manual descriptions in that scene.</p>
<p>We found that the binary-labeling model constructed for each participant showed sufficiently high performance in brain-response prediction especially when high-dimensional vectors were used (<xref ref-type="supplementary-material" rid="pcbi.1009138.s004">S4 Fig</xref>). For 1000 and 2000 vector dimensions, the performance of the binary-labeling model was rather significantly higher than that of all the word vector-based models (Wilcoxon test, p &lt; 0.05, FDR corrected); nonetheless, the performance differences between models were not large (&lt;0.006 for mean prediction accuracy and &lt;0.05 for the mean fraction of significant voxels). This result suggests that brain response to semantic information in natural scenes can be effectively predicted even using discrete word features.</p>
<p>Does this finding imply that the semantic relational structure of words, captured by word vectors [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref006">6</xref>], is ineffective in the modeling of cortical semantic representations? Comparing the word vector-based and binary-labeling model, however, cannot purely assess the effectiveness of semantic relational structure of word vectors because the procedures to transform scene descriptions to semantic features were different between these models. To examine the effectiveness more fairly, we compared the performance of brain-response prediction between the voxelwise model based on original, trained word vectors and that based on untrained word vectors. The same procedure as used for original word vectors was employed to obtain the untrained word vectors, except that the statistical training of word vectors was omitted. This procedure yielded 1000-dimensional random vectors, each assigned to each word in the vocabulary shared with original word vectors. Hence, the untrained vectors still had signatures of individual words but not the semantic relational structure of words at all. Then, the voxelwise model based on the untrained word vectors were constructed using the same procedure as used for the voxelwise model based on original, trained word vectors (see also <xref ref-type="sec" rid="sec007">Methods</xref>).</p>
<p>The performance of brain-response prediction (i.e., prediction accuracy and the fraction of significant voxels) was compared between the voxelwise model based on trained vectors and that based on untrained vectors. We found that the prediction performance of voxelwise models was consistently higher for trained vectors than for untrained vectors regardless of performance measures, word-vector types, and movie sets (<xref ref-type="supplementary-material" rid="pcbi.1009138.s005">S5 Fig</xref>). This result suggests that the semantic relational structure of words, captured by trained word vectors, is effective in the prediction of movie-evoked brain response, namely, in the modeling of cortical semantic representations.</p>
</sec>
<sec id="sec004">
<title>Localization of cortical regions highly predicted by voxelwise models</title>
<p>We next identified which cortical regions were predictable by the word vector-based models in order to determine that the models could capture the semantic representations in cortical regions which are considered to be involved in audiovisual semantic processing. For this purpose, the prediction accuracy averaged across participants was calculated in each of the 148 cortical regions that were anatomically segmented using FreeSurfer [<xref ref-type="bibr" rid="pcbi.1009138.ref024">24</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref025">25</xref>] on the basis of the Destrieux atlas [<xref ref-type="bibr" rid="pcbi.1009138.ref026">26</xref>]. Then, the averaged prediction accuracy for each region was mapped onto the cortical surface of a reference brain (<xref ref-type="fig" rid="pcbi.1009138.g004">Fig 4</xref> and <xref ref-type="table" rid="pcbi.1009138.t001">Table 1</xref>). The highly predictable regions for each word vector-based model were localized over widespread cortical regions, including the occipital, superior and inferior temporal, and posterior parietal regions. This localization was consistent with previous reports, in which the semantic representations in the brain were modeled with word features [<xref ref-type="bibr" rid="pcbi.1009138.ref012">12</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]. There was a strong correlation of mean prediction accuracy over 148 cortical regions between different types of word vectors (Pearson’s r &gt; 0.999; Spearman’s ρ &gt; 0.999) and between movie sets 1 and 2 (Pearson’s r &gt; 0.972; Spearman’s ρ &gt; 0.973); this indicates the consistency of model predictability of these cortical regions across vector types and stimulus sets. In addition, there were similar tendencies across vector dimensionality (<xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s008">S8</xref> Figs and <xref ref-type="supplementary-material" rid="pcbi.1009138.s028">S1</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s033">S6</xref> Tables). These results indicate that the word vector-based voxelwise models trained reliably showed the localization of highly predictable regions.</p>
<fig id="pcbi.1009138.g004" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g004</object-id>
<label>Fig 4</label>
<caption>
<title>Cortical mapping of prediction accuracy.</title>
<p>Participant-averaged prediction accuracy of voxelwise models (vector dimension = 1000) was mapped onto the cortical surface of a reference brain for each of fastText (<bold>A</bold> and <bold>B</bold>), GloVe (<bold>C</bold> and <bold>D</bold>), and word2vec (<bold>E</bold> and <bold>F</bold>) vectors and for each of movie sets 1 (<bold>A</bold>, <bold>C</bold>, and <bold>E</bold>) and 2 (<bold>B</bold>, <bold>D</bold>, and <bold>F</bold>). The prediction accuracy was averaged within each of the cortical regions that were anatomically segmented. Brighter colors in the surface maps indicate cortical regions that have higher prediction accuracy. We showed only regions with mean prediction accuracy above 0.11, which reaches a significance level (i.e., p = 0.05) of prediction accuracy after Bonferroni correction for multiple comparisons among 148 cortical regions (i.e., p = 0.0001 ~ 0.05/148). The five cortical regions with the highest mean prediction accuracy are numbered in a descending order separately for each type of word vectors and for each movie set. The names of these regions are shown in <xref ref-type="table" rid="pcbi.1009138.t001">Table 1</xref>. LH, left hemisphere; RH, right hemisphere; A, anterior; P, posterior.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g004" xlink:type="simple"/>
</fig>
<table-wrap id="pcbi.1009138.t001" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.t001</object-id>
<label>Table 1</label> <caption><title>Cortical regions with the highest mean prediction accuracy for each type of word vectors and for each movie set.</title></caption>
<alternatives>
<graphic id="pcbi.1009138.t001g" mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.t001" xlink:type="simple"/>
<table>
<colgroup>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
</colgroup>
<thead>
<tr>
<th align="center" colspan="2"/>
<th align="center" style="background-color:#BFBFBF">Rank</th>
<th align="center" style="background-color:#BFBFBF">Region name</th>
<th align="center" style="background-color:#BFBFBF">Prediction accuracy</th>
</tr>
</thead>
<tbody>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>fastText</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.312</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.293</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.254</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.245</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.238</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.365</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.337</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.248</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.245</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Right middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.242</td>
</tr>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>GloVe</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.320</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.301</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.258</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.251</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.241</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.368</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.340</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.251</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.249</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.244</td>
</tr>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>Word2vec</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.319</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.302</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.260</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.252</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.243</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.367</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.339</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.251</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.249</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Right middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.244</td>
</tr>
</tbody>
</table>
</alternatives>
</table-wrap>
<p>These results showed that the mean prediction accuracy in each brain region ranged up to 0.37 across word-vector types and movie sets. It might be argued that the prediction accuracy observed in this study was deemed much lower than that observed in previous studies (up to 0.6–0.8) [<xref ref-type="bibr" rid="pcbi.1009138.ref021">21</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]. However, this discrepancy is likely to be explained by the difference in spatial granularity used for calculating prediction accuracy; namely, we showed brain region-wise prediction accuracy whereas the previous studies reported voxelwise prediction accuracy. At the voxel level, our voxelwise models, including binary-labeling models, for individual brains exhibited prediction accuracy ranging up to ~0.7 (<xref ref-type="supplementary-material" rid="pcbi.1009138.s009">S9</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s010">S10</xref> Figs), which is consistent with the results from the previous studies [<xref ref-type="bibr" rid="pcbi.1009138.ref021">21</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>].</p>
<p>A similar localization was also observed when the fraction of significant voxels in each cortical region was mapped onto the cortical surface (<xref ref-type="fig" rid="pcbi.1009138.g005">Fig 5</xref> and <xref ref-type="table" rid="pcbi.1009138.t002">Table 2</xref>). The fraction was relatively large in the occipital, superior and inferior temporal, and posterior parietal regions compared with the other regions. This localization pattern was observed to be highly consistent across vector types, movie sets, and vector dimensions (<xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s013">S13</xref> Figs and <xref ref-type="supplementary-material" rid="pcbi.1009138.s034">S7</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s039">S12</xref> Tables). Taken together, these results suggest that the word vector-based models capture the semantic representations in appropriate cortical regions.</p>
<fig id="pcbi.1009138.g005" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g005</object-id>
<label>Fig 5</label>
<caption>
<title>Cortical mapping of fraction of significant voxels.</title>
<p>Participant-averaged fraction of significant voxels of voxelwise models (vector dimension = 1000) mapped onto the cortical surface of a reference brain for each of fastText (<bold>A</bold> and <bold>B</bold>), GloVe (<bold>C</bold> and <bold>D</bold>), and word2vec (<bold>E</bold> and <bold>F</bold>) vectors and for each of movie sets 1 (<bold>A</bold>, <bold>C</bold>, and <bold>E</bold>) and 2 (<bold>B</bold>, <bold>D</bold>, and <bold>F</bold>). The fraction was computed within each cortical region. Brighter colors indicate regions that have larger fraction. The five cortical regions with the highest mean fraction of significant voxels are numbered in a descending order separately for each type of word vectors and for each movie set. The names of these regions are shown in <xref ref-type="table" rid="pcbi.1009138.t002">Table 2</xref>. Other conventions are the same as in <xref ref-type="fig" rid="pcbi.1009138.g004">Fig 4</xref>.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g005" xlink:type="simple"/>
</fig>
<table-wrap id="pcbi.1009138.t002" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.t002</object-id>
<label>Table 2</label> <caption><title>Cortical regions with the highest mean fraction of significant voxels for each type of word vectors and for each movie set.</title></caption>
<alternatives>
<graphic id="pcbi.1009138.t002g" mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.t002" xlink:type="simple"/>
<table>
<colgroup>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
</colgroup>
<thead>
<tr>
<th align="center" colspan="2"/>
<th align="center" style="background-color:#BFBFBF">Rank</th>
<th align="center" style="background-color:#BFBFBF">Region name</th>
<th align="center" style="background-color:#BFBFBF">Fraction of significant voxels</th>
</tr>
</thead>
<tbody>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>fastText</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.946</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.945</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.943</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.935</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.928</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.966</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.962</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.943</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Right middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.937</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.930</td>
</tr>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>GloVe</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.953</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.950</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.948</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.938</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.937</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.969</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.968</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.946</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Right middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.939</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.935</td>
</tr>
<tr>
<td align="center" rowspan="10" style="background-color:#BFBFBF"><bold>Word2vec</bold></td>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 1</bold></td>
<td align="center" style="background-color:#F2F2F2">1</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.952</td>
</tr>
<tr>
<td align="center">2</td>
<td align="left">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.950</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">3</td>
<td align="left" style="background-color:#F2F2F2">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.948</td>
</tr>
<tr>
<td align="center">4</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.940</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">5</td>
<td align="left" style="background-color:#F2F2F2">Right superior occipital sulcus and transverse occipital sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.937</td>
</tr>
<tr>
<td align="center" rowspan="5" style="background-color:#BFBFBF"><bold>Movie set 2</bold></td>
<td align="center">1</td>
<td align="left">Right anterior occipital sulcus and preoccipital notch</td>
<td align="center">0.969</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">2</td>
<td align="left" style="background-color:#F2F2F2">Left anterior occipital sulcus and preoccipital notch</td>
<td align="center" style="background-color:#F2F2F2">0.965</td>
</tr>
<tr>
<td align="center">3</td>
<td align="left">Left middle occipital sulcus and lunatus sulcus</td>
<td align="center">0.948</td>
</tr>
<tr>
<td align="center" style="background-color:#F2F2F2">4</td>
<td align="left" style="background-color:#F2F2F2">Right middle occipital sulcus and lunatus sulcus</td>
<td align="center" style="background-color:#F2F2F2">0.941</td>
</tr>
<tr>
<td align="center">5</td>
<td align="left">Left superior occipital sulcus and transverse occipital sulcus</td>
<td align="center">0.934</td>
</tr>
</tbody>
</table>
</alternatives>
</table-wrap>
<p>It should be noted that, however, the fraction of significant voxels was not small even in the other cortical regions, including the prefrontal cortex. The minimum fraction within any individual region was 0.237, which is significantly higher than 0.05 (corresponding to the significance level of 0.05; Wilcoxson test, p &lt; 0.00001). In addition, even when the chance level of the fraction for each region was estimated from control voxelwise models in which word vectors were shuffled across vector dimensions for each vector (see also <xref ref-type="sec" rid="sec007">Methods</xref>), the fraction of significant voxels was significantly above the chance level in all the cortical regions (Wilcoxson test, p &lt; 0.00001, FDR corrected; <xref ref-type="supplementary-material" rid="pcbi.1009138.s014">S14 Fig</xref>). These results suggest that although high prediction accuracy was observed in specific regions (Figs <xref ref-type="fig" rid="pcbi.1009138.g004">4</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s008">S8</xref>), the word vector-based model can potentially capture semantic information even from other regions across the cortex.</p>
<p>In comparison between word vectors and discreate word features, we observed similar localization patterns of prediction accuracy (<xref ref-type="supplementary-material" rid="pcbi.1009138.s015">S15 Fig</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s040">S13 Table</xref>) and the fraction of significant voxels (<xref ref-type="supplementary-material" rid="pcbi.1009138.s016">S16 Fig</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s041">S14 Table</xref>). Therefore, even regarding the localization of predictable cortical regions, we did not find any clear differences between word vectors and discrete word features.</p>
<p>To assess the effectiveness of the sematic relational structure of word vectors in response prediction in each cortical region, we compared prediction performance (i.e., prediction performance and the fraction of significant voxels) between the voxelwise model based on trained word vectors and that based on untrained word vectors within each region. The results reveal that the model based on trained vectors exhibited significantly higher response-prediction performance in the majority of the 148 regions (<xref ref-type="supplementary-material" rid="pcbi.1009138.s017">S17 Fig</xref>; 130–144 regions for prediction accuracy; 136–148 regions for the fraction of significant voxels; Wilcoxon test, p &lt; 0.05, FDR corrected). Even across regions, the model based on trained vectors outperformed the model based on untrained vectors regardless of performance measures, word-vector types, and datasets (Wilcoxon test, p &lt; 0.05, FDR corrected). Although performance differences between these two models were not large, the improvement of the performance was observed comparably across regions rather than conspicuously in specific regions. These results suggest that the semantic relational structure of words, captured by word vectors, improves the modeling of semantic representations in widespread brain regions.</p>
</sec>
<sec id="sec005">
<title>Correlation between brain- and behavior-derived word dissimilarities</title>
<p>Next, we tested the correlation between the word dissimilarity matrix derived from the voxelwise models and that derived from behavioral data to clarify the behavioral correlates of modeled semantic representations. Word representations in the modeled brain space were calculated by multiplying the original fastText, GloVe, or word2vec word vectors by model weights, and the brain-derived word dissimilarity matrix was obtained from the correlation distance between all possible pairs of word representations (<xref ref-type="fig" rid="pcbi.1009138.g001">Fig 1B</xref>). Meanwhile, the behavior-derived word dissimilarity matrix was measured from the behavioral data from the word-arrangement task (<xref ref-type="fig" rid="pcbi.1009138.g002">Fig 2</xref>), in which 36 participants completed 18.8 (SD = 10.7) trials on average for each session (see <xref ref-type="sec" rid="sec007">Methods</xref> for more details). These brain- and behavior-derived word dissimilarity matrices were constructed by averaging the matrices over all brain models or all behavioral data separately for nouns and adjectives. In addition, to determine whether the behavioral correlates of word dissimilarity change through the transformation from original word vector representations to brain representations, we also calculated word vector-derived word dissimilarity matrices from the correlation distance between all possible pairs of original fastText, GloVe, or word2vec word vectors separately for nouns and adjectives.</p>
<p><xref ref-type="fig" rid="pcbi.1009138.g006">Fig 6</xref> shows the behavior-derived word dissimilarity matrices and the brain- and word vector-derived word dissimilarity matrices constructed using 1000-dimensional fastText word vectors. The brain-derived matrices were highly consistent across movie sets (Spearman’s ρ = 0.926 and 0.943 for nouns and adjectives, respectively). For both nouns (<xref ref-type="fig" rid="pcbi.1009138.g006">Fig 6A</xref>) and adjectives (<xref ref-type="fig" rid="pcbi.1009138.g006">Fig 6B</xref>), we found significant correlations between the brain- and behavior-derived matrices (brain–behavior correlations; permutation test, p &lt; 0.0001). The correlation coefficients were larger for nouns than for adjectives (p &lt; 0.0001). Although we also found significant correlations between the word vector- and behavior-derived matrices (word vector–behavior correlations; p &lt; 0.0001), the brain–behavior correlations for nouns (but not for adjectives) were significantly stronger than the word vector–behavior correlations (p &lt; 0.0001).</p>
<fig id="pcbi.1009138.g006" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g006</object-id>
<label>Fig 6</label>
<caption>
<title>Correlations between example dissimilarity matrices.</title>
<p>We have constructed word dissimilarity matrices separately for nouns (<bold>A</bold>) and adjectives (<bold>B</bold>). Each color map shows the behavior-derived matrices obtained from behavioral data (top left in <bold>A</bold> and <bold>B</bold>), the word vector-derived matrices obtained directly from 1000-dimensional fastText vectors (top right), or the brain-derived matrices obtained from voxelwise models (based on 1000-dimensional fastText vectors) for each movie set (bottom). Brighter colors in each map indicate higher dissimilarity of word pairs. Spearman’s correlation coefficients (ρ) are indicated.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g006" xlink:type="simple"/>
</fig>
<p><xref ref-type="fig" rid="pcbi.1009138.g007">Fig 7</xref> summarizes the brain–and word vector–behavior correlations for each word-vector type (i.e., fastText, GloVe, or word2vec; vector dimensionality = 1000). For all the vector types and for both nouns and adjectives, there were significant brain–behavior correlations as well as word vector–behavior correlations (permutation test, p &lt; 0.0001, FDR corrected). Such significant correlations were observed regardless of vector dimensionality (<xref ref-type="supplementary-material" rid="pcbi.1009138.s018">S18</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s020">S20</xref> Figs); although the correlation coefficients differed across vector dimensions, no clear tendency of changes in correlation coefficients across vector dimensions was determined. These results indicate that the brain-derived dissimilarity structure of semantic representations correlates with the behavior-derived dissimilarity structure regardless of vector types and dimensionality.</p>
<fig id="pcbi.1009138.g007" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g007</object-id>
<label>Fig 7</label>
<caption>
<title>Behavioral correlates of brain-derived word dissimilarity for different word vector types.</title>
<p>We computed brain–behavior and word vector–behavior correlations separately for the three types of 1000-dimensional word vectors (green, fastText; orange, GloVe; blue, word2vec). The Spearman’s correlation coefficients are separately shown for nouns (<bold>A</bold>) and adjectives (<bold>B</bold>). Marks above bars indicate the statistical significance of the correlation difference between different word vector types and between brain–behavior and word vector–behavior correlations (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g007" xlink:type="simple"/>
</fig>
<p>We then performed three different comparisons of behavioral correlations in <xref ref-type="fig" rid="pcbi.1009138.g007">Fig 7</xref>. In the comparison between nouns and adjectives, the brain–behavior correlations of noun dissimilarity were significantly stronger than those of adjective dissimilarity for all the vector types (permutation test, p &lt; 0.0001, FDR corrected). In the comparison between brain–behavior and word vector–behavior correlations, there was a significant difference only for fastText noun dissimilarity (p &lt; 0.0001, FDR corrected). In the comparison between vector types, the vector-type difference of brain–behavior correlations was partly significant but small (p &lt; 0.05, FDR corrected) whereas the word vector–behavior correlations for fastText vectors was much weaker than those for the other types of vectors (p &lt; 0.0001, FDR corrected).</p>
<p>In these cases, the populations from which we obtained these dissimilarity matrices were different between brain- and behavior-derived data. However, we observed a significant correlation for both nouns and adjectives and for both movie sets even when these matrices were collected from and averaged over the same population (six participants; permutation test, p &lt; 0.0001; <xref ref-type="supplementary-material" rid="pcbi.1009138.s021">S21 Fig</xref>). At the level of individual participants, the brain–behavior correlation of noun dissimilarity was significant for each of all the six participants from this population (p &lt; 0.01) whereas the correlation of adjective dissimilarity was significantly higher than chance level for only three participants of them (p &lt; 0.05; <xref ref-type="supplementary-material" rid="pcbi.1009138.s022">S22 Fig</xref>). Together, the word vector-based models could capture perceptual noun and adjective dissimilarities at the population level and perceptual noun dissimilarity at the individual level even when the brain- and behavior-derived word dissimilarities were originated from the same population.</p>
<p>The brain-derived dissimilarity matrices described above (Figs <xref ref-type="fig" rid="pcbi.1009138.g006">6</xref> and <xref ref-type="fig" rid="pcbi.1009138.g007">7</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s045">S18</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s022">S22</xref>) were calculated from the model weights of all cortical voxels (the number of voxels was 54914–73301 [mean ± SD = 62245 ± 5018] for movie set 1 and 52832–73301 [mean ± SD = 62150 ± 5190] for movie set 2). However, the high performance of the voxelwise models was observed in localized cortical regions (Figs <xref ref-type="fig" rid="pcbi.1009138.g004">4</xref>, <xref ref-type="fig" rid="pcbi.1009138.g005">5</xref>, <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s008">S8</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s013">S13</xref>). Hence, the correlation of brain- and behavior-derived dissimilarity matrices may be stronger when only the cortical regions with high model performance were used. To test this possibility, we constructed brain-derived dissimilarity matrices using only those voxels with the highest prediction accuracy (top 2000, 5000, 10000, 30000, or 50000 voxels). Then, their correlations with behavior-derived matrices were compared with the original correlations calculated using all cortical voxels. However, we found the strongest correlation when using all cortical voxels regardless of word vector types (Figs <xref ref-type="fig" rid="pcbi.1009138.g008">8</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s023">S23</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s025">S25</xref>). This result suggests that semantic information correlated with behavior are distributed broadly across the cortex and can be captured using word vector-based voxelwise models.</p>
<fig id="pcbi.1009138.g008" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g008</object-id>
<label>Fig 8</label>
<caption>
<title>Effects of voxel selection on brain–behavior correlation.</title>
<p>We have evaluated the correlations between brain- and behavior-derived word dissimilarity matrices when the brain-derived matrix was obtained from the selected voxels with the highest model accuracy. The correlation coefficients for fastText (<bold>A</bold> and <bold>B</bold>), GloVe (<bold>C</bold> and <bold>D</bold>), and word2vec (<bold>E</bold> and <bold>F</bold>) vectors are shown separately for nouns (<bold>A</bold>, <bold>C</bold>, and <bold>E</bold>) and adjectives (<bold>B</bold>, <bold>D</bold>, and <bold>F</bold>), while the numbers of selected voxels are changed (2000, 5000, 10000, 30000, 50000, and all voxels).</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g008" xlink:type="simple"/>
</fig>
<p>To gain a further understanding of cortical localization of behaviorally correlated sematic representations modeled by word vectors, we examined brain–behavior correlations of word dissimilarity within each cortical region. We found that although moderate or strong correlations were observed across the cortex (Spearman’s ρ = 0.356–0.621 for noun dissimilarity, 0.246–0.400 for adjective dissimilarity), the correlations were relatively weak in low-level audiovisual regions, such as the posterior occipital cortex and the superior temporal area (Figs <xref ref-type="fig" rid="pcbi.1009138.g009">9</xref> and <xref ref-type="fig" rid="pcbi.1009138.g010">10</xref>). This localization pattern of brain–behavior correlations across the cortex was deemed different from that of model prediction performance (Figs <xref ref-type="fig" rid="pcbi.1009138.g004">4</xref>, <xref ref-type="fig" rid="pcbi.1009138.g005">5</xref>, <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s008">S8</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s013">S13</xref>). This discrepancy is likely to cause the reduced brain–behavior correlation when voxels used for the analysis were selected according to high prediction accuracy (Figs <xref ref-type="fig" rid="pcbi.1009138.g008">8</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s023">S23</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s025">S25</xref>). Considering that the fraction of significantly predictable voxels was sufficiently large even in the high-level cortical regions showing strong brain–behavior correlations (<xref ref-type="fig" rid="pcbi.1009138.g005">Fig 5</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s013">S13</xref>), these results suggest that word vector-based models effectively capture semantic information in these regions, which is more closely associated with human perception than the low-level audiovisual regions.</p>
<fig id="pcbi.1009138.g009" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g009</object-id>
<label>Fig 9</label>
<caption>
<title>Cortical mapping of brain–behavior correlations of noun dissimilarity.</title>
<p>Spearman’s correlation coefficients between brain- and behavior-derived noun dissimilarity matrices were calculated within each cortical region and mapped onto the cortical surface of a reference brain for each of fastText (<bold>A</bold>), GloVe (<bold>B</bold>), and word2vec (<bold>C</bold>) models (vector dimension = 1000) and for each of movie sets 1(left) and 2 (right). Brighter colors indicate regions that have larger correlation coefficients. The five cortical regions with the highest correlation coefficients are numbered in a descending order separately for each type of word vectors and for each movie set. The names of these regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s042">S15 Table</xref>. Other conventions are the same as in <xref ref-type="fig" rid="pcbi.1009138.g004">Fig 4</xref>.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g009" xlink:type="simple"/>
</fig>
<fig id="pcbi.1009138.g010" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.g010</object-id>
<label>Fig 10</label>
<caption>
<title>Cortical mapping of brain–behavior correlations of adjective dissimilarity.</title>
<p>The same analysis as in <xref ref-type="fig" rid="pcbi.1009138.g009">Fig 9</xref> but for adjective dissimilarity. The names of the five cortical regions with the highest Spearman’s correlation coefficients for each word-vector type and each movie set are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s043">S16 Table</xref>.</p>
</caption>
<graphic mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.g010" xlink:type="simple"/>
</fig>
<p>Previous studies have reported that cortical representations in the object-selective visual areas show clear distinction between animate and inanimate categories [<xref ref-type="bibr" rid="pcbi.1009138.ref027">27</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref029">29</xref>]. Since the nouns we used for constructing word dissimilarity matrices consisted of six different semantic categories along animate/inanimate dimension (humans, non-human animals, non-animal nature things, constructs, vehicles, and other artifacts; see also <xref ref-type="sec" rid="sec007">Methods</xref>), it might be argued that the observed behavioral correlates of brain-derived word dissimilarity simply reflected such clear distinction of animate and inanimate representations. However, even when brain–behavior correlations were examined separately for each of the six categories, we found significant brain–behavior correlations for each of the “humans”, “non-animal natural things”, “vehicles”, and “other artifacts” categories (p &lt; 0.05, FDR corrected; <xref ref-type="supplementary-material" rid="pcbi.1009138.s026">S26 Fig</xref>). This result indicates that the observed brain–behavior correlations of word dissimilarity cannot be explained only by simple representational distinction between animate and inanimate categories.</p>
<p>Finally, we compared the brain-derived word dissimilarity modeled by word vectors and that modeled by discrete word features in terms of their behavioral correlates. In contrast to the performance of brain-response prediction (<xref ref-type="supplementary-material" rid="pcbi.1009138.s004">S4</xref>, <xref ref-type="supplementary-material" rid="pcbi.1009138.s015">S15</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s016">S16</xref> Figs), the brain–behavior correlations of word dissimilarity for the binary-labeling models were much weaker than those for the word vector-based models regardless of datasets and vector dimensionality (<xref ref-type="supplementary-material" rid="pcbi.1009138.s027">S27 Fig</xref>). This result suggests that the voxelwise modeling based on word vectors have a clear advantage in capturing the similarity structure of cortical semantic representations associated with human perception.</p>
</sec>
</sec>
<sec id="sec006" sec-type="conclusions">
<title>Discussion</title>
<p>We have examined the behavioral correlates of semantic representations estimated by word vector-based brain models. We constructed a voxelwise model that has the ability to predict movie-evoked fMRI signals in individual brains through a word vector space. The voxelwise models showed substantial response-prediction performance in reasonable cortical regions consistently across stimulus and parameter sets. There were significant correlations between the word dissimilarity structure derived from the voxelwise models and that derived from behavioral data. These results suggest that the semantic representations estimated from word vector-based brain models can appropriately capture the semantic relationship of words in human perception. Our findings contribute to the establishment of word vector-based brain modeling as a powerful tool in investigating human semantic processing.</p>
<p>Word vectors have been extensively leveraged for the modeling of semantic information in the human brain. Some studies have successfully predicted brain responses, through word vector spaces, evoked by audiovisual stimuli, such as movies [<xref ref-type="bibr" rid="pcbi.1009138.ref015">15</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref016">16</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref022">22</xref>], pictures [<xref ref-type="bibr" rid="pcbi.1009138.ref010">10</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref013">13</xref>], and sentences [<xref ref-type="bibr" rid="pcbi.1009138.ref014">14</xref>]. In addition, one of these studies has visualized the semantic representational space in the brain through word vector-based modeling [<xref ref-type="bibr" rid="pcbi.1009138.ref016">16</xref>]. Another line of studies has successfully recovered perceived semantic contents from brain responses by using word vector spaces [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref017">17</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref018">18</xref>]. In this way, word vectors will be a useful tool to investigate comprehensively the semantic processing in the human brain. Several studies have reported the association between original word vectors and human behavior [<xref ref-type="bibr" rid="pcbi.1009138.ref006">6</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref030">30</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref032">32</xref>]. However, no study has explicitly clarified the behavioral correlates of the brain semantic representations revealed by word vector-based models. Although previous studies in which movie-evoked brain responses were predicted from manual descriptions for the movie scenes [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref015">15</xref>] employed human behavior implicitly, they did not reveal the association between brain- and behavior-derived semantic representations. To the best of our knowledge, this is the first study demonstrating the behavioral correlates of semantic representations modeled by word vectors.</p>
<p>We demonstrated that word vector-based voxelwise models much outperformed voxelwise models with discrete word features in terms of the brain–behavior correlation of word dissimilarity (<xref ref-type="supplementary-material" rid="pcbi.1009138.s027">S27 Fig</xref>) although response-prediction performance and its cortical localization were deemed comparable between these models (<xref ref-type="supplementary-material" rid="pcbi.1009138.s004">S4</xref>, <xref ref-type="supplementary-material" rid="pcbi.1009138.s015">S15</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s016">S16</xref> Figs). This finding suggests that word vectors are more effective to model the similarity structure of cortical semantic representations associated with human perception, compared with discrete word features, which were used for voxelwise modeling in previous studies (e.g., [<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]). In addition, another advantage of word vector-based modeling is that semantic representations can be quantified using much more vocabularies [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>] than modeling based on discrete word features. This enables to comprehensively understand the representational structure of semantic information in the human brain. For example, a recent study demonstrated the multidimensional representations of natural objects on the basis of a data-driven approach using large-scale behavioral data of similarity judgements [<xref ref-type="bibr" rid="pcbi.1009138.ref033">33</xref>]. The word vector-based models capturing human word similarity judgements may be used to explore whether such a multidimensional structure of semantic representations actually exists in the human brain.</p>
<p>In addition, the word vector-based modeling with large vocabularies allows the quantification of different aspects of semantic information in the brain by using different types of words. In particular, adjectives have a potential for the visualization of impression information [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>]. We found the behavioral correlates of modeled semantic representations on nouns and even adjectives, which were consistently observed across stimulus and parameter sets (Figs <xref ref-type="fig" rid="pcbi.1009138.g007">7</xref>, <xref ref-type="supplementary-material" rid="pcbi.1009138.s018">S18</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s020">S20</xref>). These findings indicate that word vectors are deemed suitable for modeling the cortical representations of impression information. Thus, word vector-based modeling provides a useful framework for extensive investigation of semantic representations in the brain.</p>
<p>In this study, we found that the cortical regions showing high accuracy in predicting movie-evoked brain response by our models were localized primarily in the posterior and temporal cortices (Figs <xref ref-type="fig" rid="pcbi.1009138.g004">4</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s008">S8</xref>). This localization pattern of high predictability is consistent with previous reports in which cortical semantic representations were modeled from movie-evoked response [<xref ref-type="bibr" rid="pcbi.1009138.ref012">12</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]. However, when semantic models using word features are applied to brain response evoked by linguistic stimuli, e.g., sentences and speech, the brain response is highly predictable not only in sensory regions but also in the language network, including the left inferior frontal and superior temporal cortices [<xref ref-type="bibr" rid="pcbi.1009138.ref014">14</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref021">21</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref034">34</xref>]. These findings imply that the localization of high predictability depends on stimulus modalities, namely, whether stimuli are linguistic or not. This notion is supported by evidence that the language network has a relatively small contribution to semantic processing that requires no explicit linguistic processing [<xref ref-type="bibr" rid="pcbi.1009138.ref035">35</xref>]. Thus, the semantic representations modeled using the brain response evoked by nonlinguistic stimuli may be biased toward specific modalities, such as vision and audition.</p>
<p>However, the observed localization pattern of high predictability does not mean that our models fail to capture semantic information in high-level cortical regions, including the language network. We also found that the fraction of voxels being significantly predictable by our models was sufficiently high in widespread cortical regions (Figs <xref ref-type="fig" rid="pcbi.1009138.g005">5</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s014">S14</xref>). In addition, significant brain–behavior correlations of word dissimilarity were observed broadly across the cortex and rather stronger in the high-level association cortex than in the low-level audiovisual cortex (Figs <xref ref-type="fig" rid="pcbi.1009138.g009">9</xref> and <xref ref-type="fig" rid="pcbi.1009138.g010">10</xref>). In general, the neural signals in high-level cortical regions explain a larger portion of the variability in human perception and behavior than those in low-level cortical regions [<xref ref-type="bibr" rid="pcbi.1009138.ref036">36</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref038">38</xref>]. Thus, the results suggest that our word vector-based models can affectively capture semantic information in high-level cortical regions which is closely associated with human perception.</p>
<p>In the field of natural language processing, recently developed algorithms based on deep learning have exhibited state-of-the-art performance in many types of natural language tasks [<xref ref-type="bibr" rid="pcbi.1009138.ref039">39</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref041">41</xref>]. The feature representations obtained by these algorithms are also beginning to be used for the modeling of semantic representations [<xref ref-type="bibr" rid="pcbi.1009138.ref042">42</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref043">43</xref>], which show state-of-the-art performance even in brain-response prediction [<xref ref-type="bibr" rid="pcbi.1009138.ref044">44</xref>]. Hence, it is possible that the correlation between brain- and behavior-derived semantic representations improves by using such deep learning features instead of word vectors. However, some advantages are noted using word vectors compared with deep learning features: First, word vectors are easily trained with much smaller sets of parameters than deep learning features. Second, word vectors are easily interpretable owing to their direct association with the meaning of words. In contrast, deep learning features cannot be easily interpreted without sophisticated methods for improving their interpretability [<xref ref-type="bibr" rid="pcbi.1009138.ref045">45</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref046">46</xref>]. Third, there are many existing methods to improve word-vector spaces for specific types of tasks [<xref ref-type="bibr" rid="pcbi.1009138.ref013">13</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref047">47</xref>–<xref ref-type="bibr" rid="pcbi.1009138.ref051">51</xref>]. These advantages may be also beneficial for the modeling of semantic representations. Therefore, it is important to use these two types of algorithms separately depending on the aims of specific studies on semantic processing.</p>
</sec>
<sec id="sec007" sec-type="materials|methods">
<title>Methods</title>
<sec id="sec008">
<title>Ethics statement</title>
<p>The experimental protocol was approved by the ethics and safety committees of the National Institute of Information and Communications Technology. Written informed consent was obtained from all of experimental participants.</p>
</sec>
<sec id="sec009">
<title>Participants</title>
<p>In total, 52 healthy Japanese participants (21 females and 31 males; age 20–61, mean ± SD = 26.8 ± 8.9 years) were recruited for the 2 sets of fMRI experiments. Among the recruits, 36 participated in one or the other of these fMRI experiments, and 16 participated in both. In addition, another 36 healthy Japanese participants (20 females and 16 males; age 18–58, mean ± SD = 25.9 ± 10.1 years) were recruited for psychological experiments. Of these, 6 were also participants of the fMRI experiments and 30 were unique participants. All participants had normal or corrected-to-normal vision. The fMRI data, but not the psychological data, used here were also utilized in a previous publication [<xref ref-type="bibr" rid="pcbi.1009138.ref022">22</xref>].</p>
</sec>
<sec id="sec010">
<title>MRI experiments</title>
<p>Functional and anatomical MRI data were collected via a 3T Siemens MAGNETOM Prisma scanner (Siemens, Germany), with a 64-channel Siemens volume coil. Functional data were collected using a multiband gradient echo EPI sequence [<xref ref-type="bibr" rid="pcbi.1009138.ref052">52</xref>] (TR = 1000 ms; TE = 30 ms; flip angle = 60°; voxel size = 2 × 2 × 2 mm; matrix size = 96 × 96; FOV = 192 × 192 mm; the number of slices = 72; multiband factor = 6). Anatomical data were also gathered using a T1-weighted MPRAGE sequence (TR = 2530 ms; TE = 3.26 ms; flip angle = 9°; voxel size = 1 × 1 × 1 mm; matrix size = 256 × 256; FOV = 256 × 256 mm; the number of slices = 208) on the same 3T scanner.</p>
<p>In the two sets of fMRI experiments, participants were asked to view movie stimuli on a projector screen inside the scanner (27.9 × 15.5 of visual angle at 30Hz) and used MR-compatible headphones for the sounds. The participants were given no explicit task. The fMRI data for each participant upon viewing of the movies were collected in three separate recording sessions over 3 days for each set of fMRI experiments.</p>
<p>The movie stimuli consisted of Japanese television advertisements for one set of experiments (movie set 1) and Japanese web advertisements for the other set of experiments (movie set 2; see also [<xref ref-type="bibr" rid="pcbi.1009138.ref022">22</xref>]). Movie set 1 has included 420 ads broadcasted on Japanese TV between 2011 and 2017; meanwhile, movie set 2 included 368 ads broadcasted on the Internet between 2015 and 2018. The ad movies were all unique, include a wide variety of product categories (see <xref ref-type="supplementary-material" rid="pcbi.1009138.s044">S17 Table</xref> for more details). The length of each movie was 15 or 30 s. To create the movie stimuli for each experiment, the original movies in each of the movie sets 1 and 2 were sequentially concatenated in a pseudo-random order. For each movie set, 14 non-overlapping movie clips of 610 s in length were obtained.</p>
<p>Individual movie clips were then displayed in separate scans. The initial 10 s of each clip served as a dummy in order to discard hemodynamic transient signals caused by clip onset. fMRI responses collected during the 10 s dummy part were not used for modeling. Twelve clips from each movie set were only presented once. The fMRI responses to these clips were used for the training of the voxelwise models (training dataset; 7200 s in total). The other two clips for each movie set were presented four times each in four separate scans. The fMRI responses to these clips were then averaged across four scans to improve the signal-to-noise ratio [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref053">53</xref>]. The averaged responses were further used for the test of the voxelwise models (test dataset; 1200 s in total).</p>
</sec>
<sec id="sec011">
<title>MRI data pre-processing</title>
<p>Motion correction in each functional scan was carried out via the statistical parameter mapping toolbox (SPM8, <ext-link ext-link-type="uri" xlink:href="http://www.fil.ion.ucl.ac.uk/spm/software/spm8/" xlink:type="simple">http://www.fil.ion.ucl.ac.uk/spm/software/spm8/</ext-link>). For each subject, all volumes were aligned to the first image from the first functional run. Low-frequency fMRI response drift was then eliminated by subtracting median-filtered signals (within a 120-s window) from raw signals. Then, the response for each voxel was normalized by subtracting the mean response and scaling to the unit variance. FreeSurfer [<xref ref-type="bibr" rid="pcbi.1009138.ref024">24</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref025">25</xref>] was used to identify cortical surfaces from anatomical data and register these surfaces to the voxels of functional data. Then, each voxel was assigned to one of the 148 cortical regions derived from the Destrieux atlas for cortical segmentation [<xref ref-type="bibr" rid="pcbi.1009138.ref026">26</xref>].</p>
</sec>
<sec id="sec012">
<title>Word vector spaces</title>
<p>This study used fastText skip-gram [<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>], GloVe [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>], or word2vec skip-gram [<xref ref-type="bibr" rid="pcbi.1009138.ref002">2</xref>] to construct a word vector space. These algorithms have been originally developed to learn a word vector space based on word co-occurrence statistics in natural language texts.</p>
<p>The training objective of the skip-gram algorithm of fastText and word2vec is to obtain word vector representations that enable the surrounding words to be accurately predicted from a given word in a sentence [<xref ref-type="bibr" rid="pcbi.1009138.ref002">2</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>]. More formally, given a sequence of training words <italic>w</italic><sub>1</sub>, <italic>w</italic><sub>2</sub>,…,<italic>w</italic><sub><italic>T</italic></sub>, the skip-gram algorithm seeks a <italic>K</italic>-dimensional vector space that maximizes the average log probability, given as:
<disp-formula id="pcbi.1009138.e001">
<alternatives>
<graphic id="pcbi.1009138.e001g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e001" xlink:type="simple"/>
<mml:math display="block" id="M1">
<mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:mfrac><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mrow><mml:munder><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mo>−</mml:mo><mml:mi>c</mml:mi><mml:mo>≤</mml:mo><mml:mi>j</mml:mi><mml:mo>≤</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>≠</mml:mo><mml:mn>0</mml:mn></mml:mrow></mml:munder><mml:mrow><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:mspace width="0.25em"/><mml:mi mathvariant="normal">p</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mrow>
</mml:math>
</alternatives>
</disp-formula>
where <italic>c</italic> is the size of the context window, which corresponds to the number of to-be-predicted words before and after the center word <italic>w</italic><sub><italic>t</italic></sub>. Therefore, the skip-gram vector space is optimized on the basis of the local co-occurrence statistics of nearby words in the text corpus. The basic formulation of <italic>p</italic>(<italic>w</italic><sub><italic>t</italic>+<italic>j</italic></sub>|<italic>w</italic><sub><italic>t</italic></sub>) is the softmax function:
<disp-formula id="pcbi.1009138.e002">
<alternatives>
<graphic id="pcbi.1009138.e002g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e002" xlink:type="simple"/>
<mml:math display="block" id="M2">
<mml:mi>p</mml:mi><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>|</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">x</mml:mi><mml:mi mathvariant="normal">p</mml:mi><mml:mo>(</mml:mo><mml:mi>〈</mml:mi><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">v</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">v</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mi>〉</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:msubsup><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>W</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo>(</mml:mo><mml:mi>〈</mml:mi><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">v</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">v</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mi>〉</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mfrac>
</mml:math>
</alternatives>
</disp-formula>
where <bold>v</bold><sub><italic>wi</italic></sub> is the vector representation of <italic>w</italic><sub><italic>i</italic></sub>, <italic>W</italic> is the number of words in the vocabulary, and 〈<bold>v</bold><sub>1</sub>, <bold>v</bold><sub>2</sub>〉 indicates the inner product of vectors <bold>v</bold><sub>1</sub> and <bold>v</bold><sub>2</sub>. However, because of the high computational cost of this formulation, the negative sampling technique is used to produce a computationally efficient approximation of the softmax function [<xref ref-type="bibr" rid="pcbi.1009138.ref002">2</xref>]. In addition, fastText introduces sub-word modeling, which is robust for inflected and rare words [<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>].</p>
<p>GloVe learns word vector representations using a global log-bilinear regression model that is subject to the cost function:
<disp-formula id="pcbi.1009138.e003">
<alternatives>
<graphic id="pcbi.1009138.e003g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e003" xlink:type="simple"/>
<mml:math display="block" id="M3">
<mml:mi>J</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:munderover><mml:mo stretchy="false">∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>V</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi>f</mml:mi><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>−</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">log</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mrow>
</mml:math>
</alternatives>
</disp-formula>
where <italic>V</italic> is the vocabulary size, <inline-formula id="pcbi.1009138.e004"><alternatives><graphic id="pcbi.1009138.e004g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e004" xlink:type="simple"/><mml:math display="inline" id="M4"><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow></mml:msub><mml:mo>∈</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></alternatives></inline-formula> is a <italic>d</italic>-dimensional vector of word <italic>i</italic>, <inline-formula id="pcbi.1009138.e005"><alternatives><graphic id="pcbi.1009138.e005g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e005" xlink:type="simple"/><mml:math display="inline" id="M5"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">j</mml:mi></mml:mrow></mml:msub><mml:mo>∈</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></alternatives></inline-formula> is a separate context vector of word <italic>j</italic>, and <italic>b</italic><sub><italic>i</italic></sub> and <inline-formula id="pcbi.1009138.e006"><alternatives><graphic id="pcbi.1009138.e006g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e006" xlink:type="simple"/><mml:math display="inline" id="M6"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></alternatives></inline-formula> are additional biases. <italic>X</italic> is the co-occurrence matrix, where <italic>X</italic><sub><italic>ij</italic></sub> represents the number of occurrences of word <italic>i</italic> in the context of word <italic>j</italic>. <italic>f</italic>(<italic>X</italic><sub><italic>ij</italic></sub>) is a specific weighting function. GloVe combines the advantages of local context window and global matrix factorization methods [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>].</p>
<p>A vector space of each algorithm was constructed from a text corpus of the Japanese Wikipedia dump on April 1, 2020. All Japanese texts in the corpus were segmented into words and lemmatized using MeCab (<ext-link ext-link-type="uri" xlink:href="http://taku910.github.io/mecab" xlink:type="simple">http://taku910.github.io/mecab</ext-link>). Used were only nouns, verbs, and adjectives. In an attempt to improve the reliability of word-vector learning, the vocabulary size was restricted to ~100,000 words by excluding words that appeared infrequently in the corpus. The learning parameters of fastText and word2vec used were as follows: window size = 10; the number of negative samples = 5; downsampling rate = 0.001; and the number of learning epochs = 10. The learning parameters of GloVe used were as follows: window size = 10; initial learning rate = 0.05; <italic>α</italic> = 0.75; and the number of training iterations = 50. The vector spaces were constructed using five different numbers of vector dimensions (100, 300, 500, 1000, and 2000). Of these, the 1000-dimension vector space of each algorithm was used for the main analysis (Figs <xref ref-type="fig" rid="pcbi.1009138.g003">3</xref>–<xref ref-type="fig" rid="pcbi.1009138.g010">10</xref>), and the other dimensions were used to test the effect of vector dimensionality on modeling performance and behavioral correlates. To eliminate the effects of trial variations in the learning quality of vector spaces of each algorithm on the performance of voxelwise modeling, five different vector spaces were learned independently using the same text corpus and the same algorithm with the same parameters. All results were obtained from the average over five learned spaces for each parameter set.</p>
</sec>
<sec id="sec013">
<title>Movie scene descriptions</title>
<p>Manual scene descriptions using natural Japanese language were provided for every 1-s scene of each movie in movie sets 1 and 2, in a manner similar to that described previously [<xref ref-type="bibr" rid="pcbi.1009138.ref011">11</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref012">12</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref022">22</xref>]. The annotators were native Japanese speakers (movie set 1: 68 females and 28 males, age 19–62 years; movie set 2: 11 females and 2 males, age 20–56 years), who were not the fMRI participants. They were instructed to describe each scene (the middle frame of each 1-s clip) using more than 50 Japanese characters. Multiple annotators (movie set 1, 12–14 annotators; movie set 2, 5 annotators) were randomly assigned for each scene to reduce the potential effect of personal bias. The descriptions contain a variety of expressions reflecting not only objective perceptions but also the subjective perceptions of the annotator (e.g., impression, feeling, association with ideas; for more details, see <ext-link ext-link-type="uri" xlink:href="https://osf.io/3hkwd" xlink:type="simple">https://osf.io/3hkwd</ext-link>).</p>
<p>Each description for a given scene was also segmented, lemmatized, and decomposed into nouns, verbs, and adjectives via MeCab as described above; then, they were transformed into fastText, GloVe, or word2vec vectors. The word vectors were then averaged within each description. For each scene, all vectors obtained from the different descriptions were averaged. Through this procedure, a single vector (scene vector) was obtained for each 1-s scene, which was later used for modeling.</p>
</sec>
<sec id="sec014">
<title>Binary labeling of movie scenes</title>
<p>To compare discrete word features with word vectors in terms of the modeling of brain semantic representations, the binary labeling of semantic contents in each movie scene was used to construct voxelwise models. The binary labels were obtained from the presence/absence of words in the manual movie-scene descriptions. For each scene, individual words (nouns, verbs, and adjectives) in the vocabulary were counted in all descriptions for the scene. If a given word was present, the binary label of the word was 1. If not, the binary label of the word was 0. This labeling was performed over all movie scenes. Then, scene vectors with the dimensionality of 100, 300, 500, 1000, and 2000 were obtained from the binary labels of 100, 300, 500, 1000, and 2000 words that most frequently appeared in the movie-scene descriptions for each movie set. This type of semantic features is similar to ones used in a previous study on the modeling of cortical semantic representations [<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>]. However, the previous study performed the binary labeling of movie scenes using the manual labeling of single words by one annotator, unlike our binary labeling.</p>
</sec>
<sec id="sec015">
<title>Voxelwise modeling</title>
<p>The procedure of voxelwise modeling was similar to that described previously [<xref ref-type="bibr" rid="pcbi.1009138.ref012">12</xref>]. A series of fMRI responses evoked in individual <italic>N</italic> voxel by a series of <italic>S</italic> movie scenes, represented by a <italic>S</italic>×<italic>N</italic> matrix <bold>R</bold>, was modeled as a weighted linear combination of a <italic>K</italic>-dimensional scene-vector matrix <bold>V</bold> plus isotropic Gaussian noise <bold><italic>ε</italic></bold>.</p>
<disp-formula id="pcbi.1009138.e007">
<alternatives>
<graphic id="pcbi.1009138.e007g" mimetype="image" position="anchor" xlink:href="info:doi/10.1371/journal.pcbi.1009138.e007" xlink:type="simple"/>
<mml:math display="block" id="M7">
<mml:mi mathvariant="bold">R</mml:mi><mml:mo>=</mml:mo><mml:mi mathvariant="bold">V</mml:mi><mml:mi mathvariant="bold">W</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="bold">ε</mml:mi>
</mml:math>
</alternatives>
</disp-formula>
<p>A set of linear temporal filters was used to model the slow hemodynamic response and its coupling with brain response [<xref ref-type="bibr" rid="pcbi.1009138.ref053">53</xref>]. In an attempt to capture the hemodynamic delay in the responses, the <italic>S</italic>×4<italic>K</italic> scene-vector matrix was constructed by concatenating four sets of <italic>K</italic>-dimensional scene vectors with delays of 3, 4, 5, and 6 s. The 4<italic>K</italic>×<italic>N</italic> weighted matrix <bold>W</bold> was estimated using an L2-regularized linear least-squares regression, which can obtain good estimates even for models containing a large number of regressors [<xref ref-type="bibr" rid="pcbi.1009138.ref023">23</xref>].</p>
<p>The model training was conducted using the training dataset (7200 datapoints). In order to estimate the regularization parameters, the training dataset was randomly divided into two subsets containing 80% and 20% of the samples, for model fitting and validation, respectively. This random resampling procedure was repeated 10 times. Regularization parameters were optimized, according to the mean Pearson’s correlation coefficient between the predicted and measured fMRI signals for the 20% validation samples. An optimal parameter was obtained separately for each model.</p>
<p>The brain-response prediction performance of the model was evaluated using the test dataset (1200 datapoints), which was not used for model fitting or parameter estimation. The prediction accuracy of the model was quantified as the Pearson’s correlation coefficient between the predicted and average measured fMRI signals in the test dataset. In addition, the significance of prediction accuracy (Pearson’s correlation) for each of all cortical voxels was evaluated after the correction for multiple comparisons using FDR. The rate of voxels, to all cortical voxels, for which the prediction accuracy reached a significance threshold (p &lt; 0.05, FDR corrected) was calculated as the fraction of significant voxels, which was another measure of brain-response prediction performance for each model.</p>
<p>To test whether the fraction of significant voxels for each cortical region was significantly above chance level, two types of statistical tests were performed. First, using the fraction for each model of each participant as a data sample, whether or not the fraction was significantly higher than a standard significance threshold (p = 0.05) was tested using Wilcoxon test with FDR correction. Second, the chance level of the fraction for each cortical region was more conservatively estimated using control voxelwise models based on word vectors that were shuffled across vector dimensions for each vector. Five different sets of the shuffled word vectors were generated from each word-vector space and used to construct five different control voxelwise models for each participant. Then, the fraction of significant voxels was obtained for each cortical region using these control models and averaged across the five different models, which produces the chance level of the fraction for each participant. Whether the fraction of significant voxels for each original model was significantly above the chance level was tested for each region using Wilcoxon test with FDR correction while the fraction for each participant was used as a data sample.</p>
<p>Word vectors effectively capture the semantic relational structure of words as demonstrated by previous studies that applied word vectors to word similarity judgements [<xref ref-type="bibr" rid="pcbi.1009138.ref003">3</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref004">4</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref006">6</xref>]. To determine whether such semantic relational structure of word vectors is effective in the modeling of cortical semantic representations, the response-prediction performance of voxelwise models using original word vectors was compared with that using untrained word vectors. Untrained word vectors were obtained with the same procedure as used for the generation of original, trained word vectors, except that the statistical training of word vectors was omitted. This procedure yielded 1000-dimensional random vectors that were each assigned to each word in the vocabulary shared with the corresponding trained vectors. Hence, untrained vectors had signatures of individual words but not the semantic relational structure of words at all. Then, scene vectors were generated from untrained vector representations of scene descriptions.</p>
<p>Voxelwise models based on untrained vectors were constructed using the same procedure as used for voxelwise models based on trained vectors. Five different voxelwise models for each participant were constructed from five different sets of untrained vectors. The performance of brain-response prediction (i.e., prediction accuracy and the fraction of significant voxels) for each participant was obtained by averaging the performance of these five models. Finally, the performance for the models based on untrained vectors was statistically compared with that for the models based on trained vectors. Higher performance for the models based on trained vectors indicates that the semantic relational structure of word vectors is effective in the modeling of cortical semantic representations.</p>
</sec>
<sec id="sec016">
<title>Psychological experiments</title>
<p>Participants were asked to perform a word-arrangement task on a PC. This task is considered to be a modified version of the psychological task introduced previously [<xref ref-type="bibr" rid="pcbi.1009138.ref019">19</xref>]. Kriegeskorte and colleagues originally used this task to examine the behavioral correlates of cortical object representations, which were quantified using the representational similarity of fMRI signals evoked by object stimuli [<xref ref-type="bibr" rid="pcbi.1009138.ref054">54</xref>,<xref ref-type="bibr" rid="pcbi.1009138.ref055">55</xref>]. The present study used this task to study the methodological validity of word vector-based brain modeling by testing the behavioral correlates of modeled cortical semantic representations. The psychological data are available online (<ext-link ext-link-type="uri" xlink:href="https://osf.io/um3qg/" xlink:type="simple">https://osf.io/um3qg/</ext-link>).</p>
<p>In each trial of this task, the participants were required to arrange ≤60 words (nouns or adjectives) in a two-dimensional space according to their semantic relationship on a computer screen by mouse drag-and-drop operations (<xref ref-type="fig" rid="pcbi.1009138.g002">Fig 2</xref>). This paradigm has allowed for the efficient collection of the perceptual semantic dissimilarity between words from participants [<xref ref-type="bibr" rid="pcbi.1009138.ref019">19</xref>].</p>
<p>The words used in this task included 60 nouns and 60 adjectives in Japanese (<xref ref-type="table" rid="pcbi.1009138.t003">Table 3</xref>). These words were selected from the vocabulary of the fastText space (i.e., top 100,000 most frequently used words in the Japanese Wikipedia corpus). Nouns were selected in terms of the following six semantic categories: humans, non-human animals, non-animal natural things, constructs, vehicles, and other artifacts. Ten words were selected for each category. For adjectives, such category-based selection was deemed difficult due to the small number of adjectives in the vocabulary (only 473 words). Nouns and adjectives were separately used in two distinct sessions of the task, which was performed over 2 days.</p>
<p>The words were arranged in a designated circular area on the computer screen (“arena”). The words were initially displayed outside the arena. The participants used mouse drag-and-drop operations in moving each word item into the arena and arranging them. The arrangement of words was according to the participants’ own judgment of semantic similarity and dissimilarity between words. More similar or dissimilar word pairs should be closer or further apart in the arena, respectively. The participants were allowed to move any words within the arena as many times as they wished. Once all words were moved into the arena, the participants could click a button marked “Next” anytime to move on to the next trial.</p>
<p>On the initial trial of each session, the participants have arranged the entire set of 60 words (nouns or adjectives). On subsequent trials, they arranged subsets of those 60 words. After the end of each trial, the rough estimate of the word dissimilarity matrix and the evidence (0–1) for each word pairwise dissimilarity were computed as described previously [<xref ref-type="bibr" rid="pcbi.1009138.ref019">19</xref>]. Words in the subsets chosen for each trial were determined so as to increase evidence for pairwise dissimilarities of the words whose evidence estimated on the previous trial was the weakest. The session continued until the evidence of every word pair dissimilarity was above a threshold (0.75) or until the total duration of the session approached 1 h.</p>
<table-wrap id="pcbi.1009138.t003" position="float">
<object-id pub-id-type="doi">10.1371/journal.pcbi.1009138.t003</object-id>
<label>Table 3</label> <caption><title>Words used in psychological experiments and for calculating word dissimilarity matrices.</title></caption>
<alternatives>
<graphic id="pcbi.1009138.t003g" mimetype="image" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.t003" xlink:type="simple"/>
<table>
<colgroup>
<col align="left" valign="middle"/>
<col align="left" valign="middle"/>
</colgroup>
<thead>
<tr>
<th align="left" style="background-color:#BFBFBF">Nouns</th>
<th align="left" style="background-color:#BFBFBF">Adjectives</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Employee, carpenter, teacher, student, driver, player, actor, writer, doctor, researcher, monkey, dog, cat, horse, cow, sheep, bird, frog, fish, bug, star, sky, mountain, sea, river, water, flower, tree, dirt, vegetable, house, school, factory, hospital, station, airport, bridge, hotel, stadium, park, car, motorbike, train, ship, boat, airplane, helicopter, rocket, bicycle, carriage, movie, book, music, phone, PC, TV, cloth, table, cup, text</td>
<td align="left">Many, high, strong, large, absent, good, near, long, new, few, deep, broad, low, young, bad, bold, small, short, red, early, white, beautiful, black, old, weak, intense, far, detailed, childish, tender, bright, quick, correct, strict, blue, slow, sweet, thick, narrow, difficult, dim, rare, thin, heavy, light, pleasant, slender, wide, dark, hot, scary, shallow, sharp, circular, interesting, pale, sad, equal, remarkable, terrific</td>
</tr>
</tbody>
</table>
</alternatives>
<table-wrap-foot>
<fn id="t003fn001"><p>Original words were in Japanese, and, for display purposes, here they were translated into English.</p></fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="sec017">
<title>Word dissimilarity matrix</title>
<p>The brain-derived word dissimilarity matrix for the word vector-based models was estimated using the following procedure: First, the 4<italic>K</italic>×<italic>N</italic> voxelwise model weights for each participant were transformed into <italic>K</italic>×<italic>N</italic> by averaging the weights across the four sets of hemodynamic delay terms. Second, all <italic>N</italic> or top <italic>M</italic> voxels with the highest prediction accuracy were then selected from the model weights. All voxels were used in the main analysis, and selected voxels (<italic>M</italic> = 2000, 5000, 10000, 30000, and 50000) were used in an additional analysis that tested the effect of voxel selection on brain–behavior correlation. Third, the <italic>K</italic>-dimensional fastText, GloVe, or word2vec vectors for the words used in the psychological experiments (<xref ref-type="table" rid="pcbi.1009138.t003">Table 3</xref>) were multiplied by the <italic>K</italic>×<italic>N</italic> weight matrix, which yielded <italic>N</italic>- or <italic>M</italic>-dimensional word representations in the modeled brain space. Finally, a word dissimilarity matrix was calculated by the correlation distance (1 –Pearson’s correlation coefficient) between these word representations, which was averaged across participants separately for nouns and adjectives.</p>
<p>The brain-derived word dissimilarity matrix for the voxelwise model based on binary labels (binary-labeling model) was also estimated. For the binary-labeling model, the <italic>K</italic>×<italic>N</italic> matrix into which the 4<italic>K</italic>×<italic>N</italic> model weights were transformed by averaging across hemodynamic delay terms corresponded to the brain representations of <italic>K</italic> words. Accordingly, a word dissimilarity matrix of this model was calculated by the correlation distance between these <italic>N</italic>-dimensional word representations. Note that unlike the word vector-based models, the representations for the binary-labeling model contained only a limited part of the words used in the psychological experiments (<xref ref-type="table" rid="pcbi.1009138.t003">Table 3</xref>). The number of words contained depended on vector dimensionality and movie sets (<xref ref-type="supplementary-material" rid="pcbi.1009138.s045">S18 Table</xref>).</p>
<p>The behavior-derived word dissimilarity matrix was evaluated from multiple word-subset arrangements on the word-arrangement task. The dissimilarity for a given word pair was estimated as a weighted average of the scale-adjusted dissimilarity estimates from individual arrangements as has been described previously [<xref ref-type="bibr" rid="pcbi.1009138.ref019">19</xref>]. These estimates produced a word dissimilarity matrix for the entire set of words. In this way, the word dissimilarity matrix was estimated separately for nouns and adjectives and was averaged across participants.</p>
<p>In addition, a word vector-derived word dissimilarity matrix was also examined in order to test whether the behavioral correlates of semantic dissimilarity structures changed through the transformation from original word vector representations to brain word representations. The word vector-derived matrix was computed by the correlation distance between fastText, GloVe, or word2vec vectors of the same word sets (<xref ref-type="table" rid="pcbi.1009138.t003">Table 3</xref>) separately for nouns and adjectives.</p>
<p>The behavioral correlates of brain- or word vector-derived data were evaluated using Spearman’s correlation between the brain- and behavior-derived word dissimilarity matrices (brain–behavior correlation) and between the word vector- and behavior-derived word dissimilarity matrices (word vector–behavior correlation). The upper triangular portion of each matrix was used for calculating correlation coefficients. For the binary-labeling model, correlation coefficients were calculated using only the words contained in the vocabulary of the model (<xref ref-type="supplementary-material" rid="pcbi.1009138.s045">S18 Table</xref>).</p>
<p>A permutation test was performed to evaluate the statistical significance of brain–behavior and word vector–behavior correlations. In each repetition of this test, the rows of a brain- or word vector-derived word dissimilarity matrix were randomly shuffled. Then, this shuffled brain- or word vector-derived matrix was used to calculate a Spearman’s correlation with the original behavior-derived matrix. This procedure was repeated 10,000 times to obtain a null distribution of correlation coefficients and estimate the p value of an actual correlation coefficient. The difference of correlation coefficients between different matrix pairs was also tested in a similar manner. In this case, the difference of brain–or word vector–behavior correlation coefficients between different matrix pairs was calculated in each repetition while a brain- or word vector-derived matrix was shuffled separately for each of the different matrix pairs. This procedure was repeated 10,000 times to estimate a null distribution of the correlation coefficient differences.</p>
</sec>
</sec>
<sec id="sec018" sec-type="supplementary-material">
<title>Supporting information</title>
<supplementary-material id="pcbi.1009138.s001" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s001" xlink:type="simple">
<label>S1 Fig</label>
<caption>
<title>Model performance comparison across fastText vector dimensions.</title>
<p><bold>A</bold>) Mean prediction accuracy of voxelwise models based on the fastText vector space with different vector dimensions. Error bars indicate standard error of the mean (SEM). <bold>B</bold>) Difference of mean prediction accuracies between different vector dimensions. The difference was evaluated separately for movie sets 1 (left) and 2 (right). The color of each cell represents the accuracy difference of the dimension on the x-axis minus the dimension on y-axis (red, positive values; blue, negative values). The mark in each cell indicates the statistical significance of the difference (Wilcoxon test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, n.s., p &gt; 0.05, FDR corrected). <bold>C</bold>) Fractions of significant voxels for voxelwise models with different vector dimensions. The same conventions were used as in <bold>A</bold>. <bold>D</bold>) Difference in fractions of significant voxels between different vector dimensions. The same conventions were used as in <bold>B</bold>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s002" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s002" xlink:type="simple">
<label>S2 Fig</label>
<caption>
<title>Model performance comparison across GloVe vector dimensions.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s001">S1 Fig</xref> but for GloVe vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s003" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s003" xlink:type="simple">
<label>S3 Fig</label>
<caption>
<title>Model performance comparison across word2vec vector dimensions.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s001">S1 Fig</xref> but for word2vec vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s004" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s004" xlink:type="simple">
<label>S4 Fig</label>
<caption>
<title>Comparison of prediction performance with voxelwise models based on discrete word features.</title>
<p>To compare discrete word features with word vectors in terms of the modeling of brain semantic representations, voxelwise models with the binary labeling of movie scenes (binary-labeling models) were constructed for individual participants. Prediction accuracy (top) and the fraction of significant voxels (bottom) for the binary-labeling models (pink bars) are shown separately for the vector dimensionality of 100, 300, 500,1000, and 2000 (from left to right), along with prediction performance for the word vector-based models (green bars, fastText; orange bars, GloVe; blue bars, word2vec). Error bars indicate SEM. Marks above bars indicate the statistical significance of the performance difference between the binary-labeling model and each of the word vector-based models (Wilcoxon test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s005" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s005" xlink:type="simple">
<label>S5 Fig</label>
<caption>
<title>Comparison of model prediction performance between trained and untrained word vectors.</title>
<p>To test whether the semantic relational structure of words, captured by word vectors, is effective in the modeling of cortical semantic representations, we compared trained (original) and untrained word vectors in terms of the model performance of brain-response prediction. We obtained untrained vectors by randomly assigning a 1000-dimensional random vector to each word in the same vocabulary as used for trained vectors. Hence, untrained vectors had signatures of individual words but not the semantic relational structure of words. Voxelwise models based on untrained vectors were constructed using the same procedure as used for voxelwise models based on trained vectors. <bold>A)</bold> Prediction accuracy for models based on trained vectors (x-axis) and ones based on untrained vectors (y-axis) separately shown for each word-vector type (top, fastText; middle, GloVe; bottom, word2vec) and each dataset (left, movie set 1; right, movie set 2). Each dot represents mean prediction accuracy over all voxels of each brain. The models based on trained vectors exhibited significantly higher prediction accuracy than those based on untrained vectors regardless of vector types and datasets (Wilcoxon test, p &lt; 0.00001, FDR corrected). <bold>B)</bold> The fraction of significant voxels for models based on trained vectors (x-axis) and ones based on untrained vectors (y-axis). Each dot represents the fraction for each brain. The models based on trained vectors exhibited significantly higher fraction than those based on untrained vectors regardless of vector types and datasets (p &lt; 0.00001, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s006" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s006" xlink:type="simple">
<label>S6 Fig</label>
<caption>
<title>Cortical mapping of prediction accuracy for other dimensions of fastText vectors.</title>
<p>The mean prediction accuracy of fastText vector-based voxelwise models in each brain region was mapped onto the cortical surface (<bold>A</bold>, vector dimension = 100; <bold>B</bold>, 300; <bold>C</bold>, 500; <bold>D</bold>, 2000). The same conventions were used as in <xref ref-type="fig" rid="pcbi.1009138.g004">Fig 4</xref>. The five numbered cortical regions with the highest mean prediction accuracy for each dimension and dataset are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s028">S1 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s007" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s007" xlink:type="simple">
<label>S7 Fig</label>
<caption>
<title>Cortical mapping of prediction accuracy for other dimensions of GloVe vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6 Fig</xref> but for GloVe vector-based models. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s029">S2 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s008" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s008" xlink:type="simple">
<label>S8 Fig</label>
<caption>
<title>Cortical mapping of prediction accuracy for other dimensions of word2vec vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6 Fig</xref> but for word2vec vector-based models. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s030">S3 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s009" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s009" xlink:type="simple">
<label>S9 Fig</label>
<caption>
<title>Cortical mapping of prediction accuracy at the voxel level.</title>
<p>Voxelwise prediction accuracy of each word vector-based voxelwise model (vector dimension = 1000) for a representative participant was mapped onto the cortical flat map of the participant (from top to bottom: fastText, GloVe, word2vec, and binary labeling models; left: movie set 1, right: movie set 2). Brighter colors on the cortical maps indicate voxels with higher prediction accuracy. Only voxels with prediction accuracy above 0.10 are shown. The values of voxelwise prediction accuracy range up to 0.709. White lines on the cortical maps denote representative sulci: CoS, collateral sulcus; STS, superior temporal sulcus; TOS, transverse occipital sulcus; IPS, intraparietal sulcus; SyF, sylvian fissure; CeS, central sulcus; SFS, superior frontal sulcus; IFS, inferior frontal sulcus.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s010" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s010" xlink:type="simple">
<label>S10 Fig</label>
<caption>
<title>Upper range of voxelwise prediction accuracy.</title>
<p>The distribution of the maximum values of voxelwise prediction accuracy for individual participants was shown separately for each model (<bold>A</bold> and <bold>B</bold>: fastText; <bold>C</bold> and <bold>D</bold>: GloVe; <bold>E</bold> and <bold>F</bold>: word2vec; <bold>G</bold> and <bold>H</bold>: binary-labeling) and each dataset (<bold>A</bold>, <bold>C</bold>, <bold>E</bold>, and <bold>G</bold>: movie set 1; <bold>B</bold>, <bold>D</bold>, <bold>F</bold>, and <bold>H</bold>: movie set 2). The vertical dashed line in each panel indicates the mean value of the maximum prediction accuracy averaged across participants.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s011" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s011" xlink:type="simple">
<label>S11 Fig</label>
<caption>
<title>Cortical mapping of the fraction of significant voxels for other dimensions of fastText vectors.</title>
<p>The mean fraction of significant voxels in each brain region for fastText vector-based models was mapped onto the cortical surface (<bold>A</bold>, vector dimension = 100; <bold>B</bold>, 300; <bold>C</bold>, 500; <bold>D</bold>, 2000). The same conventions were used as in <xref ref-type="fig" rid="pcbi.1009138.g005">Fig 5</xref>. The five numbered cortical regions with the highest mean fraction of significant voxels for each dimension and dataset are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s034">S7 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s012" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s012" xlink:type="simple">
<label>S12 Fig</label>
<caption>
<title>Cortical mapping of the fraction of significant voxels for other dimensions of GloVe vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11 Fig</xref> but for GloVe vector-based models. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s035">S8 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s013" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s013" xlink:type="simple">
<label>S13 Fig</label>
<caption>
<title>Cortical mapping of the fraction of significant voxels for other dimensions of word2vec vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11 Fig</xref> but for word2vec vector-based models. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s036">S9 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s014" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s014" xlink:type="simple">
<label>S14 Fig</label>
<caption>
<title>Comparison of the fraction of significant voxels between original and control word vector-based models.</title>
<p>To test the significance of the fraction of significant voxels for original models, we compared the original fraction with the fraction of significant voxels for control models. The control models used word vectors randomly shuffled across vector dimensions for each vector and thereby produced the chance level of the fraction of significant voxels. The fraction of significant voxels in each cortical region for these original (y-axis) and control (x-axis) models is shown separately for each word-vector type (<bold>A</bold> and <bold>B</bold>: fastText; <bold>C</bold> and <bold>D</bold>: GloVe; <bold>E</bold> and <bold>F</bold>: word2vec) and each dataset (<bold>A</bold>, <bold>C</bold>, and <bold>E</bold>: movie set 1; <bold>B</bold>, <bold>D</bold>, and <bold>F</bold>: movie set 2). Each dot represents the fraction in each cortical region. Regardless of the word-vector types and datasets, the fraction for the original models was significantly higher than that for the control models in all the cortical regions (Wilcoxon test, p &lt; 0.00001, FDR corrected). This result indicates that the fraction of significant voxels for the original models is sufficiently above chance level.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s015" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s015" xlink:type="simple">
<label>S15 Fig</label>
<caption>
<title>Cortical mapping of prediction accuracy for binary-labeling models.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6 Fig</xref> but for binary-labeling models with the vector dimensionality of 1000 and 2000. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s040">S13 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s016" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s016" xlink:type="simple">
<label>S16 Fig</label>
<caption>
<title>Cortical mapping of the fraction of significant voxels for binary-labeling models.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11 Fig</xref> but for binary-labeling models with the vector dimensionality of 1000 and 2000. The numbered cortical regions are shown in <xref ref-type="supplementary-material" rid="pcbi.1009138.s041">S14 Table</xref>.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s017" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s017" xlink:type="simple">
<label>S17 Fig</label>
<caption>
<title>Comparison of model prediction performance in each cortical region between trained and untrained word vectors.</title>
<p>Prediction performance of voxelwise models (<bold>A</bold>, prediction accuracy; <bold>B</bold>, the fraction of significant voxels) in each cortical region was compared between trained (x-axis) and untrained word vectors (y-axis). The performance is shown separately for each word-vector type (top, fastText; middle, GloVe; bottom, word2vec) and each dataset (left in each of <bold>A</bold> and <bold>B</bold>, movie set 1; right, movie set 2). Each dot represents the mean performance averaged over participants for each region. The difference of model prediction performance between trained and untrained vectors was tested within each region while the performance for each participant was used as a data sample. Filled and open dots indicate that the region-wise difference was significant or not, respectively (Wilcoxon test, p &lt; 0.05, FDR corrected). Then, the difference of model prediction performance between trained and untrained vectors was tested across regions while the mean performance in each region was used as a data sample. The difference across regions were significant regardless of prediction performance measures, word-vector types, and datasets (Wilcoxon test, p &lt; 0.00001, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s018" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s018" xlink:type="simple">
<label>S18 Fig</label>
<caption>
<title>Correlations between dissimilarity matrices for all dimensions of fastText vectors.</title>
<p><bold>A</bold>) Brain–behavior and word vector–behavior correlations of noun dissimilarity matrices for different dimensions of fastText vectors. <bold>B</bold>) Difference between brain–behavior correlation coefficients of different vector dimensions for nouns. The color of each cell represents the coefficient difference of the dimension on the x-axis minus the dimension on the y-axis (red, positive values; blue, negative values). The mark in each cell indicates the statistical significance of the difference (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, n.s., p &gt; 0.05, FDR corrected). (<bold>C</bold> and <bold>D</bold>) The same analyses were performed as in <bold>A</bold> and <bold>B</bold> using adjective dissimilarity matrices.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s019" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s019" xlink:type="simple">
<label>S19 Fig</label>
<caption>
<title>Correlations between dissimilarity matrices for all dimensions of GloVe vectors.</title>
<p>The same analysis as in S18 but for GloVe vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s020" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s020" xlink:type="simple">
<label>S20 Fig</label>
<caption>
<title>Correlations between dissimilarity matrices for all dimensions of word2vec vectors.</title>
<p>The same analysis as in S18 but for word2vec vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s021" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s021" xlink:type="simple">
<label>S21 Fig</label>
<caption>
<title>Correlations between brain- and behavior-derived word dissimilarity matrices from the same participant population.</title>
<p>We constructed word dissimilarity matrices using behavioral and voxelwise-model data obtained from the same population of 6 participants for movie set 1. <bold>A</bold>, <bold>B</bold>) Behavior- and brain-derived dissimilarity matrices for nouns (<bold>A</bold>) and adjectives (<bold>B</bold>). The brain-derived matrices were obtained using 1000-dimensional fastText vectors. The same conventions were used as in <xref ref-type="fig" rid="pcbi.1009138.g006">Fig 6</xref>. <bold>C</bold>) Brain–behavior correlations for nouns and adjectives. All the correlation coefficients were significantly higher than chance level (permutation test, p &lt; 0.0001, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s022" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s022" xlink:type="simple">
<label>S22 Fig</label>
<caption>
<title>Brain–behavior correlations of word dissimilarity at the individual level.</title>
<p>We calculated the Spearman’s correlation between brain- and behavior-derived word dissimilarity matrices for each of the 6 participants who had both brain and behavioral data. The correlation coefficients are shown separately for nouns (<bold>A</bold>) and adjectives (<bold>B</bold>) and for each type of word vectors (dimensionality = 1000). Marks above bars indicate the statistical significance of the correlation coefficients (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s023" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s023" xlink:type="simple">
<label>S23 Fig</label>
<caption>
<title>Difference of brain–behavior correlation coefficients between each pair of different voxel selections for fastText vector-based models.</title>
<p>The color of each cell represents the difference between each pair of the number of selected voxels; the difference is the brain–behavior correlation coefficient for the dimension on the x-axis minus the coefficient for the dimension on the y-axis (red, positive values; blue, negative values). The mark in each cell indicates the statistical significance of the difference (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s024" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s024" xlink:type="simple">
<label>S24 Fig</label>
<caption>
<title>Difference of brain–behavior correlation coefficients between each pair of different voxel selections for GloVe vector-based models.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s023">S23 Fig</xref> but for GloVe vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s025" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s025" xlink:type="simple">
<label>S25 Fig</label>
<caption>
<title>Difference of brain–behavior correlation coefficients between each pair of different voxel selections for word2vec vector-based models.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s023">S23 Fig</xref> but for word2vec vectors.</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s026" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s026" xlink:type="simple">
<label>S26 Fig</label>
<caption>
<title>Brain–behavior correlation for each noun category.</title>
<p>The nouns used for constructing word dissimilarity matrices consisted of the six categories (humans, non-human animals, non-animal natural things, constructs, vehicles, and other artifacts; 10 nouns in each category). We calculated the correlation between behavior- and brain-derived word dissimilarity matrices obtained from each noun category. The Pearson’s correlation coefficients for each category were shown separately for each model (green, fastText; orange, GloVe; blue, word2vec) and for each movie set (<bold>A</bold>, movie set 1; <bold>B</bold>, movie set 2). Marks above bars indicate the statistical significance of the correlation coefficients (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s027" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s027" xlink:type="simple">
<label>S27 Fig</label>
<caption>
<title>Comparison of brain–behavior correlation with binary-labeling models.</title>
<p>The correlation of brain- and behavior-derived word dissimilarity matrices was computed for binary-labeling models. The correlation coefficients for these models (pink bars) are separately shown for nouns (top) and adjectives (bottom) and for the vector dimensionality of 300, 500, 1000, and 2000 (from left to right), along with the coefficients of brain–behavior correlation for the word vector-based models (green bars, fastText; orange bars, GloVe; blue bars, word2vec). In this case, because the vocabulary of the binary-labeling models changed depending on vector dimensionality and datasets (<xref ref-type="supplementary-material" rid="pcbi.1009138.s043">S16 Table</xref>), the number of nouns and adjectives used for calculating the brain-behavior correlation of both binary-labeling and word vector-based models were different across vector dimensions and datasets. Consequently, the correlation coefficients for the word vector-based models differed from those shown in other figures (Figs <xref ref-type="fig" rid="pcbi.1009138.g006">6</xref>–<xref ref-type="fig" rid="pcbi.1009138.g007">7</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s018">S18</xref>–<xref ref-type="supplementary-material" rid="pcbi.1009138.s020">S20</xref>). Marks below bars indicate the statistical significance of correlation coefficients above chance level whereas marks above bars indicate the statistical significance of the correlation difference between the binary-labeling model and each of the word vector-based models (permutation test, ***p &lt; 0.0001, **p &lt; 0.01, *p &lt; 0.05, FDR corrected).</p>
<p>(TIF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s028" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s028" xlink:type="simple">
<label>S1 Table</label>
<caption>
<title>Cortical regions with the highest mean prediction accuracy for each of the other dimensional fastText vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s029" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s029" xlink:type="simple">
<label>S2 Table</label>
<caption>
<title>Cortical regions with the highest mean prediction accuracy for each of the other dimensional GloVe vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s030" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s030" xlink:type="simple">
<label>S3 Table</label>
<caption>
<title>Cortical regions with the highest mean prediction accuracy for each of the other dimensional word2vec vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s031" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s031" xlink:type="simple">
<label>S4 Table</label>
<caption>
<title>Consistency of inter-regional patterns of prediction accuracy for fastText vectors across vector dimensionality and across movie sets.</title>
<p>We calculated the mean prediction accuracy of fastText vector-based models within each cortical region (Figs <xref ref-type="fig" rid="pcbi.1009138.g004">4</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s006">S6</xref>) and compared the inter-regional patterns of prediction accuracy between arbitrary pairs of vector dimensionality and movie sets. The value in each cell in the upper and lower part of the table denotes the Pearson’s or Spearman’s correlation coefficient, respectively, of the inter-regional patterns between each pair.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s032" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s032" xlink:type="simple">
<label>S5 Table</label>
<caption>
<title>Consistency of inter-regional patterns of prediction accuracy for GloVe vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s031">S4 Table</xref> but for GloVe vectors.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s033" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s033" xlink:type="simple">
<label>S6 Table</label>
<caption>
<title>Consistency of inter-regional patterns of prediction accuracy for word2vec vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s031">S4 Table</xref> but for word2vec vectors.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s034" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s034" xlink:type="simple">
<label>S7 Table</label>
<caption>
<title>Cortical regions with the highest mean fraction of significant voxels for each of the other dimensional fastText vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s035" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s035" xlink:type="simple">
<label>S8 Table</label>
<caption>
<title>Cortical regions with the highest mean fraction of significant voxels for each of the other dimensional GloVe vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s036" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s036" xlink:type="simple">
<label>S9 Table</label>
<caption>
<title>Cortical regions with the highest mean fraction of significant voxels for each of the other dimensional word2vec vectors.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s037" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s037" xlink:type="simple">
<label>S10 Table</label>
<caption>
<title>Consistency of inter-regional patterns of significant-voxel fraction for fastText vectors across vector dimensionality and across movie sets.</title>
<p>As in the case of the mean prediction accuracy (<xref ref-type="supplementary-material" rid="pcbi.1009138.s031">S4 Table</xref>), the inter-regional patterns of significant-voxel fraction of fastText vector-based models (Figs <xref ref-type="fig" rid="pcbi.1009138.g005">5</xref> and <xref ref-type="supplementary-material" rid="pcbi.1009138.s011">S11</xref>) were compared between arbitrary pairs of vector dimensionality and movie sets. The value in each cell in the upper and lower part of the table denotes the Pearson’s or Spearman’s correlation coefficient, respectively, of the inter-regional patterns between each pair.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s038" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s038" xlink:type="simple">
<label>S11 Table</label>
<caption>
<title>Consistency of inter-regional patterns of significant-voxel fraction for GloVe vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s037">S10 Table</xref> but for GloVe vectors.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s039" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s039" xlink:type="simple">
<label>S12 Table</label>
<caption>
<title>Consistency of inter-regional patterns of significant-voxel fraction for word2vec vectors.</title>
<p>The same analysis as in <xref ref-type="supplementary-material" rid="pcbi.1009138.s037">S10 Table</xref> but for word2vec vectors.</p>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s040" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s040" xlink:type="simple">
<label>S13 Table</label>
<caption>
<title>Cortical regions with the highest mean prediction accuracy for each of 1000- and 2000-dimensional binary-labeling models.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s041" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s041" xlink:type="simple">
<label>S14 Table</label>
<caption>
<title>Cortical regions with the highest mean fraction of significant voxels for each of 1000- and 2000-dimensional binary-labeling models.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s042" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s042" xlink:type="simple">
<label>S15 Table</label>
<caption>
<title>Cortical regions with the highest brain–behavior correlations of noun dissimilarity for each word-vector type.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s043" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s043" xlink:type="simple">
<label>S16 Table</label>
<caption>
<title>Cortical regions with the highest brain–behavior correlations of adjective dissimilarity for each word-vector type.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s044" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s044" xlink:type="simple">
<label>S17 Table</label>
<caption>
<title>Number of stimulus movie clips in individual product/service categories.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
<supplementary-material id="pcbi.1009138.s045" mimetype="image/tiff" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s045" xlink:type="simple">
<label>S18 Table</label>
<caption>
<title>Number of words contained in the vocabulary of the binary-labeling model.</title>
<p>(TIFF)</p>
</caption>
</supplementary-material>
</sec>
</body>
<back>
<ack>
<p>We thank Ms. Hitomi Koyama, Mr. Yusuke Nakano, Mr. Koji Takashima, Mr. Takeshi Matsuda, Mr. Susumu Minamiyama, Ms. Mami Yamashita, Mr. Ryo Yano, Ms. Risa Matsumoto, Mr. Masato Okino, and Mr. Akira Nagaoka for their analytical and experimental support.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="pcbi.1009138.ref001"><label>1</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Deerwester</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Dumais</surname> <given-names>ST</given-names></name>, <name name-style="western"><surname>Furnas</surname> <given-names>GW</given-names></name>, <name name-style="western"><surname>Landauer</surname> <given-names>TK</given-names></name>, <name name-style="western"><surname>Harshman</surname> <given-names>R</given-names></name>. <article-title>Indexing by latent semantic analysis</article-title>. <source>J Am Soc Inf Sci</source>. <year>1990</year>;<volume>41</volume>: <fpage>391</fpage>–<lpage>407</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref002"><label>2</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Mikolov</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Chen</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Corrado</surname> <given-names>G</given-names></name>, <name name-style="western"><surname>Dean</surname> <given-names>J</given-names></name>. <article-title>Distributed representations of words and phrases and their compositionality</article-title>. <source>Adv Neural Inf Process Syst</source>. <year>2013</year>;<volume>26</volume>: <fpage>3111</fpage>–<lpage>3119</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref003"><label>3</label><mixed-citation publication-type="other" xlink:type="simple">Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. Proceedings of the Empiricial Methods in Natural Language Processing (EMNLP 2014). 2014. pp. 1532–1543.</mixed-citation></ref>
<ref id="pcbi.1009138.ref004"><label>4</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Bojanowski</surname> <given-names>P</given-names></name>, <name name-style="western"><surname>Grave</surname> <given-names>E</given-names></name>, <name name-style="western"><surname>Joulin</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Mikolov</surname> <given-names>T</given-names></name>. <article-title>Enriching word vectors with subword information</article-title>. <source>Trans Assoc Comput Linguist</source>. <year>2017</year>;<volume>5</volume>: <fpage>135</fpage>–<lpage>146</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref005"><label>5</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Lund</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Burgess</surname> <given-names>C</given-names></name>. <article-title>Producing high-dimensional semantic spaces from lexical co-occurrence</article-title>. <source>Behav Res methods, instruments, Comput</source>. <year>1996</year>;<volume>28</volume>: <fpage>203</fpage>–<lpage>208</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref006"><label>6</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Mikolov</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Yih</surname> <given-names>W</given-names></name>, <name name-style="western"><surname>Zweig</surname> <given-names>G</given-names></name>. <article-title>Linguistic regularities in continuous space word representations</article-title>. <source>HLT-NAACL</source>. <year>2013</year>. pp. <fpage>746</fpage>–<lpage>751</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref007"><label>7</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Tang</surname> <given-names>D</given-names></name>, <name name-style="western"><surname>Wei</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Yang</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Zhou</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Liu</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Qin</surname> <given-names>B</given-names></name>. <article-title>Learning sentiment-specific word embedding</article-title>. <source>ACL</source>. <year>2014</year>. pp. <fpage>1555</fpage>–<lpage>1565</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref008"><label>8</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Li</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Pan</surname> <given-names>Q</given-names></name>, <name name-style="western"><surname>Yang</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Wang</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Tang</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Cambria</surname> <given-names>E</given-names></name>. <article-title>Learning word representations for sentiment analysis</article-title>. <source>Cognit Comput</source>. <year>2017</year>;<volume>9</volume>: <fpage>843</fpage>–<lpage>851</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref009"><label>9</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Wang</surname> <given-names>Z</given-names></name>, <name name-style="western"><surname>Ittycheriah</surname> <given-names>A</given-names></name>. <article-title>FAQ-based question answering via word alignment</article-title>. <source>arXiv</source>. <year>2015</year>; 1507.02628.</mixed-citation></ref>
<ref id="pcbi.1009138.ref010"><label>10</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Güçlü</surname> <given-names>U</given-names></name>, <name name-style="western"><surname>van Gerven</surname> <given-names>MAJ</given-names></name>. <article-title>Semantic vector space models predict neural responses to complex visual stimuli</article-title>. <source>arXiv</source>. <year>2015</year>; 1510.04738.</mixed-citation></ref>
<ref id="pcbi.1009138.ref011"><label>11</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Nishida</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Nishimoto</surname> <given-names>S</given-names></name>. <article-title>Decoding naturalistic experiences from human brain activity via distributed representations of words</article-title>. <source>Neuroimage</source>. <year>2018</year>;<volume>180</volume>: <fpage>232</fpage>–<lpage>242</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuroimage.2017.08.017" xlink:type="simple">10.1016/j.neuroimage.2017.08.017</ext-link></comment> <object-id pub-id-type="pmid">28801255</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref012"><label>12</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Nishida</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Matsumoto</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Yoshikawa</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Son</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Murakami</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Hayashi</surname> <given-names>R</given-names></name>, <etal>et al</etal>. <article-title>Reduced intra- and inter-individual diversity of semantic representations in the brains of schizophrenia patients</article-title>. <source>bioRxiv</source>. <year>2020</year>; 2020.06.03.132928. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1101/2020.06.03.132928" xlink:type="simple">10.1101/2020.06.03.132928</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref013"><label>13</label><mixed-citation publication-type="other" xlink:type="simple">Ruan Y-P, Ling Z-H, Hu Y. Exploring semantic representation in brain activity using word embeddings. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 2016. pp. 669–679.</mixed-citation></ref>
<ref id="pcbi.1009138.ref014"><label>14</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Wehbe</surname> <given-names>L</given-names></name>, <name name-style="western"><surname>Murphy</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Talukdar</surname> <given-names>P</given-names></name>, <name name-style="western"><surname>Fyshe</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Ramdas</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Mitchell</surname> <given-names>T</given-names></name>. <article-title>Simultaneously uncovering the patterns of brain regions involved in different story reading Subprocesses</article-title>. <source>PLoS One</source>. <year>2014</year>;<volume>9</volume>: <fpage>1</fpage>–<lpage>19</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1371/journal.pone.0112575" xlink:type="simple">10.1371/journal.pone.0112575</ext-link></comment> <object-id pub-id-type="pmid">25426840</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref015"><label>15</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Vodrahalli</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Chen</surname> <given-names>P-H</given-names></name>, <name name-style="western"><surname>Liang</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Baldassano</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Chen</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Yong</surname> <given-names>E</given-names></name>, <etal>et al</etal>. <article-title>Mapping between fMRI responses to movies and their natural language annotations</article-title>. <source>Neuroimage</source>. <year>2018</year>;<volume>180</volume>: <fpage>223</fpage>–<lpage>231</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuroimage.2017.06.042" xlink:type="simple">10.1016/j.neuroimage.2017.06.042</ext-link></comment> <object-id pub-id-type="pmid">28648889</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref016"><label>16</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Nishida</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Huth</surname> <given-names>AG</given-names></name>, <name name-style="western"><surname>Gallant</surname> <given-names>JL</given-names></name>, <name name-style="western"><surname>Nishimoto</surname> <given-names>S</given-names></name>. <article-title>Word statistics in large-scale texts explain the human cortical semantic representation of objects, actions, and impressions</article-title>. <source>Soc Neurosci Abstr</source>. <year>2015</year>;<volume>45</volume>: <fpage>333.13</fpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref017"><label>17</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Pereira</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Lou</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Pritchett</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Ritter</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Gershman</surname> <given-names>SJ</given-names></name>, <name name-style="western"><surname>Kanwisher</surname> <given-names>N</given-names></name>, <etal>et al</etal>. <article-title>Toward a universal decoder of linguistic meaning from brain activation</article-title>. <source>Nat Commun</source>. <year>2018</year>;<volume>9</volume>: <fpage>963</fpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1038/s41467-018-03068-4" xlink:type="simple">10.1038/s41467-018-03068-4</ext-link></comment> <object-id pub-id-type="pmid">29511192</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref018"><label>18</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Wang</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Zhang</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Wang</surname> <given-names>H</given-names></name>, <name name-style="western"><surname>Lin</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Zong</surname> <given-names>C</given-names></name>. <article-title>Fine-grained neural decoding with distributed word representations</article-title>. <source>Inf Sci (Ny)</source>. <year>2020</year>;<volume>507</volume>: <fpage>256</fpage>–<lpage>272</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.ins.2019.08.043" xlink:type="simple">10.1016/j.ins.2019.08.043</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref019"><label>19</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Kriegeskorte</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Mur</surname> <given-names>M</given-names></name>. <article-title>Inverse MDS: Inferring dissimilarity structure from multiple item arrangements</article-title>. <source>Front Psychol</source>. <year>2012</year>;<volume>3</volume>: <fpage>1</fpage>–<lpage>13</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fpsyg.2012.00001" xlink:type="simple">10.3389/fpsyg.2012.00001</ext-link></comment> <object-id pub-id-type="pmid">22279440</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref020"><label>20</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Naselaris</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Olman</surname> <given-names>CA</given-names></name>, <name name-style="western"><surname>Stansbury</surname> <given-names>DE</given-names></name>, <name name-style="western"><surname>Ugurbil</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Gallant</surname> <given-names>JL</given-names></name>. <article-title>A voxel-wise encoding model for early visual areas decodes mental images of remembered scenes</article-title>. <source>Neuroimage</source>. <year>2015</year>;<volume>105</volume>: <fpage>215</fpage>–<lpage>228</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuroimage.2014.10.018" xlink:type="simple">10.1016/j.neuroimage.2014.10.018</ext-link></comment> <object-id pub-id-type="pmid">25451480</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref021"><label>21</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Huth</surname> <given-names>AG</given-names></name>, <name name-style="western"><surname>de Heer</surname> <given-names>WA</given-names></name>, <name name-style="western"><surname>Griffiths</surname> <given-names>TL</given-names></name>, <name name-style="western"><surname>Theunissen</surname> <given-names>FE</given-names></name>, <article-title>Gallant, Jack L. Natural speech reveals the semantic maps that tile human cerebral cortex</article-title>. <source>Nature</source>. <year>2016</year>;<volume>532</volume>: <fpage>453</fpage>–<lpage>458</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1038/nature17637" xlink:type="simple">10.1038/nature17637</ext-link></comment> <object-id pub-id-type="pmid">27121839</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref022"><label>22</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Nishida</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Nakano</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Blanc</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Maeda</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Kado</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Nishimoto</surname> <given-names>S</given-names></name>. <article-title>Brain-mediated transfer learning of convolutional neural networks</article-title>. <source>Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence</source>. <year>2020</year>. pp. <fpage>5281</fpage>–<lpage>5288</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1609/aaai.v34i04.5974" xlink:type="simple">10.1609/aaai.v34i04.5974</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref023"><label>23</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Huth</surname> <given-names>AG</given-names></name>, <name name-style="western"><surname>Nishimoto</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Vu</surname> <given-names>AT</given-names></name>, <name name-style="western"><surname>Gallant</surname> <given-names>JL</given-names></name>. <article-title>A continuous semantic space describes the representation of thousands of object and action categories across the human brain</article-title>. <source>Neuron</source>. <year>2012</year>;<volume>76</volume>: <fpage>1210</fpage>–<lpage>1224</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuron.2012.10.014" xlink:type="simple">10.1016/j.neuron.2012.10.014</ext-link></comment> <object-id pub-id-type="pmid">23259955</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref024"><label>24</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Dale</surname> <given-names>AM</given-names></name>, <name name-style="western"><surname>Fischl</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Sereno</surname> <given-names>MI</given-names></name>. <article-title>Cortical surface-based analysis. I. Segmentation and surface reconstruction</article-title>. <source>Neuroimage</source>. <year>1999</year>;<volume>9</volume>: <fpage>179</fpage>–<lpage>194</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1006/nimg.1998.0395" xlink:type="simple">10.1006/nimg.1998.0395</ext-link></comment> <object-id pub-id-type="pmid">9931268</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref025"><label>25</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Fischl</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Sereno</surname> <given-names>MI</given-names></name>, <name name-style="western"><surname>Dale</surname> <given-names>AM</given-names></name>. <article-title>Cortical surface-based analysis. II: Inflation, flattening, and a surface-based coordinate system</article-title>. <source>Neuroimage</source>. <year>1999</year>;<volume>9</volume>: <fpage>195</fpage>–<lpage>207</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1006/nimg.1998.0396" xlink:type="simple">10.1006/nimg.1998.0396</ext-link></comment> <object-id pub-id-type="pmid">9931269</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref026"><label>26</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Destrieux</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Fischl</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Dale</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Halgren</surname> <given-names>E</given-names></name>. <article-title>Automatic parcellation of human cortical gyri and sulci using standard anatomical nomenclature</article-title>. <source>Neuroimage</source>. <year>2010</year>;<volume>53</volume>: <fpage>1</fpage>–<lpage>15</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuroimage.2010.06.010" xlink:type="simple">10.1016/j.neuroimage.2010.06.010</ext-link></comment> <object-id pub-id-type="pmid">20547229</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref027"><label>27</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Connolly</surname> <given-names>AC</given-names></name>, <name name-style="western"><surname>Guntupalli</surname> <given-names>JS</given-names></name>, <name name-style="western"><surname>Gors</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Hanke</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Halchenko</surname> <given-names>YO</given-names></name>, <name name-style="western"><surname>Wu</surname> <given-names>Y-C</given-names></name>, <etal>et al</etal>. <article-title>The representation of biological classes in the human brain</article-title>. <source>J Neurosci</source>. <year>2012</year>;<volume>32</volume>: <fpage>2608</fpage>–<lpage>2618</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1523/JNEUROSCI.5547-11.2012" xlink:type="simple">10.1523/JNEUROSCI.5547-11.2012</ext-link></comment> <object-id pub-id-type="pmid">22357845</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref028"><label>28</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Kriegeskorte</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Mur</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Ruff</surname> <given-names>DA</given-names></name>, <name name-style="western"><surname>Kiani</surname> <given-names>R</given-names></name>, <name name-style="western"><surname>Bodurka</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Esteky</surname> <given-names>H</given-names></name>, <etal>et al</etal>. <article-title>Matching Categorical Object Representations in Inferior Temporal Cortex of Man and Monkey</article-title>. <source>Neuron</source>. <year>2008</year>;<volume>60</volume>: <fpage>1126</fpage>–<lpage>1141</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.neuron.2008.10.043" xlink:type="simple">10.1016/j.neuron.2008.10.043</ext-link></comment> <object-id pub-id-type="pmid">19109916</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref029"><label>29</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Konkle</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Caramazza</surname> <given-names>A</given-names></name>. <article-title>Tripartite organization of the ventral stream by animacy and object size</article-title>. <source>J Neurosci</source>. <year>2013</year>;<volume>33</volume>: <fpage>10235</fpage>–<lpage>10242</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1523/JNEUROSCI.0983-13.2013" xlink:type="simple">10.1523/JNEUROSCI.0983-13.2013</ext-link></comment> <object-id pub-id-type="pmid">23785139</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref030"><label>30</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Hill</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Reichart</surname> <given-names>R</given-names></name>, <name name-style="western"><surname>Korhonen</surname> <given-names>A</given-names></name>. <article-title>SimLex-999: Evaluating Semantic Models With (Genuine) Similarity Estimation.</article-title> <source>Comput Linguist.</source> <year>2015</year>;<volume>41</volume>: <fpage>665</fpage>–<lpage>695</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1162/COLI%5Fa%5F00237" xlink:type="simple">10.1162/COLI_a_00237</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref031"><label>31</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Pereira</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Gershman</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Ritter</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Botvinick</surname> <given-names>M</given-names></name>. <article-title>A comparative evaluation of off-the-shelf distributed semantic representations for modelling behavioural data</article-title>. <source>Cogn Neuropsychol</source>. <year>2016</year>;<volume>33</volume>: <fpage>175</fpage>–<lpage>190</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/02643294.2016.1176907" xlink:type="simple">10.1080/02643294.2016.1176907</ext-link></comment> <object-id pub-id-type="pmid">27686110</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref032"><label>32</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Iordan</surname> <given-names>MC</given-names></name>, <name name-style="western"><surname>Giallanza</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Ellis</surname> <given-names>CT</given-names></name>, <name name-style="western"><surname>Beckage</surname> <given-names>NM</given-names></name>, <name name-style="western"><surname>Cohen</surname> <given-names>JD</given-names></name>. <article-title>Context Matters: Recovering Human Semantic Structure from Machine Learning Analysis of Large-Scale Text Corpora</article-title>. <source>arXiv</source>. <year>2020</year>; <year>1910</year>.06954.</mixed-citation></ref>
<ref id="pcbi.1009138.ref033"><label>33</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Hebart</surname> <given-names>MN</given-names></name>, <name name-style="western"><surname>Zheng</surname> <given-names>CY</given-names></name>, <name name-style="western"><surname>Pereira</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Baker</surname> <given-names>CI</given-names></name>. <article-title>Revealing the multidimensional mental representations of natural objects underlying human similarity judgements</article-title>. <source>Nat Hum Behav.</source> <year>2020</year>;<volume>4</volume>: <fpage>1173</fpage>–<lpage>1185</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1038/s41562-020-00951-3" xlink:type="simple">10.1038/s41562-020-00951-3</ext-link></comment> <object-id pub-id-type="pmid">33046861</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref034"><label>34</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Deniz</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Nunez-Elizalde</surname> <given-names>AO</given-names></name>, <name name-style="western"><surname>Huth</surname> <given-names>AG</given-names></name>, <name name-style="western"><surname>Gallant</surname> <given-names>JL</given-names></name>. <article-title>The representation of semantic information across human cerebral cortex during listening versus reading is invariant to stimulus modality</article-title>. <source>J Neurosci</source>. <year>2019</year>;<volume>39</volume>: <fpage>7722</fpage>–<lpage>7736</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1523/JNEUROSCI.0675-19.2019" xlink:type="simple">10.1523/JNEUROSCI.0675-19.2019</ext-link></comment> <object-id pub-id-type="pmid">31427396</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref035"><label>35</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Ivanova</surname> <given-names>AA</given-names></name>, <name name-style="western"><surname>Mineroff</surname> <given-names>Z</given-names></name>, <name name-style="western"><surname>Zimmerer</surname> <given-names>V</given-names></name>, <name name-style="western"><surname>Kanwisher</surname> <given-names>N</given-names></name>, <name name-style="western"><surname>Varley</surname> <given-names>R</given-names></name>, <name name-style="western"><surname>Fedorenko</surname> <given-names>E</given-names></name>. <article-title>The language network is recruited but not required for non-verbal semantic processing</article-title>. <source>bioRxiv</source>. <year>2019</year>; 696484. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1101/696484" xlink:type="simple">10.1101/696484</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref036"><label>36</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Boly</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Seth</surname> <given-names>AK</given-names></name>, <name name-style="western"><surname>Wilke</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Ingmundson</surname> <given-names>P</given-names></name>, <name name-style="western"><surname>Baars</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Laureys</surname> <given-names>S</given-names></name>, <etal>et al</etal>. <article-title>Consciousness in humans and non-human animals: recent advances and future directions</article-title>. <source>Front Psychol</source>. <year>2013</year>;<volume>4</volume>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fpsyg.2013.00625" xlink:type="simple">10.3389/fpsyg.2013.00625</ext-link></comment> <object-id pub-id-type="pmid">24198791</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref037"><label>37</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Cole</surname> <given-names>MW</given-names></name>, <name name-style="western"><surname>Ito</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Braver</surname> <given-names>TS</given-names></name>. <article-title>The behavioral relevance of task information in human prefrontal cortex</article-title>. <source>Cereb Cortex</source>. <year>2016</year>;<volume>26</volume>: <fpage>2497</fpage>–<lpage>2505</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1093/cercor/bhv072" xlink:type="simple">10.1093/cercor/bhv072</ext-link></comment> <object-id pub-id-type="pmid">25870233</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref038"><label>38</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Sakai</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Passingham</surname> <given-names>RE</given-names></name>. <article-title>Prefrontal set activity predicts rule-specific neural processing during subsequent cognitive performance</article-title>. <source>J Neurosci</source>. <year>2006</year>;<volume>26</volume>: <fpage>1211</fpage>–<lpage>8</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1523/JNEUROSCI.3887-05.2006" xlink:type="simple">10.1523/JNEUROSCI.3887-05.2006</ext-link></comment> <object-id pub-id-type="pmid">16436608</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref039"><label>39</label><mixed-citation publication-type="book" xlink:type="simple"><name name-style="western"><surname>Peters</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Neumann</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Iyyer</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Gardner</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Clark</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Lee</surname> <given-names>K</given-names></name>, <etal>et al</etal>. <source>Deep contextualized word representations. Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers).</source> <publisher-loc>Stroudsburg, PA, USA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>; <year>2018</year>. pp. <fpage>2227</fpage>–<lpage>2237</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.18653/v1/N18-1202" xlink:type="simple">10.18653/v1/N18-1202</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref040"><label>40</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Devlin</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Chang</surname> <given-names>M-W</given-names></name>, <name name-style="western"><surname>Lee</surname> <given-names>K</given-names></name>, <name name-style="western"><surname>Toutanova</surname> <given-names>K</given-names></name>. <article-title>BERT: Pre-training of deep bidirectional transformers for language understanding</article-title>. <source>arXiv.</source> <year>2018</year>; <year>1810</year>.04805.</mixed-citation></ref>
<ref id="pcbi.1009138.ref041"><label>41</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Yang</surname> <given-names>Z</given-names></name>, <name name-style="western"><surname>Dai</surname> <given-names>Z</given-names></name>, <name name-style="western"><surname>Yang</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Carbonell</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Salakhutdinov</surname> <given-names>RR</given-names></name>, <name name-style="western"><surname>Le Q</surname> <given-names>V</given-names></name>. <article-title>XLNet: Generalized autoregressive pretraining for language understanding</article-title>. <source>Advances in neural information processing systems</source>. <year>2019</year>. pp. <fpage>5753</fpage>–<lpage>5763</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref042"><label>42</label><mixed-citation publication-type="book" xlink:type="simple"><name name-style="western"><surname>Jain</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Huth</surname> <given-names>AG</given-names></name>. <chapter-title>Incorporating context into language encoding models for fMRI</chapter-title>. <source>Proceedings of the 32Nd International Conference on Neural Information Processing Systems</source>. <publisher-loc>USA</publisher-loc>: <publisher-name>Curran Associates Inc.</publisher-name>; <year>2018</year>. pp. <fpage>6629</fpage>–<lpage>6638</lpage>.</mixed-citation></ref>
<ref id="pcbi.1009138.ref043"><label>43</label><mixed-citation publication-type="book" xlink:type="simple"><name name-style="western"><surname>Jat</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Tang</surname> <given-names>H</given-names></name>, <name name-style="western"><surname>Talukdar</surname> <given-names>P</given-names></name>, <name name-style="western"><surname>Mitchell</surname> <given-names>T</given-names></name>. <chapter-title>Relating simple sentence representations in deep neural networks and the brain</chapter-title>. <source>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.</source> <publisher-loc>Stroudsburg, PA, USA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>; <year>2019</year>. pp. <fpage>5137</fpage>–<lpage>5154</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.18653/v1/P19-1507" xlink:type="simple">10.18653/v1/P19-1507</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref044"><label>44</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Schrimpf</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Blank</surname> <given-names>I</given-names></name>, <name name-style="western"><surname>Tuckute</surname> <given-names>G</given-names></name>, <name name-style="western"><surname>Kauf</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Hosseini</surname> <given-names>EA</given-names></name>, <name name-style="western"><surname>Kanwisher</surname> <given-names>N</given-names></name>, <etal>et al</etal>. <article-title>Artificial neural networks accurately predict language processing in the brain</article-title>. <source>bioRxiv</source>. <year>2020</year>; 2020.06.26.174482. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1101/2020.06.26.174482" xlink:type="simple">10.1101/2020.06.26.174482</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref045"><label>45</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Doshi-Velez</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Kim</surname> <given-names>B</given-names></name>. <article-title>Towards a rigorous science of interpretable machine learning</article-title>. <source>arXiv</source>. <year>2017</year>; 1702.08608.</mixed-citation></ref>
<ref id="pcbi.1009138.ref046"><label>46</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Guidotti</surname> <given-names>R</given-names></name>, <name name-style="western"><surname>Monreale</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Ruggieri</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Turini</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Giannotti</surname> <given-names>F</given-names></name>, <name name-style="western"><surname>Pedreschi</surname> <given-names>D</given-names></name>. <article-title>A survey of methods for explaining black box models</article-title>. <source>ACM Comput Surv</source>. <year>2019</year>;<volume>51</volume>: <fpage>1</fpage>–<lpage>42</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/3236009" xlink:type="simple">10.1145/3236009</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref047"><label>47</label><mixed-citation publication-type="book" xlink:type="simple"><name name-style="western"><surname>Ling</surname> <given-names>W</given-names></name>, <name name-style="western"><surname>Dyer</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Black</surname> <given-names>AW</given-names></name>, <name name-style="western"><surname>Trancoso</surname> <given-names>I</given-names></name>. <source>Two/too simple adaptations of Word2Vec for syntax problems. Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>. <publisher-loc>Stroudsburg, PA, USA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>; <year>2015</year>. pp. <fpage>1299</fpage>–<lpage>1304</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3115/v1/N15-1142" xlink:type="simple">10.3115/v1/N15-1142</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref048"><label>48</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Faruqui</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Dodge</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Jauhar</surname> <given-names>SK</given-names></name>, <name name-style="western"><surname>Dyer</surname> <given-names>C</given-names></name>, <name name-style="western"><surname>Hovy</surname> <given-names>E</given-names></name>, <name name-style="western"><surname>Smith</surname> <given-names>NA</given-names></name>. <article-title>Retrofitting word vectors to semantic lexicons</article-title>. <source>arXiv</source>. <year>2014</year>; 1411.4166.</mixed-citation></ref>
<ref id="pcbi.1009138.ref049"><label>49</label><mixed-citation publication-type="book" xlink:type="simple"><name name-style="western"><surname>Liu</surname> <given-names>Q</given-names></name>, <name name-style="western"><surname>Jiang</surname> <given-names>H</given-names></name>, <name name-style="western"><surname>Wei</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Ling</surname> <given-names>Z-H</given-names></name>, <name name-style="western"><surname>Hu</surname> <given-names>Y</given-names></name>. <chapter-title>Learning semantic word embeddings based on ordinal knowledge constraints</chapter-title>. <source>Proceedings of ACL</source>. <publisher-loc>Stroudsburg, PA, USA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>; <year>2015</year>. pp. <fpage>1501</fpage>–<lpage>1511</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3115/v1/P15-1145" xlink:type="simple">10.3115/v1/P15-1145</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref050"><label>50</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Fyshe</surname> <given-names>A</given-names></name>, <name name-style="western"><surname>Talukdar</surname> <given-names>PP</given-names></name>, <name name-style="western"><surname>Murphy</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Mitchell</surname> <given-names>TM</given-names></name>. <article-title>Interpretable semantic vectors from a joint model of brain- and text-based meaning</article-title>. <source>ACL</source>. <year>2014</year>. pp. <fpage>489</fpage>–<lpage>499</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3115/v1/p14-1046" xlink:type="simple">10.3115/v1/p14-1046</ext-link></comment> <object-id pub-id-type="pmid">26166940</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref051"><label>51</label><mixed-citation publication-type="other" xlink:type="simple">Barrett M, Bingel J, Hollenstein N, Rei M, Søgaard A. Sequence classification with human attention. CoNLL 2018 - 22nd Conf Comput Nat Lang Learn Proc. 2018; 302–312. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.18653/v1/k18-1030" xlink:type="simple">10.18653/v1/k18-1030</ext-link></comment></mixed-citation></ref>
<ref id="pcbi.1009138.ref052"><label>52</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Moeller</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Yacoub</surname> <given-names>E</given-names></name>, <name name-style="western"><surname>Olman</surname> <given-names>CA</given-names></name>, <name name-style="western"><surname>Auerbach</surname> <given-names>E</given-names></name>, <name name-style="western"><surname>Strupp</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Harel</surname> <given-names>N</given-names></name>, <etal>et al</etal>. <article-title>Multiband multislice GE-EPI at 7 tesla, with 16-fold acceleration using partial parallel imaging with application to high spatial and temporal whole-brain fMRI</article-title>. <source>Magn Reson Med</source>. <year>2010</year>;<volume>63</volume>: <fpage>1144</fpage>–<lpage>1153</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/mrm.22361" xlink:type="simple">10.1002/mrm.22361</ext-link></comment> <object-id pub-id-type="pmid">20432285</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref053"><label>53</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Nishimoto</surname> <given-names>S</given-names></name>, <name name-style="western"><surname>Vu</surname> <given-names>AT</given-names></name>, <name name-style="western"><surname>Naselaris</surname> <given-names>T</given-names></name>, <name name-style="western"><surname>Benjamini</surname> <given-names>Y</given-names></name>, <name name-style="western"><surname>Yu</surname> <given-names>B</given-names></name>, <name name-style="western"><surname>Gallant</surname> <given-names>JL</given-names></name>. <article-title>Reconstructing visual experiences from brain activity evoked by natural movies</article-title>. <source>Curr Biol</source>. <year>2011</year>;<volume>21</volume>: <fpage>1641</fpage>–<lpage>1646</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.cub.2011.08.031" xlink:type="simple">10.1016/j.cub.2011.08.031</ext-link></comment> <object-id pub-id-type="pmid">21945275</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref054"><label>54</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Mur</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Meys</surname> <given-names>M</given-names></name>, <name name-style="western"><surname>Bodurka</surname> <given-names>J</given-names></name>, <name name-style="western"><surname>Goebel</surname> <given-names>R</given-names></name>, <name name-style="western"><surname>Bandettini</surname> <given-names>PA</given-names></name>, <name name-style="western"><surname>Kriegeskorte</surname> <given-names>N</given-names></name>. <article-title>Human object-similarity judgments reflect and transcend the primate-IT object representation</article-title>. <source>Front Psychol</source>. <year>2013</year>;<volume>4</volume>: <fpage>1</fpage>–<lpage>22</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fpsyg.2013.00001" xlink:type="simple">10.3389/fpsyg.2013.00001</ext-link></comment> <object-id pub-id-type="pmid">23382719</object-id></mixed-citation></ref>
<ref id="pcbi.1009138.ref055"><label>55</label><mixed-citation publication-type="journal" xlink:type="simple"><name name-style="western"><surname>Charest</surname> <given-names>I</given-names></name>, <name name-style="western"><surname>Kievit</surname> <given-names>RA</given-names></name>, <name name-style="western"><surname>Schmitz</surname> <given-names>TW</given-names></name>, <name name-style="western"><surname>Deca</surname> <given-names>D</given-names></name>, <name name-style="western"><surname>Kriegeskorte</surname> <given-names>N</given-names></name>. <article-title>Unique semantic space in the brain of each beholder predicts perceived similarity</article-title>. <source>Proc Natl Acad Sci U S A</source>. <year>2014</year>;<volume>111</volume>: <fpage>14545</fpage>–<lpage>14570</lpage>. <comment>doi: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1073/pnas.1402594111" xlink:type="simple">10.1073/pnas.1402594111</ext-link></comment> <object-id pub-id-type="pmid">25246586</object-id></mixed-citation></ref>
</ref-list>
</back>
<sub-article article-type="aggregated-review-documents" id="pcbi.1009138.r001" specific-use="decision-letter">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r001</article-id>
<title-group>
<article-title>Decision Letter 0</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name name-style="western">
<surname>Graham</surname>
<given-names>Lyle J.</given-names>
</name>
<role>Deputy Editor</role>
</contrib>
<contrib contrib-type="author">
<name name-style="western">
<surname>Cai</surname>
<given-names>Ming Bo</given-names>
</name>
<role>Associate Editor</role>
</contrib>
</contrib-group>
<permissions>
<copyright-year>2021</copyright-year>
<copyright-holder>Graham, Cai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="article" id="rel-obj001" link-type="peer-reviewed-article"/>
<custom-meta-group>
<custom-meta>
<meta-name>Submission Version</meta-name>
<meta-value>0</meta-value>
</custom-meta>
</custom-meta-group>
</front-stub>
<body>
<p>
<named-content content-type="letter-date">1 Dec 2020</named-content>
</p>
<p>Dear Dr. Nishida,</p>
<p>Thank you very much for submitting your manuscript "Behavioral correlates of cortical semantic representations modeled by word vectors" for consideration at PLOS Computational Biology.</p>
<p>As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.</p>
<p>Dear Dr. Nishida and co-authors,</p>
<p>I am very glad to handle this submission. Now three reviews have returned for the manuscript. Overall, reviewers found the paper interesting. However, they also raised important issues. I agree with the reviewers that these issues should be addressed before the paper can be considered for acceptance.</p>
<p>One concern is novelty. Two reviewers worry that the paper has not provided significant new insights and have pointed out relevant literature that should be compared to explain the novelty more clearly.</p>
<p>Another is the validity of methodology and sufficiency of analysis. Reviewer 1 and 2 both asked for additional analysis that can help readers better evaluate the conclusion. Clarification of some of the analysis is needed before they can assess the validity of the claims made. I agree with all these and think they should be carefully addressed. When I read the manuscript, I had the same question as the second point in concern (2) of reviewer 1: what the neural similarity would look like if simple binary labels of the existence of object corresponding to the nouns are used to build design matrix and estimate neural patterns. Please also address other issues raised.</p>
<p>I noticed that fraction of significant voxels, accuracy, dissimilarity matrix have been provided in the data repository which allows plotting the figures in the manuscript. However, such quantities still do not allow readers to reproduce the findings (including these derived quantities themselves) without the data underlying them. For example, several questions of Reviewers #1 and #2 cannot be answered based on these derived quantities. PLOS journals require authors to make all data necessary to replicate their study’s findings publicly available without restriction at the time of publication. Please follow the Data Availability policy of PLOS (<ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/data-availability" xlink:type="simple">https://journals.plos.org/ploscompbiol/s/data-availability</ext-link>) to provide data that allows reproduction of the findings in the manuscript.</p>
<p>We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.</p>
<p>When you are ready to resubmit, please upload the following:</p>
<p>[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.</p>
<p>[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).</p>
<p>Important additional instructions are given below your reviewer comments.</p>
<p>Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.</p>
<p>Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.</p>
<p>Sincerely,</p>
<p>Ming Bo Cai</p>
<p>Associate Editor</p>
<p>PLOS Computational Biology</p>
<p>Lyle Graham</p>
<p>Deputy Editor</p>
<p>PLOS Computational Biology</p>
<p>***********************</p>
<p>Reviewer's Responses to Questions</p>
<p><bold>Comments to the Authors:</bold></p>
<p><bold>Please note here if the review is uploaded as an attachment.</bold></p>
<p>Reviewer #1: Summary</p>
<p>The authors investigate the extent to which artificial models of language (NLP) can predict human neural responses and align with human behavioral judgments. They find significant agreement between brain, model, and behavior and conclude that current word embedding models are able to capture a meaningful amount of variance of both neural representation and semantic perception in humans.</p>
<p>Overall Impressions</p>
<p>I think this question is very interesting and I think that progress in this domain would be very welcome and useful for everyone who is interested in semantics, neural correlates of perception, and/or prediction of neural representations (myself included). I also like the approach (which is consistent with prior work), but I’m not fully convinced that the present results provide significant new insights into the problem in their current form. My main concerns about the current manuscript relate to the choice of embedding model, the lower effect sizes compared to similar prior work, and not fully accounting for known neural biases in how the effects are analyzed. However, I believe that these concerns could be fully addressed by additional analyses and discussion (as detailed below).</p>
<p>Major Comments</p>
<p>(1) The embedding model selected has weaker performance than openly available alternatives: Pereira et al. 2018 shows that multiple skip-gram models (e.g., word-2-vec, which is purportedly not as ‘good’ as fastText) can predict concept similarity up to r~=0.70. There’s also some recent work from Iordan et al. 2019 that pushes this to r~=0.90 for certain categories. In the current work, fastText is only at r~=0.30 with behavior and the authors argue that this means the brain mapping helps (r=~0.60), but it may be that other word embedding models can already do much better than any brain mapping at predicting behavior for this particular set of adjectives and/or nouns, but that the authors did not pick a good enough embedding model. To provide evidence that this is not the case, the authors could (1) test other embedding models listed above and show that they have similar behavioral performance for their set of concepts; and/or (2) show that at least one such other model (especially if it outperforms fastText) can also give improved predictions when used to predict neural representations similar to the current analysis.</p>
<p>(2) Neural pattern prediction accuracy is much lower and spatially different from prior work: The pattern of model performance (Fig. 4, top) is very different from that of prior work and performance is much lower (i.e., r~=0.10-0.30 compared to e.g. Fig. 1 of Huth et al. 2016, r=0.60 in high-level regions). Spatially, in the present work, prediction is highest in LO, which is below chance in the Huth language model; also, current model has little to no significant frontal cortex predictions. One could argue that this is because Huth et al. 2016 tries to use a language model to predict auditory stories, not movies. But then, performance of the current model is even weaker when compared to Huth et al. 2012 (r values up to 0.80) which uses simple binary labelling of the objects occurring in videos to predict neural response. Hence, I am not convinced that this language model is the best approach to model visual content, e.g., complex movies. To provide evidence that this is a good/comparable/better alternative than existing methods, the authors could use their current data to automatically generate a binary matrix of objects (similar to Huth et al. 2012) for every timepoint and use this to perform neural prediction, instead of using the word vector representations for the objects/concepts. If this matches/improves performance, then this could show evidence that the word vector approach either doesn’t add anything or may even hinder recovery of semantic information by introducing external biases.</p>
<p>(3) Correlations don’t control for the known animate/inanimate distinction in VTC: One of the most salient distinctions in patterns of activity in occipito-temporal cortex is between animate and inanimate stimuli/categories. From the list on line 522, I see a clear human-animal-manmade clustering of nouns which is also consistent with the behavioral matrix in Fig. 4. If the model simply picks up on the highest variance dimension of representation across cortical patterns (e.g., animacy), then this could drive most of the correlation effects observed. However, we already know this animacy effect is present in the brain (Connolly et al. 2012, Konkle &amp; Caramazza 2013, Kriegeskorte et al. 2008, etc.), so, in this case, this model would tell us nothing new about semantic knowledge representation in the brain. To address this issue, the authors could try to characterize and control for the effects of animacy (and the human-animal distinction within that) when measuring how well the model predicts behavior and/or the brain (i.e., regress this distinction out or focus their analysis on similarity matrices within these superordinate categories). If the effects survive, then this would indeed tell us that the models capture something interesting and novel about the semantic structure of neural representations and their connections with behavioral measures / perception, beyond animacy.</p>
<p>Minor Comments</p>
<p>(4) line 145: point to the corresponding supplementary figure (S5?). Authors could also consider running a Friedman test to assess non-monotonicity statistically.</p>
<p>(5) line 151: how do you define ‘appropriate’? It would be helpful if you mentioned briefly in the text that you used Freesurfer anatomical segmentation.</p>
<p>(6) Fig. 4: it would be helpful to have labels for these brain regions in the caption / text / supplement. Is the yellow one in a/b LO?</p>
<p>(7) Fig. 4: A voxel-wise map similar to Huth et al. 2012/2016 would be helpful here to compare how this method can model semantic information in the brain (i.e., same figure, but with voxel-level granularity, not just Freesurfer ROI).</p>
<p>(8) line 273: the effect sizes are really low (1-2% neural variance explained) and the authors are over-inflating the value of the procedure here.</p>
<p>(9) It would be useful to discuss why you chose fastText over other models. This could also be addressed in conjuction with Major Comment #1 above.</p>
<p>(10) Multiband factor 6 is quite high and the TR=1s is quite fast -- this may cause very low SNR (which maybe would explain low r values compared to similar work?). Is there a reason you did not use a more conventional fMRI sequence (e.g., multiband 2/4, 1.5-2s TR)?</p>
<p>(11) line 228: The authors should be a little more careful with the strength of the conclusions they draw from their results.</p>
<p>(12) line 310: I don’t agree with the argument that the language regions are not involved in processing this type of movie info – I suspect that an activation GLM analysis on the movie data would show weak, but consistent activation in language regions that the current model cannot capture.</p>
<p>(13) line 330: I don't necessarily agree that word embeddings are more explainable than deep network representations; they are just as opaque. I tried to find work supporting the authors’ claim but couldn’t. Please provide some references here and expand on this explanation or please reword.</p>
<p>Reviewer #2: I have now had a chance for an in-depth review of “Behavioral correlates of cortical semantic representations modeled by word vectors” by Nishida et al. The manuscript is well written and the overall problem domain is derived clearly. In this project, the authors use encoding models based on a FastText embedding space to predict fMRI brain data and subsequently use the predicted beta values to compute a representational dissimilarity matrix (RDM) for a novel set of words. The resulting RDM is compared to a behavioural RDM obtained via inverse MDS on the same words. Using this elaborate pipeline, the authors come to the conclusion that the predicted brain data and behavioural judgments are correlated, indicating the feasibility of using semantic embedding spaces for brain and behavioural modelling.</p>
<p>The overall approach is interesting, especially since the brain data was obtained via a visual/sensory modality, whereas the encoding model is derived from more abstract textual scene descriptions. To my understanding this cross-domain encoding approach acts as a “filtering” function that highlights brain regions related to semantics rather than low-level visual brain regions. This strength of the paper could have been highlighted a bit more. As stated above, the paper is well written, and computationally elegant. However, I have a few comments, as detailed below, which I hope will help the authors improve their work.</p>
<p>Major 1: Take-home message and novelty</p>
<p>While I see the elegance and computational complexity of the author’s endeavour, I am a bit lost as to what the actual take home message of the work may be and where and how the authors advance beyond what is known about the brain from prior work. I therefore invite the authors to elaborate on this.</p>
<p>At multiple occasions, the authors underline the novelty of the work, stating that “[…] there has been no study explicitly examining whether the modeled semantic representations actually capture our perception of semantic information.” and “However, whether the semantic representations modeled by word vectors accurately reflect the semantic perception of humans is yet to be determined.". However, there is a large body of research investigating exactly these questions (see [1-6] for examples of previous work most of which remain uncited).</p>
<p>Major 2: Technical validity</p>
<p>Going through the manuscript, I found myself continuously wondering about the exact statistical procedures run for the different tests described. This made it quite difficult to judge the actual merit of the claims made. Below are examples of cases that need technical clarification/adjustments:</p>
<p>a. The authors write “the fraction of significant voxels that reach their prediction accuracy to a threshold value”. It remains unclear, however, how significance was established, what the threshold value was, what statistical test was used, which data was tested, etc. As a second example, the authors write “Only regions with mean prediction accuracy above 0.11 (corresponding to p = 0.0001 ~ 0.05/150 regions) are shown”. Can the authors clarify what exactly they did here to derive the statistics?</p>
<p>b. The minimum fraction of significant voxels within any individual region is reported as 0.237. This seems very high, given that the average model prediction accuracy across all voxels is reported as “0.09”. To better understand this discrepancy, I would like the authors to comment on (a) how exactly the test set was derived, (b) how exactly the authors tested for significance (see above comment), and (c) how many training/test datapoints were used. Secondly, testing against chance, after having fitted hundreds of free parameters, seems like a somewhat low bar (although I acknowledge the fact that this is prediction on test data). As a more conservative control, I suggest the authors use an untrained embedding network as a control model as this would give the claims for actual semantic embedding throughout the cortex more explanatory power/merit.</p>
<p>c. “There was a strong correlation of mean prediction accuracy over 150 cortical regions between movie sets 1 and 2 (Pearson’s r = 0.974);“ Please repeat this analysis with a robust Spearman correlation, as deviations from normality (such as outliers) can easily increase the Pearson correlation coefficient but this is not meaningful. Please note that nearby regions are autocorrelated, which may violate the independence assumption of the classic Pearson statistic.</p>
<p>d. To my understanding, the significance of correlations between RDMs (Figure 5) is computed via bootstrapping individual RDM cells. I do not think that this is technically valid, as the RDM cells have dependencies that are violated by this approach (the N*N-1 cells are all derived from only N experimental conditions). The more standard approach, which I advise the authors to take, would be to perform a permutation test on the RDM conditions (i.e. permuting the rows/columns of one of the two RDMs to derive a null distribution against which the empirical correlation value can be tested). Moreover, Spearman correlations are more commonly used for RDM comparisons across domains.</p>
<p>Minor:</p>
<p>1. Which exact atlas was used? This is not mentioned anywhere. Relatedly, were the 150 regions tested all regions of the atlas, or were ROIs excluded?</p>
<p>2. Please consider citing [7], which is a novel paper in the domain.</p>
<p>3. Could the authors attempt a spatial localization of brain-behavioural semantic processing by computing RDMs for each ROI and correlate it with the behavioural data?</p>
<p>I sincerely hope that the above comments will be perceived as constructive and that they may help assist the authors in making their contribution stronger. It is an exciting project and overall computational approach.</p>
<p>Signed</p>
<p>Tim Kietzmann</p>
<p>[1] <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/pdf/1910.06954.pdf" xlink:type="simple">https://arxiv.org/pdf/1910.06954.pdf</ext-link></p>
<p>[2] Mikolov, T., Yih, S. W. &amp; Zweig, G. Linguistic Regularities in Continuous Space Word Representations. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 746-751 (2013).</p>
<p>[3] <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1408.3456" xlink:type="simple">https://arxiv.org/abs/1408.3456</ext-link></p>
<p>[4] Pereira, F., Gershman, S., Ritter, S. &amp; Botvinick, M. A comparative evaluation of off- the-shelf distributed semantic representations for modelling behavioural data. Cogn. Neuropsychol. 33, 175–190 (2016).</p>
<p>[5] <ext-link ext-link-type="uri" xlink:href="https://pubmed.ncbi.nlm.nih.gov/29777825/" xlink:type="simple">https://pubmed.ncbi.nlm.nih.gov/29777825/</ext-link></p>
<p>[6] <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/pdf/1805.07644.pdf" xlink:type="simple">https://arxiv.org/pdf/1805.07644.pdf</ext-link></p>
<p>[7] <ext-link ext-link-type="uri" xlink:href="https://www.nature.com/articles/s41562-020-00951-3" xlink:type="simple">https://www.nature.com/articles/s41562-020-00951-3</ext-link></p>
<p>Reviewer #3: Although a number of studies demonstrated correspondences between brain activity and word embeddings, the behavioral validation of these studies has been lacking. The authors address this issue by using representational similarity analysis on behavioral data and fMRI data in order the compare the two, revealing a significant correlation. It is a clearly written manuscript with well-designed experiments and analyses. I really liked that it has a clear message and delivers it well without going into tangential research questions which would distract from the main results. I would be happy to recommend accepting the manuscript for publication after the few minor issues that I list below are addressed.</p>
<p>- It is not immediately clear how many participants were involved in the study and which tasks each participant performed. Although this information is given later in the methods section, it appears rather late in the manuscript. A table or a few sentences to clearly state this earlier in the manuscript could make it easier for the reader. On another note, just out of curiosity, I see that the number of participants was quite large in comparison to what we are used to seeing in similar studies. Was there a specific motivation for collecting such a large fMRI dataset?</p>
<p>- Not a big deal, but in Figure 4, the colors that are overlaid on to the brain surfaces are transparent, which makes it a bit confusing to read the results with the brain surface also having brightness differences. Making the results opaque would be preferable, in my opinion. At the same time, I see the importance of conveying the anatomical information to the reader, so I leave it to the authors to decide what to do with this comment.</p>
<p>- How was the variability of the RDMs between participants? Have the analyses been performed also at the individual level (e.g. comparing each participant’s behavioral data with their own brain data)? If not, why not?</p>
<p>**********</p>
<p><bold>Have all data underlying the figures and results presented in the manuscript been provided?</bold></p>
<p>Large-scale datasets should be made available via a public repository as described in the <italic>PLOS Computational Biology</italic> <ext-link ext-link-type="uri" xlink:href="http://journals.plos.org/ploscompbiol/s/data-availability" xlink:type="simple">data availability policy</ext-link>, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.</p>
<p>Reviewer #1: Yes</p>
<p>Reviewer #2: Yes</p>
<p>Reviewer #3: Yes</p>
<p>**********</p>
<p>PLOS authors have the option to publish the peer review history of their article (<ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/editorial-and-peer-review-process#loc-peer-review-history" xlink:type="simple">what does this mean?</ext-link>). If published, this will include your full peer review and any attached files.</p>
<p>If you choose “no”, your identity will remain anonymous but your review may still be made public.</p>
<p><bold>Do you want your identity to be public for this peer review?</bold> For information about this choice, including consent withdrawal, please see our <ext-link ext-link-type="uri" xlink:href="https://www.plos.org/privacy-policy" xlink:type="simple">Privacy Policy</ext-link>.</p>
<p>Reviewer #1: No</p>
<p>Reviewer #2: <bold>Yes: </bold>Tim C Kietzmann</p>
<p>Reviewer #3: No</p>
<p><underline>Figure Files:</underline></p>
<p>While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, <underline><ext-link ext-link-type="uri" xlink:href="https://pacev2.apexcovantage.com/" xlink:type="simple">https://pacev2.apexcovantage.com</ext-link></underline>. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at <underline><email xlink:type="simple">figures@plos.org</email></underline>.</p>
<p><underline>Data Requirements:</underline></p>
<p>Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: <ext-link ext-link-type="uri" xlink:href="http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5" xlink:type="simple">http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5</ext-link>.</p>
<p><underline>Reproducibility:</underline></p>
<p>To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions, please see <underline><ext-link ext-link-type="uri" xlink:href="http://journals.plos.org/plospathogens/s/submission-guidelines" xlink:type="simple">http://journals.plos.org/compbiol/s/submission-guidelines#loc-materials-and-methods</ext-link></underline></p>
</body>
</sub-article>
<sub-article article-type="author-comment" id="pcbi.1009138.r002">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r002</article-id>
<title-group>
<article-title>Author response to Decision Letter 0</article-title>
</title-group>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="peer-reviewed-article" id="rel-obj002" link-type="rebutted-decision-letter" object-id="10.1371/journal.pcbi.1009138.r001" object-id-type="doi" object-type="decision-letter"/>
<custom-meta-group>
<custom-meta>
<meta-name>Submission Version</meta-name>
<meta-value>1</meta-value>
</custom-meta>
</custom-meta-group>
</front-stub>
<body>
<p>
<named-content content-type="author-response-date">31 Mar 2021</named-content>
</p>
<supplementary-material id="pcbi.1009138.s046" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s046" xlink:type="simple">
<label>Attachment</label>
<caption>
<p>Submitted filename: <named-content content-type="submitted-filename">Nishida-BehavCorrSemantic-ResponseLetter02.docx</named-content></p>
</caption>
</supplementary-material>
</body>
</sub-article>
<sub-article article-type="aggregated-review-documents" id="pcbi.1009138.r003" specific-use="decision-letter">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r003</article-id>
<title-group>
<article-title>Decision Letter 1</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name name-style="western">
<surname>Graham</surname>
<given-names>Lyle J.</given-names>
</name>
<role>Deputy Editor</role>
</contrib>
<contrib contrib-type="author">
<name name-style="western">
<surname>Cai</surname>
<given-names>Ming Bo</given-names>
</name>
<role>Associate Editor</role>
</contrib>
</contrib-group>
<permissions>
<copyright-year>2021</copyright-year>
<copyright-holder>Graham, Cai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="article" id="rel-obj003" link-type="peer-reviewed-article"/>
<custom-meta-group>
<custom-meta>
<meta-name>Submission Version</meta-name>
<meta-value>1</meta-value>
</custom-meta>
</custom-meta-group>
</front-stub>
<body>
<p>
<named-content content-type="letter-date">16 May 2021</named-content>
</p>
<p>Dear Dr. Nishida,</p>
<p>Thank you very much for submitting your manuscript "Behavioral correlates of cortical semantic representations modeled by word vectors" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.</p>
<p>We are very glad to have received your revision. Reviewers are generally satisfied with the updated version. There are still a few comments from Reviewer #2 that need addressing. It would be great if you can improve the manuscript based on these sugestions.</p>
<p>Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.</p>
<p>When you are ready to resubmit, please upload the following:</p>
<p>[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out</p>
<p>[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).</p>
<p>Important additional instructions are given below your reviewer comments.</p>
<p>Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.</p>
<p>Sincerely,</p>
<p>Ming Bo Cai</p>
<p>Associate Editor</p>
<p>PLOS Computational Biology</p>
<p>Lyle Graham</p>
<p>Deputy Editor</p>
<p>PLOS Computational Biology</p>
<p>***********************</p>
<p>A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact <email xlink:type="simple">ploscompbiol@plos.org</email> immediately:</p>
<p>[LINK]</p>
<p>We are very glad to have received your revision. Reviewers are generally satisfied with the updated version. There are still a few comments from Reviewer #2 that need addressing. It would be great if you can improve the manuscript based on these sugestions.</p>
<p>Reviewer's Responses to Questions</p>
<p><bold>Comments to the Authors:</bold></p>
<p><bold>Please note here if the review is uploaded as an attachment.</bold></p>
<p>Reviewer #1: I now had a chance to carefully read the new version and the responses, and I think the authors did a really nice job on the revisions. I would be happy to see this paper published.</p>
<p>Reviewer #2: I thank the authors for considering my earlier comments. I am happy with the current version of the manuscript. Below are some minor points that I think would further improve the paper (especially point 1).</p>
<p>Minor:</p>
<p>1. In response to one of my earlier comments, the authors derived a random control model (“we emulated an untrained embedding network […] using word vectors for which the vector representations were shuffled across vector dimensions for each word.). This approach seems overly lenient, as no sensible betas can be derived for the encoding model if the dimensions are shuffled randomly for each datapoint. I therefore do not see the immediate use in this control. What I had originally thought of, but perhaps not communicated precisely enough, was that the authors take an untrained DNN (rather than a semantics trained DNN) and run it through their pipeline. This ensures that the encoding model can make “sensible” predictions, while the model activation patterns themselves are not driven based on semantics, as the model is untrained.</p>
<p>2. Typo: Author summery —&gt; Author summary</p>
<p>3. Figure 7 - would it make sense for the authors to add the binary category control model as an additional bar?</p>
<p>4. I would advise not calling binary noun/category descriptions “primitive” (l 678).</p>
<p>Reviewer #3: The authors have addressed all of my points in the revision. I have no more concerns left and would be happy to see the manuscript published in its current form.</p>
<p>**********</p>
<p><bold>Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?</bold></p>
<p>The <ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/materials-and-software-sharing" xlink:type="simple">PLOS Data policy</ext-link> requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.</p>
<p>Reviewer #1: Yes</p>
<p>Reviewer #2: Yes</p>
<p>Reviewer #3: Yes</p>
<p>**********</p>
<p>PLOS authors have the option to publish the peer review history of their article (<ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/editorial-and-peer-review-process#loc-peer-review-history" xlink:type="simple">what does this mean?</ext-link>). If published, this will include your full peer review and any attached files.</p>
<p>If you choose “no”, your identity will remain anonymous but your review may still be made public.</p>
<p><bold>Do you want your identity to be public for this peer review?</bold> For information about this choice, including consent withdrawal, please see our <ext-link ext-link-type="uri" xlink:href="https://www.plos.org/privacy-policy" xlink:type="simple">Privacy Policy</ext-link>.</p>
<p>Reviewer #1: No</p>
<p>Reviewer #2: <bold>Yes: </bold>Tim C Kietzmann</p>
<p>Reviewer #3: No</p>
<p>Figure Files:</p>
<p>While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, <ext-link ext-link-type="uri" xlink:href="https://pacev2.apexcovantage.com" xlink:type="simple">https://pacev2.apexcovantage.com</ext-link>. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at <email xlink:type="simple">figures@plos.org</email>.</p>
<p>Data Requirements:</p>
<p>Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: <ext-link ext-link-type="uri" xlink:href="http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5" xlink:type="simple">http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5</ext-link>.</p>
<p>Reproducibility:</p>
<p>To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at <ext-link ext-link-type="uri" xlink:href="https://plos.org/protocols?utm_medium=editorial-email&amp;utm_source=authorletters&amp;utm_campaign=protocols" xlink:type="simple">https://plos.org/protocols?utm_medium=editorial-email&amp;utm_source=authorletters&amp;utm_campaign=protocols</ext-link></p>
<p>References:</p>
<p>Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.</p>
<p><italic>If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.</italic></p>
</body>
</sub-article>
<sub-article article-type="author-comment" id="pcbi.1009138.r004">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r004</article-id>
<title-group>
<article-title>Author response to Decision Letter 1</article-title>
</title-group>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="peer-reviewed-article" id="rel-obj004" link-type="rebutted-decision-letter" object-id="10.1371/journal.pcbi.1009138.r003" object-id-type="doi" object-type="decision-letter"/>
<custom-meta-group>
<custom-meta>
<meta-name>Submission Version</meta-name>
<meta-value>2</meta-value>
</custom-meta>
</custom-meta-group>
</front-stub>
<body>
<p>
<named-content content-type="author-response-date">24 May 2021</named-content>
</p>
<supplementary-material id="pcbi.1009138.s047" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" position="float" xlink:href="info:doi/10.1371/journal.pcbi.1009138.s047" xlink:type="simple">
<label>Attachment</label>
<caption>
<p>Submitted filename: <named-content content-type="submitted-filename">Nishida-BehavCorrSemantic-ResponseLetter04.docx</named-content></p>
</caption>
</supplementary-material>
</body>
</sub-article>
<sub-article article-type="aggregated-review-documents" id="pcbi.1009138.r005" specific-use="decision-letter">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r005</article-id>
<title-group>
<article-title>Decision Letter 2</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name name-style="western">
<surname>Graham</surname>
<given-names>Lyle J.</given-names>
</name>
<role>Deputy Editor</role>
</contrib>
<contrib contrib-type="author">
<name name-style="western">
<surname>Cai</surname>
<given-names>Ming Bo</given-names>
</name>
<role>Associate Editor</role>
</contrib>
</contrib-group>
<permissions>
<copyright-year>2021</copyright-year>
<copyright-holder>Graham, Cai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="article" id="rel-obj005" link-type="peer-reviewed-article"/>
<custom-meta-group>
<custom-meta>
<meta-name>Submission Version</meta-name>
<meta-value>2</meta-value>
</custom-meta>
</custom-meta-group>
</front-stub>
<body>
<p>
<named-content content-type="letter-date">1 Jun 2021</named-content>
</p>
<p>Dear Dr. Nishida,</p>
<p>We are pleased to inform you that your manuscript 'Behavioral correlates of cortical semantic representations modeled by word vectors' has been provisionally accepted for publication in PLOS Computational Biology.</p>
<p>Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.</p>
<p>Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.</p>
<p>IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.</p>
<p>Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.</p>
<p>Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. </p>
<p>Best regards,</p>
<p>Ming Bo Cai</p>
<p>Associate Editor</p>
<p>PLOS Computational Biology</p>
<p>Lyle Graham</p>
<p>Deputy Editor</p>
<p>PLOS Computational Biology</p>
<p>***********************************************************</p>
<p>Reviewer's Responses to Questions</p>
<p><bold>Comments to the Authors:</bold></p>
<p><bold>Please note here if the review is uploaded as an attachment.</bold></p>
<p>Reviewer #2: The authors have addressed my remaining concerns. I suggest acceptance.</p>
<p>**********</p>
<p><bold>Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?</bold></p>
<p>The <ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/materials-and-software-sharing" xlink:type="simple">PLOS Data policy</ext-link> requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.</p>
<p>Reviewer #2: None</p>
<p>**********</p>
<p>PLOS authors have the option to publish the peer review history of their article (<ext-link ext-link-type="uri" xlink:href="https://journals.plos.org/ploscompbiol/s/editorial-and-peer-review-process#loc-peer-review-history" xlink:type="simple">what does this mean?</ext-link>). If published, this will include your full peer review and any attached files.</p>
<p>If you choose “no”, your identity will remain anonymous but your review may still be made public.</p>
<p><bold>Do you want your identity to be public for this peer review?</bold> For information about this choice, including consent withdrawal, please see our <ext-link ext-link-type="uri" xlink:href="https://www.plos.org/privacy-policy" xlink:type="simple">Privacy Policy</ext-link>.</p>
<p>Reviewer #2: No</p>
</body>
</sub-article>
<sub-article article-type="editor-report" id="pcbi.1009138.r006" specific-use="acceptance-letter">
<front-stub>
<article-id pub-id-type="doi">10.1371/journal.pcbi.1009138.r006</article-id>
<title-group>
<article-title>Acceptance letter</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name name-style="western">
<surname>Graham</surname>
<given-names>Lyle J.</given-names>
</name>
<role>Deputy Editor</role>
</contrib>
<contrib contrib-type="author">
<name name-style="western">
<surname>Cai</surname>
<given-names>Ming Bo</given-names>
</name>
<role>Associate Editor</role>
</contrib>
</contrib-group>
<permissions>
<copyright-year>2021</copyright-year>
<copyright-holder>Graham, Cai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<related-object document-id="10.1371/journal.pcbi.1009138" document-id-type="doi" document-type="article" id="rel-obj006" link-type="peer-reviewed-article"/>
</front-stub>
<body>
<p>
<named-content content-type="letter-date">18 Jun 2021</named-content>
</p>
<p>PCOMPBIOL-D-20-01475R2 </p>
<p>Behavioral correlates of cortical semantic representations modeled by word vectors</p>
<p>Dear Dr Nishida,</p>
<p>I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.</p>
<p>The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript. </p>
<p>Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.</p>
<p>Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work! </p>
<p>With kind regards,</p>
<p>Katalin Szabo</p>
<p>PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom <email xlink:type="simple">ploscompbiol@plos.org</email> | Phone +44 (0) 1223-442824 | <ext-link ext-link-type="uri" xlink:href="http://ploscompbiol.org" xlink:type="simple">ploscompbiol.org</ext-link> | @PLOSCompBiol</p>
</body>
</sub-article>
</article>