Abstract
In order to describe how humans represent meaning in the brain, one must be able to account for not just concrete words but, critically, also abstract words which lack a physical referent. Hebbian formalism and optimization are basic principles of brain function, and they provide an appealing approach for modeling word meanings based on word co-occurrences. Here, we built a model of the semantic space based on word statistics in a large text corpus, which was able to decode items from brain signals. In the model, word abstractness emerged from the statistical regularities of the language environment. This salient property of the model co-varied, at 280–420 ms after word presentation, with activity in the left-hemisphere frontal, anterior temporal and superior parietal cortex that have been linked with processing of abstract words. In light of these results, we propose that the neural encoding of word meanings is importantly grounded in language through statistical regularities.