• Keine Ergebnisse gefunden

Instead of evaluating the model on how well it predicts RTs for individual words or acoustic durations for individual words or segments, we can ask whether the model properly mirrors higher-level properties of the language system. More specifically, we can ask whether qualitative differences in lexical processing can be accounted for as resulting from differences in the distributional properties of cues and outcomes. If discriminative learning is on the right track, these differences should emerge straightforwardly from discriminative learning.

First consider the systematic differences in lexical processing that have been observed for English and Hebrew. For English, transposing letters (e.g., writingperhapsaspehraps) is relatively harmless.

Readers can reconstruct the intended meaning, and when reading for content, typos of this kind often escape notice. However, in Hebrew, letter transpositions severely disrupt reading (Frost, 2012).

To understand why, Baayen (2012) applied incremental NDL to English and Hebrew datasets of comparable size, using letter pairs as cues. Hebrew emerged as using a smaller set of bigram cues,

while using these cues more intensively, such that Hebrew bigram cues tended to have stronger weights to lexome outcomes compared to English. Interestingly, the disruption caused by transposing letters can be estimated by calculating the extent to which a word’s activation decreases when a pair of letters is transposed. Median disruption in Hebrew was almost 6.6 times that for English, which explains why Hebrew is more vulnerable to letter transpositions than English. Other systemic differences between Hebrew and English reading, such as the relative immunity of loanwords in Hebrew to letter transpositions, and the absence of facilitation from form priming in Hebrew, also follow straightforwardly from discriminative learning. Thus, discriminative learning helps clarify that indeed, as hypothesized by Frost (2012), the very different distributional properties of the Hebrew and English lexicons lie at the heart of the qualitative differences observed experiment-ally for lexical processing in these languages.

A second example of a system property that discriminative learning handles well is a double dissociation observed for English. Under impairment, English speakers with memory loss tend to have greater difficulties with irregular verbs, whereas speakers with phonological impairment tend to have more problems with regular verbs. Joanisse and Seidenberg (1999) used a connectionist network to model this dissociation, and observed that they had to selectively add noise to the semantic units of their model in order to get it to work properly. Interestingly, Joanisse and Seidenberg (1999) used one-hot encoding for their semantic representations, under the assumption that regular and irregular verbs do not differ systematically in meaning. However, irregular verbs have denser semantic neighborhoods than do regular verbs (Baayen and Moscoso del Prado Martín, 2005).

Heitmeier and Baayen (2021) therefore used LDL-EoL to model the double dissociation of regularity by impairment, as LDL makes it possible to work with corpus-based semantic vectors that reflect the differences in semantic density of regular and irregular verbs. Their simulation studies show that it is indeed the greater semantic density of English irregular verbs that renders them more fragile under semantic impairment.

A third example of using discriminative learning at the system level concerns second and third language acquisition. Chuang et al. (2020a) carried out a series of simulations with translation

equivalents from German, English, Mandarin, and Dutch, using both incremental LDL and LDL-EoL. As expected, the onset of L2 learning has a substantial influence on learning, as does the amount of L2 input. Another factor that emerged as important is the number of homophones in a language. Within the L1, homophones can be handled fairly well. However, when additional languages are learned, the homophones become more fragile and begin to suffer from intrusion, specifically if the translation equivalents of the other languages are not themselves homophonous.

The fragility of homophones fits well with the problems that L2 learners have with L2 words that have multiple senses that do not overlap with the senses of their L1 translation equivalents. Take English cut for example. One can cut with an axe, a knife, or scissors, whereas in Dutch and Mandarin the different cutting actions have to be expressed by three different verbs: hakken, knippen, andsnijdenin Dutch andkǎn, qiē, andjiǎnin Mandarin. L1 English speakers, when learning Dutch or Mandarin as L2, will thus have to learn to make more fine-grained decisions as to which verb to use as translation equivalent for Englishcut. From this perspective,cutactually has three meanings, and hence is a homophone. It straightforwardly follows that LDL predicts they are inevitably difficult for non-native speakers to learn.

5. General considerations

In computational linguistics, deep learning networks are the state-of-the-art. NDL/LDL contrasts with deep learning in several ways. First, there are no hidden layers. Second, the mathematics of multivariate multiple regression guarantees interpretational transparency. For the endstate of learning, convergence is to the global optimum. Third, LDL works well in sparse high dimensional spaces, whereas deep learning typically works with denser spaces of much lower dimensionality.

Since discriminative networks typically have many thousands of units, they implement ‘wide learning’.

Importantly, dimension reduction can lead to considerable reduction in LDL’s prediction accuracy.

For low-dimensional compressed spaces, deep learning is required, as mappings between compressed low-dimensional spaces typically are non-linear. In LDL, it is advisable to match the dimensionality of form vectors with that of the semantic vectors, for the mathematics to work best. As form matrices

using binary coding to specify the presence or absence of n-grams or n-phones are very sparse, networks mapping form onto meaning can be simplified by removing connections with weights very close to zero (see, e.g. Arnold et al., 2017; Milin et al., 2017b), with hardly any loss of accuracy.

Fourth, NDL and LDL hardly have any hyperparameters. For incremental learning, the learning rate is a free hyperparameter. For form vectors based on units such as triphones or trigrams, the dimensionality of the form space is determined by the number of different triphones in the data, and is not a free hyperparameter. And since the dimension of semantic vectors should match that of the form vectors, this is not a free hyperparameter either. Thus, how the model performs is determined by, firstly, the way form and meaning vectors are constructed (see Milin et al., 2017b, for discussion of the choice of representation for solving nonlinear classification with NDL), and secondly, by the distributional properties of the data. From this perspective, NDL/LDL provides a statistical tool that is driven almost exclusively by the data.

Breiman (2001) distinguished between two modeling cultures, the data modeling culture of statistics, and the algorithmic modeling culture of machine learning. The goal of algorithmic modeling is to obtain precise predictions, and if the model remains a black box or is theoretically uninterpretable (as is often the case for deep learning models as well as with recursive partitioning models) this is not an issue. The goal of data modeling, by contrast, is to obtain a statistical model that can generate the observed data. The resulting model provides the analyst with insight into what could be the mechanisms that give rise to the data. Although LDL can be seen as the simplest possible way of implementing machine learning for lexical learning, it is in spirit much closer to statistical data analysis. It is designed to assist the analyst to better understand the relations between sublexical distributional properties and text-derived distributional semantic properties.

Because LDL implementslinearmultivariate multiple regression, it inherits from the linear model the limitation that nonlinear functional relations (as addressed by, e.g., the generalized additive model, Wood, 2017) cannot be discovered. Here, there is room for considerable improvement.

An important property of the present formalization of Word and Paradigm Morphology is that morphological processing is not construed as involving parallel operations on symbols of form

and meaning. Figuring out the proper forms in production, or the proper meanings in compre-hension, is left to the network. Importantly, the network dynamically creates form and meaning representations, rather than retrieving them from some static lexical repository. Thus, NDL and LDL move away from classical compositional models in which stems and exponents have to be parsed out for comprehension or combined for production. However, the present approach is ana-lytical in the sense that semantic vectors for inflected words are obtained by summing the semantic vectors of pertinent lexomes.

It has been argued that phonemes, stems, and exponents have corresponding specialized neural areas in the brain (Bozic et al., 2010; Cibelli et al., 2015). Since NDL and LDL are statistical models of lexical processing, they make no claims about the topological organization of discrimination networks in neural tissue. However, it is mathematically possible to impose topological organization on form and semantic units in a 2-dimensional plane. As shown in Baayen et al. (2018) and Shafaei-Bajestan et al. (2020), phone-like and morph-like clusters can emerge in such 2-D maps. (As shown by Heitmeier and Baayen (2021), such maps make it possible to implement topologically localized lesions in simulation studies of aphasia.) Thus, NDL and LDL are not necessarily at odds with experimental evidence that is usually understood as supporting the neural reality of phonemes and morphemes. Instead of taking phonemes and morphemes to be the atomic units of a linguistic calculus, they can be understood as emergent properties of discriminative learning networks. Crucially, these low-level statistical networks provide much more precise predictions for the fine details of lexical processing than can be obtained with high-level abstract symbolic systems (for a discussion of NDL and statistical learning in language acquisition, see Baayen et al., 2016b).

However, since high-level symbolic descriptions are part of the cultural embedding of language, and have uncontestable pedagogical value, a full model of the mental lexicon will have to comprise not only low-level discriminative learning, but also high-level explicit symbolic learning that is grafted on top of low-level implicit learning.

Vectors and matrix multiplication

In what follows, we provide a brief conceptual introduction to vectors and matrices, using as example shopping lists and total expenses depending on supermarket. Consider Jack, who wants to buy two apples and three pairs, and Jill, who wants to buy four apples and two pairs. Their respective shopping lists can be represented as vectors, ordered sequences of numbers: (2,3) for Jack and (4,2)for Jill. We bring these two shopping list vectors together in a 2×2 matrixC.

C =

At the Edeka supermarket, apples coste2 and pears e3. At the Rewe supermarket, they coste3 ande2. We bundle this information in a second matrix,F.

F =

We can now calculate the total expenses at each of the supermarkets for both Jack and Jill. For Jack, buying 2 apples and 3 pears at Edeka will coste2×2 + 3×3 = 13. This combination of weighting numbers by price, and adding, is formalized by matrix multiplication.

S =CF =

The numbers going into the calculation of the expenses for Jack at Edeka are bolded: the row vector of Jack’s shopping list inCis paired with the column vector for Edeka inF, the elements of these vectors are multiplied pairwise, and then summed. The other three total expenses listed in S are

obtained in the same way. Jack is better off going to Rewe for his shopping, whereas Jill is better of going to Edeka.

When modeling comprehension with LDL, we knowC and we also know S, and we want to estimateF. WithF in hand, we can transform a ‘shopping list’ of triphones into a list of ‘semantic expenses’.

Links to digital materials

An R package for discriminative learning,WpmWithLdl, is available at http://www.sfs.uni-tuebingen.de/~hbaayen/publications/WpmWithLdl_1.0.tar.gz. An optimized implemen-tation of LDL in julia,JudiLing, is available athttps://megamindhenry.github.io/JudiLing.

jl/stable/.

Reference List

Arnold, D., Tomaschek, F., Lopez, F., Sering, T., and Baayen, R. H. (2017). Words from spontaneous conversational speech can be recognized with human-like accuracy by an error-driven learning algorithm that discriminates between meanings straight from smart acoustic features, bypassing the phoneme as recognition unit. PLOS ONE, 12(4):e0174623.

Baayen, R. H. (2012). Learning from the bible: computational modelling of the costs of letter transpositions and letter exchanges in reading classical hebrew and modern english. Lingue e linguaggio, 11(2):123–146.

Baayen, R. H., Chuang, Y.-Y., and Blevins, J. P. (2018). Inflectional morphology with linear mappings. The Mental Lexicon, 13(2):232–270.

Baayen, R. H., Chuang, Y.-Y., Shafaei-Bajestan, E., and Blevins, J. (2019). The discriminative lexicon: A unified computational model for the lexicon and lexical processing in comprehension

and production grounded not in (de)composition but in linear discriminative learning.

Complexity.

Baayen, R. H., Milin, P., Filipović Durdević, D., Hendrix, P., and Marelli, M. (2011). An amorphous model for morphological processing in visual comprehension based on naive discriminative learning. Psychological Review, 118:438–482.

Baayen, R. H., Milin, P., and Ramscar, M. (2016a). Frequency in lexical processing. Aphasiology, 30(11):1174–1220.

Baayen, R. H. and Moscoso del Prado Martín, F. (2005). Semantic density and past-tense formation in three Germanic languages. Language, 81:666–698.

Baayen, R. H., Shaoul, C., Willits, J., and Ramscar, M. (2016b). Comprehension without segmentation: A proof of concept with naive discriminative learning. Language, Cognition, and Neuroscience, 31(1):106–128.

Baayen, R. H. and Smolka, E. (2020). Modelling morphological priming in German with naive discriminative learning. Frontiers in Communication, section Language Sciences. preprint on PsyArXiv, doi:10.31234/osf.io/nj39v.

Balota, D., Cortese, M., Sergent-Marshall, S., Spieler, D., and Yap, M. (2004). Visual word recognition for single-syllable words. Journal of Experimental Psychology: General, 133:283–

316.

Blevins, J. P. (2016). Word and paradigm morphology. Oxford University Press.

Bozic, M., Tyler, L. K., Ives, D. T., Randall, B., and Marslen-Wilson, W. D. (2010). Bihemispheric foundations for human speech comprehension.Proceedings of the National Academy of Sciences, 107(40):17439–17444.

Breiman, L. (2001). Random forests. Machine Learning, 45:5–32.

Burnard, L. (1995). Users guide for the British National Corpus. British National Corpus consortium, Oxford university computing service.

Chuang, Y.-Y., Bell, M. J., Banke, I., and Baayen, R. H. (2020a). Bilingual and multilingual mental lexicon: a modeling study with linear discriminative learning. Language Learning, pages 1–73.

Chuang, Y.-Y., Loo, K., Blevins, J. P., and Baayen, R. H. (2020b). Estonian case inflection made simple. A case study in Word and Paradigm morphology with Linear Discriminative Learning. In Körtvélyessy, L. and Štekauer, P., editors,Advances in Morphology, pages 119–141. Cambridge University Press.

Chuang, Y.-Y., Vollmer, M.-L., Shafaei-Bajestan, E., Gahl, S., Hendrix, P., and Baayen, R. H.

(2020c). The processing of nonword form and meaning in production and comprehension:

A computational modeling approach using linear discriminative learning. Behavior Research Methods, pages 1–51.

Cibelli, E. S., Leonard, M. K., Johnson, K., and Chang, E. F. (2015). The influence of lexical statistics on temporal lobe cortical dynamics during spoken word listening. Brain and language, 147:66–75.

Dalal, N. and Triggs, B. (2005). Histograms of oriented gradients for human detection. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), volume 1, pages 886–893.

Danks, D. (2003). Equilibria of the Rescorla-Wagner model. Journal of Mathematical Psychology, 47(2):109–121.

Divjak, D., Milin, P., Ez-zizi, A., Józefowski, J., and Adam, C. (2020). What is learned from exposure: an error-driven approach to productivity in language. Language, Cognition and Neuroscience, pages 1–24.

Ellis, N. C. (2006). Language acquisition as rational contingency learning. Applied Linguistics, 27(1):1–24.

Ellis, N. C. (2013). Second language acquisition. The Routledge Handbook of Second Language Acquisition, page 193.

Frost, R. (2012). Towards a universal model of reading. Behavioral and Brain Sciences, page in press.

Grainger, J., Dufau, S., Montant, M., Ziegler, J. C., and Fagot, J. (2012). Orthographic processing in baboons (papio papio). Science, 336(6078):245–248.

Hannagan, T., Ziegler, J. C., Dufau, S., Fagot, J., and Grainger, J. (2014). Deep learning of orthographic representations in baboons. PLOS-one, 9:e84843.

Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., Prenger, R., Satheesh, S., Sengupta, S., Coates, A., et al. (2014). Deep speech: Scaling up end-to-end speech recognition.

arXiv preprint arXiv:1412.5567.

Heitmeier, M. and Baayen, R. H. (2021). Simulating phonological and semantic impairment of English tense inflection with Linear Discriminative Learning. The Mental Lexicon, in press.

PsyArXiv.

Joanisse, M. F. and Seidenberg, M. S. (1999). Impairments in verb morphology after brain injury:

a connectionist model. Proceedings of the National Academy of Sciences, 96:7592–7597.

Kapatsinski, V. (2018). Changing minds changing tools: From learning theory to language acquisition to language change. MIT Press.

Keuleers, E., Lacey, P., Rastle, K., and Brysbaert, M. (2012). The british lexicon project: Lexical decision data for 28,730 monosyllabic and disyllabic english words.Behavior Research Methods, 44(1):287–304.

Landauer, T. and Dumais, S. (1997). A solution to Plato’s problem: The latent semantic analysis theory of acquisition, induction and representation of knowledge.Psychological Review, 104(2):211–240.

Liberman, A. and Mattingly, I. (1985). The motor theory of speech perception revised. Cognition, 21:1–36.

Linke, M., Broeker, F., Ramscar, M., and Baayen, R. H. (2017). Are baboons learning

“orthographic” representations? probably not. PLOS-ONE, 12(8):e0183876.

Lund, K. and Burgess, C. (1996). Producing high-dimensional semantic spaces from lexical co-occurrence. Behaviour Research Methods, Instruments, and Computers, 28(2):203–208.

Luo, X., Chuang, Y. Y., and Baayen, R. H. (2021). Judiling: an implementation in Julia of Linear Discriminative Learning algorithms for language modeling.

Matthews, P. H. (1974). Morphology. An Introduction to the Theory of Word Structure. Cambridge University Press, Cambridge.

Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.

Milin, P., Divjak, D., and Baayen, R. H. (2017a). A learning perspective on individual differences in skilled reading: Exploring and exploiting orthographic and semantic discrimination cues.

Journal of Experimental Psychology: Learning, Memory, and Cognition.

Milin, P., Feldman, L. B., Ramscar, M., Hendrix, P., and Baayen, R. H. (2017b). Discrimination in lexical decision. PLOS-one, 12(2):e0171935.

Milin, P., Madabushi, H. T., Croucher, M., and Divjak, D. (2020). Keeping it simple:

Implementation and performance of the proto-principle of adaptation and learning in the language sciences. PsyArXiv.

Nixon, J. S. (2020). Of mice and men: Speech sound acquisition as discriminative learning from prediction error, not just statistical tracking. Cognition, 197:104081.

Perrone-Bertolotti, M., Kujala, J., Vidal, J. R., Hamame, C. M., Ossandon, T., Bertrand, O., Minotti, L., Kahane, P., Jerbi, K., and Lachaux, J.-P. (2012). How silent is silent reading?

intracerebral evidence for top-down activation of temporal voice areas during reading. Journal of Neuroscience, 32(49):17554–17562.

Plag, I., Homann, J., and Kunter, G. (2017). Homophony and morphology: The acoustics of word-final S in English. Journal of Linguistics, 53(1):181–216.

Ramscar, M., Dye, M., and Klein, J. (2013a). Children value informativity over logic in word learning. Psychological Science, 24(6):1017–1023.

Ramscar, M., Dye, M., and McCauley, S. M. (2013b). Error and expectation in language learning:

The curious absence of mouses in adult speech. Language, 89(4):760–793.

Ramscar, M., Hendrix, P., Shaoul, C., Milin, P., and Baayen, R. H. (2014). Nonlinear dynamics of lifelong learning: the myth of cognitive decline. Topics in Cognitive Science, 6:5–42.

Ramscar, M. and Yarlett, D. (2007). Linguistic self-correction in the absence of feedback: A new approach to the logical problem of language acquisition. Cognitive Science, 31(6):927–960.

Ramscar, M., Yarlett, D., Dye, M., Denny, K., and Thorpe, K. (2010). The effects of feature-label-order and their implications for symbolic learning. Cognitive Science, 34(6):909–957.

Rastle, K., Davis, M. H., and New, B. (2004). The broth in my brother’s brothel: Morpho-orthographic segmentation in visual word recognition.Psychonomic Bulletin & Review, 11:1090–

1098.

Rescorla, R. A. (1988). Pavlovian conditioning. It’s not what you think it is.American Psychologist, 43(3):151–160.

Rescorla, R. A. and Wagner, A. R. (1972). A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement. In Black, A. H. and Prokasy, W. F., editors, Classical conditioning II: Current research and theory, pages 64–99. Appleton Century Crofts, New York.

Sering, K., Stehwien, N., and Gao, Y. (2019). create_vtl_corpus: Synthesizing a speech corpus with vocaltractlab (version v1.0.0). Zenodo.http://doi.org/10.5281/zenodo.2548895.

Shafaei-Bajestan, E., Tari, M. M., and Baayen, R. H. (2020). LDL-AURIS: Error-driven learning in modeling spoken word recognition. arXiv.

Shaoul, C. and Westbury, C. (2010). Exploring lexical co-occurrence space using hidex. Behavior Research Methods, 42(2):393–413.

Tomaschek, F., Plag, I., Ernestus, M., and Baayen, R. H. (2019). Modeling the duration of word-final s in english with naive discriminative learning. Journal of Linguistics, 57(1):123–161.

Tucker, B. V., Brenner, D., Danielson, D. K., Kelley, M. C., Nenadić, F., and Sims, M. (2019a). The massive auditory lexical decision (mald) database.Behavior research methods, 51(3):1187–1204.

Tucker, B. V., Sims, M., and Baayen, R. H. (2019b). Opposing forces on acoustic duration.

PsyArXiv.

Widrow, B. and Hoff, M. E. (1960). Adaptive switching circuits.1960 WESCON Convention Record Part IV, pages 96–104.

Wood, S. N. (2017). Generalized Additive Models. Chapman & Hall/CRC, New York.