Thanks to the speakers, participants, sponsors, & organizers that made this 3rd edition of Compℹ︎La -at CCS2024 at Exeter University- possible❗
Alvaro Corral
The brevity law can be precisely formulated in terms of a scaling law: results for different languages.
Abstract
Recently, it has been shown that the well-known negative correlation between word frequency and word length (the so-called “brevity law”) can be more clearly quantified by means of a scaling law. This new law states that the distribution of word frequencies conditioned to fixed word length displays a scaling form in which the frequency n scales as a power of the length l, as ld. The negative value of the exponent d determines the brevity effect. However, previous research showed only “anecdotical” evidence, as the scaling law was only established for the English language and word length was defined in terms of number of characters. Here we expand those results to include other languages and other (more linguistically motivated) definitions of word length. We pay special attention to the possible (strong or weak) universality of the results. Also, these findings allow for a new interpretation of the classical Zipf’s law for word frequencies.
Christine Cuskley
Rethinking variation as a core property of adaptive language systems
Abstract
Frequency plays a key role in language dynamics in domains spanning learning, memory, and communication. For example, frequency correlates with morphological irregularity (high frequency words are more likely to be irregularly inflected) and high frequency units are also more likely to be regularised by learners (i.e., exhibit less intra- and inter-speaker variation). In this paper, I present an agent-based model which looks at the tension between these two patterns. Intuition might suggest these dynamics would lead to systems which gradually eliminate variation, but this is clearly not the pattern we find in natural languages (e.g., Cuskley et al., 2014; Cuskley et al., 2017). Using a variant of the regularity game (Cuskley et al., 2018), this model reframes questions of variation in terms of how language systems maintain frequency dependent variation, exploring the relative contributions of maturation (i.e., that child learners have different biases than adult learners), memory, and socio-communicative pressures. In particular, the tension between learning/memory and communication supports dynamics similar to stabilizing selection in biology: for highly frequent forms, the model shows frequency dependent selection against morphologically regular (i.e., type-dominant) rules where this would disrupt communicative success in populations. The work concludes with an echo of calls to reorient thinking about language variation: in many more traditional approaches to linguistics, the draw of universalist explanations still tacitly frames variation as a phenomena which must be explained away in order to reveal "core" properties of human language. Instead, I argue we should refocus on variation itself as a core adaptive feature of language, shifting to precise questions regarding how language maintains variation as fodder for selection, and thus, adaptation.
Conor Houghton, Seth Bullock, and Jack Bunyan
A new iterated learning model of language evolution
Abstract
Languages tend to be expressive and compositional. The emergence of these properties is modelled in the iterated learning model. In the model a tutor teaches a pupil their language, in this case a mapping from binary vector `meanings' to `signals', by exposing them to a set of meaning-signal pairs. Subsequently, the pupil becomes the tutor for a new naive pupil. Successive cycles of learning and generalization lead to a stable, expressive, compositional language. However, while an agent's map from signals to meanings is implemented by their neural network, in the original iterated language model the complementary map from meanings to signals is calculated through `obversion' which becomes impractical for larger languages and is also unrealistic, requiring consideration of all possible meaning-signal pairs. We propose a more computationally tractable and more realistic model in which both the encoder and decoder are neural networks, concatenated together as an autoencoder, and pupils are required to learn from a mix of unsupervised and supervised examples, as children do. This suggests that the `internalization' of utterances may be a crucial step in the development of language.
Anna Jon-And
Exploring system-internal dynamics in the evolution of grammar
Abstract
The study of general mechanisms of language change tends to focus on either universal cultural selective pressures, such as learnability, expressivity or ease of production/perception, or on extralinguistic conditions such as speaker population size or proportions of second language speakers. However, factors intrinsic to the linguistic system in place may also play an important role in constraining or triggering change. Languages are built up of structural traits, forming the grammar of the language. Grammatical traits exhibit empirical dependencies, that have traditionally been considered universal. Recently the focus has shifted to scrutinizing lineage-specific characteristics, suggesting that language evolution is partially guided by local dynamics leading to distinct basins of attraction. Our aim here is to explore whether these basins can be identified through the examination of interactions between grammatical traits at a global level. We use Grambank, a new database of 195 traits for 2,467 language varieties, to create a network of linguistic traits, where the weight of the edge between two traits is determined by their likelihood of co-occurence in languages. The network is weighted to correct for common ancestry between languages, using scores of lexical proximity from A Global Lexical Database (GLED). Preliminary results from community detection in the network indicate attraction between similar patterns concerning bound or free morphology respectively and word order, that go beyond language families.
Adam Lipowski and Dorota Lipowska
Language ambiguities in signaling game - the stabilizing role of context
Maxi San Miguel
Prestige and volatility in Language Contact Dynamics
Abstract
Language contact refers to a situation of two languages which are spoken in a given society. The main question is then about dominance, extinction or coexistence of the two languages. Following the milestone paper by Abrams and Strogatz [1] two main parameters, prestige and volatility, appear in different models considering these questions. I will address the role of covert prestige in terms of different linguistic preferences for the two speech communities. I will also consider the role of heterogeneous volatility, either as Heterogeneous Language volatility or as Heterogeneous Social volatility. Given the many possible outcomes depending on parameter values, I will argue that we might need data about these parameters themselves, rather than fitting their values to reproduce language use data with a specific model.
[1] D. M. Abrams and S. H. Strogatz, “Modelling the dynamics of language death”, Nature 424, 900–900 (2003).
Tsuyoshi Mizuguchi, T. Yamamoto, S. Yamada and T. Suzuki
Return map analysis of word sequences
Abstract
The structure of written texts is analyzed by focusing on word sequences. As a method, word sequences in texts are transformed into rank sequences of the occurrence frequency of each word. The features of text are extracted by comparing the return map of its rank sequence with its shuffled data. Written texts of several different languages are analyzed. The results show that there is a negative correlation in the rank of adjacent words in many languages, and features of return maps of the same language texts are similar. A clustering structure which implies the relation to language (sub)family is observed. The sequences of word length are also analyzed.
Javier Martin Arista:
Historical language generation. Structural evaluation of prosified Old English poetry
Abstract
The aim of this paper is to propose and assess metrics that can be used for the structural evaluation of text of a historical language generated by AI. More specifically, the paper focuses on the generation of Old English, of which the total of written records is slightly over 3 million words (Taylor et al. 2003; Healey et al. 2004) and clearly insuJicient for AI models. The procedure of text generation is prompting with an AI chat (Petroni et al. 2021) , which is inputted a poetic text of Old English and is asked to produce a prose version of the poetry. Zero-shot learning and few-shot learning (with examples of prosification of other texts) are carried out (Gao et al. 2021). The output is assessed as to quantity and as to quality. On the side of quantity, the ratio of words and the ratio of unique words of the new version are calculated. As for quality, the meaning and the structure of the text are taken into account. Meaning, including lexical attestedness and collocation, is given more weight (60%) than structure, which includes inflection and word order and has a 40% weight. An Overall Acceptability Score is put forward that constitutes a weighted average of four diJerent scores: Attestedness Score, Inflection Score, Collocation Score and Word Order Score. Scaling factors adjust for the fact that the diJerent scores make reference to diJerent units (words, pairs and sentences). The resulting Overall Acceptability Score is checked against an analysis of the constructions in the generated prose. Conclusions are expected regarding the weights of the scores.
References
Gao, Tianyu, Adam Fischz & Danqi Cheny. 2021. Making Pre-trained Language Models Better Few-shot Learners. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, pages 3816–3830.
Healey, A. (ed.), J. Price-Wilkin & X. Xiang. 2004. The Dictionary of Old English web corpus. Toronto: Dictionary of Old English Project, Centre for Medieval Studies, University of Toronto.
Petroni, F., T. Rocktäschel, P. Lewis, A. Bakhtin, Y. Wu, A. H. Miller & S. Riedel. 2019. Language Models as Knowledge Bases? Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pages 2463–2473.
Taylor, A., A. Warner, S. Pintzuk & F. Beths 2003. The York-Toronto-Helsinki Parsed Corpus of Old English Prose [https://www-users.york.ac.uk/~lang22/YcoeHome1.htm].
Gabriel Cachoa-Ocampo, Bibiana Obregón-Quintana
Modeling Language Survivable: The Case of Indigenous Languages in Mexico
Jérôme Michaud
Combinatorial evolution and self-organization of linguistic systems
Pablo Rosillo-Rodes, Maxi San Miguel, and David Sánchez
Lexical analysis of a Flamenco lyrics corpus
Erich Round, Louise Esher, Sacha Beniamine
Modelling the evolution of inflection systems