Journal of Language Modelling
Latest Publications


TOTAL DOCUMENTS

102
(FIVE YEARS 32)

H-INDEX

5
(FIVE YEARS 2)

Published By "Institute Of Computer Science, Polish Academy Of Sciences"

2299-8470, 2299-856x

2021 ◽  
Vol 9 (2) ◽  
pp. 195-223
Author(s):  
Carmen Klaussner ◽  
Carl Vogel ◽  
Arnab Bhattacharya

This work offers an investigation into linguistic changes in a corpus of literary authors hypothesised to be possibly attributable to the effects of ageing. In part, the analysis replicates an earlier study into these effects, but adds to it by explicitly analysing and modelling competing factors, specifically the influence of background language change. Our results suggest that it is likely that this underlying change in language usage is the primary force for the change observed in the linguistic variables that was previously attributed to linguistic ageing.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Tamar Johnson ◽  
Kexin Gao ◽  
Kenny Smith ◽  
Hugh Rabagliati ◽  
Jennifer Culbertson

Research on cross-linguistic differences in morphological paradigms reveals a wide range of variation on many dimensions, including the number of categories expressed, the number of unique forms, and the number of inflectional classes. However, in an influential paper, Ackerman & Malouf (2013) argue that there is one dimension on which languages do not differ widely: in predictive structure. Predictive structure in a paradigm describes the extent to which forms predict each other, called i-complexity. Ackerman & Malouf (2013) show that although languages differ according to measure of surface paradigm complexity, called e-complexity, they tend to have low i-complexity. They conclude that morphological paradigms have evolved under a pressure for low i-complexity, such that even paradigms with very high e-complexity are relatively easy to learn so long as they have low i-complexity. While this would potentially explain why languages are able to maintain large paradigms, recent work by Johnson et al. (submitted) suggests that both neural networks and human learners may actually be more sensitive to e-complexity than i-complexity. Here we will build on this work, reporting a series of experiments under more realistic learning conditions which confirm that indeed, across a range of paradigms that vary in either e- or i-complexity, neural networks (LSTMs) are sensitive to both, but show a larger effect of e-complexity (and other measures associated with size and diversity of forms). In human learners, we fail to find any effect of i-complexity at all. Further, analysis of a large number of randomly generated paradigms show that e- and i-complexity are negatively correlated: paradigms with high e-complexity necessarily show low i-complexity.These findings suggest that the observations made by Ackerman & Malouf (2013) for natural language paradigms may stem from the nature of these measures rather than learning pressures specially attuned to i-complexity.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Ezer Rasin ◽  
Iddo Berger ◽  
Nur Lan ◽  
Itamar Shefi ◽  
Roni Katzir

A linguistic theory reaches explanatory adequacy if it arrives at a linguistically-appropriate grammar based on the kind of input available to children. In phonology, we assume that children can succeed even when the input consists of surface evidence alone, with no corrections or explicit paradigmatic information – that is, in learning from distributional evidence. We take the grammar to include both a lexicon of underlying representations and a mapping from the lexicon to surface forms. Moreover, this mapping should be able to express optionality and opacity, among other textbook patterns. This learning challenge has not yet been addressed in the literature. We argue that the principle of Minimum Description Length (MDL) offers the right kind of guidance to the learner – favoring generalizations that are neither overly general nor overly specific – and can help the learner overcome the learning challenge. We illustrate with an implemented MDL learner that succeeds in learning various linguistically-relevant patterns from small corpora.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Dakotah Jay Lambert ◽  
Jonathan Rawski ◽  
Jeffrey Heinz

We derive well-understood and well-studied subregular classes of formal languages purely from the computational perspective of algorithmic learning problems. We parameterise the learning problem along dimensions of representation and inference strategy. Of special interest are those classes of languages whose learning algorithms are necessarily not prohibitively expensive in space and time, since learners are often exposed to adverse conditions and sparse data. Learned natural language patterns are expected to be most like the patterns in these classes, an expectation supported by previous typological and linguistic research in phonology. A second result is that the learning algorithms presented here are completely agnostic to choice of linguistic representation. In the case of the subregular classes, the results fall out from traditional model-theoretic treatments of words and strings. The same learning algorithms, however, can be applied to model-theoretic treatments of other linguistic representations such as syntactic trees or autosegmental graphs, which opens a useful direction for future research.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Brandon Prickett

A number of experiments have demonstrated what seems to be a bias in human phonological learning for patterns that are simpler according to Formal Language Theory (Finley and Badecker 2008; Lai 2015; Avcu 2018). This paper demonstrates that a sequence-to-sequence neural network (Sutskever et al. 2014), which has no such restriction explicitly built into its architecture, can successfully capture this bias. These results suggest that a bias for patterns that are simpler according to Formal Language Theory may not need to be explicitly incorporated into models of phonological learning.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Ezer Rasin ◽  
Roni Katzir ◽  
Tim O'Donnell

2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Noam Chomsky

The goal of theory construction is explanation: for language, theory for particular languages (grammar) and for the faculty of language FoL (the innate endowment for language acquisition). A primitive notion of simplicity of grammars is number of symbols, but this is too crude. An improved measure distinguishes grammars that capture genuine properties of language from those that do not. The theory of FoL must meet the empirical conditions of learnability (under extreme poverty of stimulus), and evolvability (given the limited but not insignificant evidence available). Recent work provides promising insights into how these twin conditions may be satisfied.


2021 ◽  
Vol 8 (2) ◽  
Author(s):  
Ralf Naumann ◽  
Thomas Gamerschlag

Based on both syntactic and semantic criteria, Stewart (2001) and, following him, Baker and Stewart (1999), distinguish two types of serial verb constructions (SVC) and one type of covert coordination (CC) in Edo. In this article, we present an analysis of these constructions, using Type Logical Grammar (TLG) with an event-based semantic component. We choose as base logic the non-associative Lambek calculus augmented with two unary multiplicative connectives (NL(◊, □)). SVCs and CCs are interpreted as complex event structures. The complex predicates underlying these structures are derived from simple verbs by means of a constructor. SVCs and CCs differ in terms of which part of the complex event structure is denoted. For SVCs, this is the sum of all events in the structure whereas for a CC this is only the first event in the sequence. The two verbs in an SVC and a CC are treated asymmetrically by assuming that the first verb has an extended subcategorization frame. The additional argument is of type vp (possibly modally decorated). Constraints on word order and the realization of arguments are accounted for using structural rules like permutation and contraction. The application of these rules is enforced by making use of the unary connectives.


2021 ◽  
Vol 8 (2) ◽  
Author(s):  
Marie Candito ◽  
Mathieu Constant ◽  
Carlos Ramisch ◽  
Agata Savary ◽  
Bruno Guillaume ◽  
...  

We present the enrichment of a French treebank of various genres with a new annotation layer for multiword expressions (MWEs) and named entities (NEs).1 Our contribution with respect to previous work on NE and MWE annotation is the particular care taken to use formal criteria, organized into decision flowcharts, shedding some light on the interactions between NEs and MWEs. Moreover, in order to cope with the well-known difficulty to draw a clear-cut frontier between compositional expressions and MWEs, we chose to use sufficient criteria only. As a result, annotated MWEs satisfy a varying number of sufficient criteria, accounting for the scalar nature of the MWE status. In addition to the span of the elements, annotation includes the subcategory of NEs (e.g., person, location) and one matching sufficient criterion for non-verbal MWEs (e.g., lexical substitution). The 3,099 sentences of the treebank were double-annotated and adjudicated, and we paid attention to cross-type consistency and compatibility with thesyntactic layer. Overall inter-annotator agreement on non-verbal MWEs and NEs reached 71.1%. The released corpus contains 3,112 annotated NEs and 3,440 MWEs, and is distributed under an open license.


2021 ◽  
Vol 8 (2) ◽  
Author(s):  
Peter Dekker ◽  
Willem Zuidema

In this paper, we investigate how the prediction paradigm from machine learning and Natural Language Processing (NLP) can be put to use in computational historical linguistics. We propose word prediction as an intermediate task, where the forms of unseen words in some target language are predicted from the forms of the corresponding words in a source language. Word prediction allows us to develop algorithms for phylogenetic tree reconstruction, sound correspondence identification and cognate detection, in ways close to attested methods for linguistic reconstruction. We will discuss different factors, such as data representation and the choice of machine learning model, that have to be taken into account when applying prediction methods in historical linguistics. We present our own implementations and evaluate them on different tasks in historical linguistics.


Sign in / Sign up

Export Citation Format

Share Document