scholarly journals An Information-Theoretic Analysis of the Impact of Task Similarity on Meta-Learning

Author(s):  
Sharu Theresa Jose ◽  
Osvaldo Simeone
2021 ◽  
pp. 1-1
Author(s):  
Alexandros E. Tzikas ◽  
Panagiotis D. Diamantoulakis ◽  
George K. Karagiannidis

Entropy ◽  
2021 ◽  
Vol 23 (1) ◽  
pp. 126
Author(s):  
Sharu Theresa Jose ◽  
Osvaldo Simeone

Meta-learning, or “learning to learn”, refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that use either separate within-task training and test sets, like model agnostic meta-learning (MAML), or joint within-task training and test sets, like reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel individual task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Matteo Pellegrini

AbstractThis paper provides a fully word-based, abstractive analysis of predictability in Latin verb paradigms. After reviewing previous traditional and theoretically grounded accounts of Latin verb inflection, a procedure is outlined where the uncertainty in guessing the content of paradigm cells given knowledge of one or more inflected wordforms is measured by means of the information-theoretic notions of unary and n-ary implicative entropy, respectively, in a quantitative approach that uses the type frequency of alternation patterns between wordforms as an estimate of their probability of application. Entropy computations are performed by using the Qumin toolkit on data taken from the inflected lexicon LatInfLexi. Unary entropy values are used to draw a mapping of the verbal paradigm in zones of full interpredictability, composed of cells that can be inferred from one another with no uncertainty. N-ary entropy values are used to extract categorical and near principal part sets, that allow to fill the rest of the paradigm with little or no uncertainty. Lastly, the issue of the impact of information on the derivational relatedness of lexemes on uncertainty in inflectional predictions is tackled, showing that adding a classification of verbs in derivational families allows for a relevant reduction of entropy, not only for derived verbs, but also for simple ones.


2011 ◽  
Vol 4 ◽  
pp. 183-192 ◽  
Author(s):  
Khan Md. Mahfuzus Salam ◽  
Tetsuro Nishino ◽  
Kazutoshi Sasahara ◽  
Miki Takahasi ◽  
Kazuo Okanoya

Author(s):  
Subhashish Banerjee ◽  
Ashutosh Kumar Alok ◽  
R. Srikanth ◽  
Beatrix C. Hiesmayr

Sign in / Sign up

Export Citation Format

Share Document