parse trees
Recently Published Documents


TOTAL DOCUMENTS

122
(FIVE YEARS 18)

H-INDEX

14
(FIVE YEARS 2)

Author(s):  
Peter beim Graben ◽  
Markus Huber ◽  
Werner Meyer ◽  
Ronald Römer ◽  
Matthias Wolff

AbstractVector symbolic architectures (VSA) are a viable approach for the hyperdimensional representation of symbolic data, such as documents, syntactic structures, or semantic frames. We present a rigorous mathematical framework for the representation of phrase structure trees and parse trees of context-free grammars (CFG) in Fock space, i.e. infinite-dimensional Hilbert space as being used in quantum field theory. We define a novel normal form for CFG by means of term algebras. Using a recently developed software toolbox, called FockBox, we construct Fock space representations for the trees built up by a CFG left-corner (LC) parser. We prove a universal representation theorem for CFG term algebras in Fock space and illustrate our findings through a low-dimensional principal component projection of the LC parser state. Our approach could leverage the development of VSA for explainable artificial intelligence (XAI) by means of hyperdimensional deep neural computation.


2021 ◽  
Vol 5 (OOPSLA) ◽  
pp. 1-24
Author(s):  
Xiaodong Jia ◽  
Ashish Kumar ◽  
Gang Tan

In this paper, we present a derivative-based, functional recognizer and parser generator for visibly pushdown grammars. The generated parser accepts ambiguous grammars and produces a parse forest containing all valid parse trees for an input string in linear time. Each parse tree in the forest can then be extracted also in linear time. Besides the parser generator, to allow more flexible forms of the visibly pushdown grammars, we also present a translator that converts a tagged CFG to a visibly pushdown grammar in a sound way, and the parse trees of the tagged CFG are further produced by running the semantic actions embedded in the parse trees of the translated visibly pushdown grammar. The performance of the parser is compared with a popular parsing tool ANTLR and other popular hand-crafted parsers. The correctness of the core parsing algorithm is formally verified in the proof assistant Coq.


Author(s):  
May Kyi Nyein ◽  
Khin Mar Soe

Word reordering has remained one of the challenging problems for machine translation when translating between language pairs with different word orders e.g. English and Myanmar. Without reordering between these languages, a source sentence may be translated directly with similar word order and translation can not be meaningful. Myanmar is a subject-objectverb (SOV) language and an effective reordering is essential for translation. In this paper, we applied a pre-ordering approach using recurrent neural networks to pre-order words of the source Myanmar sentence into target English’s word order. This neural pre-ordering model is automatically derived from parallel word-aligned data with syntactic and lexical features based on dependency parse trees of the source sentences. This can generate arbitrary permutations that may be non-local on the sentence and can be combined into English-Myanmar machine translation. We exploited the model to reorder English sentences into Myanmar-like word order as a preprocessing stage for machine translation, obtaining improvements quality comparable to baseline rule-based pre-ordering approach on asian language treebank (ALT) corpus.


2021 ◽  
Author(s):  
Atul Sahay ◽  
Ayush Maheshwari ◽  
Ritesh Kumar ◽  
Ganesh Ramakrishnan ◽  
Manjesh Kumar Hanawal ◽  
...  

Author(s):  
Nasrin Taghizadeh ◽  
Heshaam Faili

We describe a cross-lingual adaptation method based on syntactic parse trees obtained from the Universal Dependencies (UD), which are consistent across languages, to develop classifiers in low-resource languages. The idea of UD parsing is to capture similarities as well as idiosyncrasies among typologically different languages. In this article, we show that models trained using UD parse trees for complex NLP tasks can characterize very different languages. We study two tasks of paraphrase identification and relation extraction as case studies. Based on UD parse trees, we develop several models using tree kernels and show that these models trained on the English dataset can correctly classify data of other languages, e.g., French, Farsi, and Arabic. The proposed approach opens up avenues for exploiting UD parsing in solving similar cross-lingual tasks, which is very useful for languages for which no labeled data is available.


2021 ◽  
pp. 395-403
Author(s):  
Yuri Engelhardt ◽  
Clive Richards

AbstractA ‘universal grammar’ for the full spectrum of visualization types is discussed. The grammar enables the analysis of any type of visualization regarding its syntactic constituents, such as the types of visual encodings and visual components that are used. Such an analysis of a type of visualization, describing its compositional syntax, can be represented as a specification tree. Colour coded tree branches between constituent types enforce the combination rules visually. We discuss how these specification trees differ from linguistic parse trees, and how visual statements differ from verbal statements. The grammar offers a basis for generating visualization options, and the potential for formalization and for machine-readable specifications. This may serve as a basis for a system providing computer-generated visualization advice.


2021 ◽  
Vol 102 ◽  
pp. 01011
Author(s):  
Duc Tran Vu ◽  
John Blake

In this paper, we describe the design and development of the first release of an online question generator. This pedagogic tool enables learners of English to generate open-ended, closed-ended and tag questions for a target sentence. Learners input a sentence (i.e. declarative statement) and select the type or types of questions to generate. Question generation is a non-trivial task involving numerous processes including syntactic transformation and pronoun selection. Syntactic transformation was achieved through the use of rules based on parse trees while the selection of interrogative pronouns was achieved using matching potential question foci with a linguistic knowledge encoded condition set. Lessons learned are detailed to help other researchers avoid or attempt to ameliorate the pitfalls and problems encountered in this study.


2020 ◽  
Vol 12 (12) ◽  
pp. 218
Author(s):  
Dario Onorati ◽  
Pierfrancesco Tommasino ◽  
Leonardo Ranaldi ◽  
Francesca Fallucchi ◽  
Fabio Massimo Zanzotto

The dazzling success of neural networks over natural language processing systems is imposing an urgent need to control their behavior with simpler, more direct declarative rules. In this paper, we propose Pat-in-the-Loop as a model to control a specific class of syntax-oriented neural networks by adding declarative rules. In Pat-in-the-Loop, distributed tree encoders allow to exploit parse trees in neural networks, heat parse trees visualize activation of parse trees, and parse subtrees are used as declarative rules in the neural network. Hence, Pat-in-the-Loop is a model to include human control in specific natural language processing (NLP)-neural network (NN) systems that exploit syntactic information, which we will generically call Pat. A pilot study on question classification showed that declarative rules representing human knowledge, injected by Pat, can be effectively used in these neural networks to ensure correctness, relevance, and cost-effective.


2020 ◽  
Vol 29 (1) ◽  
pp. 43-57 ◽  
Author(s):  
Matia Vannoni ◽  
Elliott Ash ◽  
Massimo Morelli

Bureaucratic discretion and executive delegation are central topics in political economy and political science. The previous empirical literature has measured discretion and delegation by manually coding large bodies of legislation. Drawing from computational linguistics, we provide an automated procedure for measuring discretion and delegation in legal texts to facilitate large-scale empirical analysis. The method uses information in syntactic parse trees to identify legally relevant provisions, as well as agents and delegated actions. We undertake two applications. First, we produce a measure of bureaucratic discretion by looking at the level of legislative detail for US states and find that this measure increases after reforms giving agencies more independence. This effect is consistent with an agency cost model, where a more independent bureaucracy requires more specific instructions (less discretion) to avoid bureaucratic drift. Second, we construct measures of delegation to governors in state legislation. Consistent with previous estimates using non-text metrics, we find that executive delegation increases under unified government.


Sign in / Sign up

Export Citation Format

Share Document