generative language
Recently Published Documents


TOTAL DOCUMENTS

33
(FIVE YEARS 15)

H-INDEX

8
(FIVE YEARS 1)

2021 ◽  
Author(s):  
Richard W. Shuai ◽  
Jeffrey A. Ruffolo ◽  
Jeffrey J. Gray

Successful development of monoclonal antibodies (mAbs) for therapeutic applications is hindered by developability issues such as low solubility, low thermal stability, high aggregation, and high immunogenicity. The discovery of more developable mAb candidates relies on high-quality antibody libraries for isolating candidates with desirable properties. We present Immunoglobulin Language Model (IgLM), a deep generative language model for generating synthetic libraries by re-designing variable-length spans of antibody sequences. IgLM formulates antibody design as an autoregressive sequence generation task based on text-infilling in natural language. We trained IgLM on approximately 558M antibody heavy- and light-chain variable sequences, conditioning on each sequence's chain type and species-of-origin. We demonstrate that IgLM can be applied to generate synthetic libraries that may accelerate the discovery of therapeutic antibody candidates


2021 ◽  
Author(s):  
Ellen Jiang ◽  
Edwin Toh ◽  
Alejandra Molina ◽  
Aaron Donsbach ◽  
Carrie J Cai ◽  
...  

2021 ◽  
pp. 1-14
Author(s):  
Ethan Porter ◽  
Yamil R. Velez

Abstract Although placebo conditions are ubiquitous in survey experiments, little evidence guides common practices for their use and selection. How should scholars choose and construct placebos? First, we review the role of placebos in published survey experiments, finding that placebos are used inconsistently. Then, drawing on the medical literature, we clarify the role that placebos play in accounting for nonspecific effects (NSEs), or the effects of ancillary features of experiments. We argue that, in the absence of precise knowledge of NSEs that placebos are adjusting for, researchers should average over a corpus of many placebos. We demonstrate this agnostic approach to placebo construction through the use of GPT-2, a generative language model trained on a database of over 1 million internet news pages. Using GPT-2, we devise 5,000 distinct placebos and administer two experiments (N = 2,975). Our results illustrate how researchers can minimize their role in placebo selection through automated processes. We conclude by offering tools for incorporating computer-generated placebo text vignettes into survey experiments and developing recommendations for best practice.


Author(s):  
Caitlyn McLoughlin

This chapter considers John Capgrave’s Life of St Katherine within a queer genealogical framework in order to contribute to a queer historical archive. Procreative and generative language and detail early in the narrative explicitly open the vita to a reading that considers the text itself as offspring in a genealogical line of reproduced texts. Thus the chapter also understands this textual procreation as representative of a particularly genderqueer temporality. Finally, this chapter offers an intersectional consideration of Katherine’s characterization in order to assess how different social privileges and subjugations effect and enable her veneration, complicating essentialist notions of gendered social position and female sanctity.


2021 ◽  
Author(s):  
Douglas Summers-Stay ◽  
Claire Bonial ◽  
Clare Voss

Author(s):  
Ariel Goldstein ◽  
Zaid Zada ◽  
Eliav Buchnik ◽  
Mariano Schain ◽  
Amy Price ◽  
...  

Departing from classical rule-based linguistic models, advances in deep learning have led to the development of a new family of self-supervised deep language models (DLMs). These models are trained using a simple self-supervised autoregressive objective, which aims to predict the next word in the context of preceding words in real-life corpora. After training, autoregressive DLMs are able to generate new 'context-aware' sentences with appropriate syntax and convincing semantics and pragmatics. Here we provide empirical evidence for the deep connection between autoregressive DLMs and the human language faculty using a 30-min spoken narrative and electrocorticographic (ECoG) recordings. Behaviorally, we demonstrate that humans have a remarkable capacity for word prediction in natural contexts, and that, given a sufficient context window, DLMs can attain human-level prediction performance. Next, we leverage DLM embeddings to demonstrate that many electrodes spontaneously predict the meaning of upcoming words, even hundreds of milliseconds before they are perceived. Finally, we demonstrate that contextual embeddings derived from autoregressive DLMs capture neural representations of the unique, context-specific meaning of words in the narrative. Our findings suggest that deep language models provide an important step toward creating a biologically feasible computational framework for generative language.


2020 ◽  
Vol 389 ◽  
pp. 93-107
Author(s):  
Jinmeng Wu ◽  
Tingting Mu ◽  
Jeyarajan Thiyagalingam ◽  
John Y. Goulermas

Sign in / Sign up

Export Citation Format

Share Document