Latent language models have recently gained attention for their potential as an alternative to (or proxy for) explicit knowledge bases (kbs).
In this position paper, we examine this hypothesis, identify strengths and limitations of both latent language models and knowledge bases, and discuss the complementary nature of the two paradigms.
Lexicon-free speech recognition naturally deals with the problem of
out-of-vocabulary (OOV) words. In this paper, we show that character-based
language models (LM) can perform as well as word-based LM
In this paper, we propose a novel architecture called Composition Attention
Grammars (CAGs) that recursively compose subtrees into a single vector
representation with a composition function, and selec
There are concerns that the ability of language models (LMs) to generate high
quality synthetic text can be misused to launch spam, disinformation, or
propaganda. Therefore, the research community is
We address the general task of structured commonsense reasoning : given a natural language input, the goal is to generate a graph such as an event such as an event or a reasoning-graph.
To employ large language models (lms) for this task, existing approaches ``serialize''the output graph as a flat list of nodes and edges.
Language models (LMs) have proven surprisingly successful at capturing
factual knowledge by completing cloze-style fill-in-the-blank questions such as
"Punta Cana is located in _." However, while know
As language models (LMs) scale, they develop many novel behaviors, good and
bad, exacerbating the need to evaluate how they behave. Prior work creates
evaluations with crowdwork (which is time-consumi
Prompting language models (LMs) with training examples and task descriptions
has been seen as critical to recent successes in few-shot learning. In this
work, we show that finetuning LMs in the few-sh
Recently, neural language models (LMs) have demonstrated impressive abilities
in generating high-quality discourse. While many recent papers have analyzed
the syntactic aspects encoded in LMs, there h
Recently, retrieval-augmented language models have shown to improve over standard neural language models, by accessing information retrieved from a large datastore, in addition to their standard, parametric, next-word prediction.
To this end, we perform a careful analysis of the various dimensions over which knn-lm diverges from standard lms, and investigate these dimensions one by one.
Inspired by evidence that pretrained language models (LMs) encode commonsense
knowledge, recent work has applied LMs to automatically populate commonsense
knowledge graphs (CKGs). However, there is a
Large pretrained language models (lms) can not only perform remarkably well on a range of natural language processing (nlp) tasks but also start improving on reasoning tasks such as arithmetic induction, symbolic manipulation, and commonsense reasoning with increasing size of models.
However, it is still unclear what the underlying capabilities of these models are.
Relational knowledge bases (kbs) are established tools for world knowledgerepresentation in machines, but they usually sacrifice some data modeling flexibility for these advantages because they adhere to a manually engineered schema.
We propose a novel taxonomy for relational knowledgerepresentation in contextual lms based on the level of kb supervision provided, considering both works that probe lms for implicit relational knowledge acquired during self-supervised pretraining on unstructured text alone, and works that explicitly supervise lms at the level of kb entities and/or relationships.
Language models (LMs) are typically trained once on a large-scale corpus and
used for years without being updated. However, in a dynamic world, new entities
constantly arise. We propose a framework to
Recent years have witnessed a new paradigm of building natural language processing (NLP) systems: general-purpose, pre-trained language models (LMs) are fine-tuned with simple downstream models to att
Pre-trained language models (LMs), such as BERT (Devlin et al., 2018) and its
variants, have led to significant improvements on various NLP tasks in past
years. However, a theoretical framework for st
Symbolic knowledge graphs (KGs) have been constructed either by expensive
human crowdsourcing or with domain-specific complex information extraction
pipelines. The emerging large pretrained language m
We present a text-only approach to augment language models with non-differentiable tools, and an iterative"self-play"technique to bootstrap performance starting from few tooldemonstrations.
At a given model scale,
tool augmented language models significantly outperform non-augmented language models.
This survey reviews works in which language models (LMs) are augmented with
reasoning skills and the ability to use tools. The former is defined as
decomposing a potentially complex task into simpler
Pretrained neural language models (LMs) are prone to generating racist,
sexist, or otherwise toxic language which hinders their safe deployment. We
investigate the extent to which pretrained LMs can b