We propose a novel post-processing approach, rethinking withretrieval (rr), which retrieves relevant external knowledge based on the decomposed reasoning steps obtained from the chain-of-thought (cot) prompting.
This lightweight approach does not require additional training or fine-tuning and is not limited by the input length of large language models (llms).
Recent literature has shown that large language models (LLMs) are generally
excellent few-shot reasoners to solve text reasoning tasks. However, the
capability of LLMs on table reasoning tasks is yet
Large Language Models (LLMs) have been transformative. They are pre-trained
foundational models that can be adapted with fine tuning to many different
natural language tasks, each of which previously
Recent advances in large language models (llm) now enable software developers to generate code based on a naturallanguage prompt.
Within computer science education, researchers are exploring the potential for llms to generate code explanations and programming assignments using carefully crafted prompts.
We are currently witnessing dramatic advances in the capabilities of Large
Language Models (LLMs). They are already being adopted in practice and
integrated into many systems, including integrated dev
Recent large language models (LLMs) have demonstrated remarkable performance
on a variety of natural language processing (NLP) tasks, leading to intense
excitement about their applicability across var
We propose smoothquant, a training-free, accuracy-preserving, and general-purpose post-training quantization (ptq) solution to enable 8-bit weight, 8-bit activation (w8a8) quantization for large language models (llms) that can be implemented efficiently.
Our work offers a turn-key solution that reduces hardware costs and democratizes llms.
Large language models (LLMs) have shown exceptional performance on a variety
of natural language tasks. Yet, their capabilities for HTML understanding --
i.e., parsing the raw HTML of a webpage, with
Are large language models (LLMs) like GPT-3 psychologically safe? In this
work, we design unbiased prompts to evaluate LLMs systematically from a
psychological perspective. Firstly, we test the person
Sentence Simplification aims to rephrase complex sentences into simpler
sentences while retaining original meaning. Large Language models (LLMs) have
demonstrated the ability to perform a variety of n
Large Language Models (LLMs) especially ChatGPT have produced impressive
results in various areas, but their potential human-like psychology is still
largely unexplored. Existing works study the virtu
We introduce a new in-context learning paradigm to measure Large Language
Models' (LLMs) ability to learn novel words during inference. In particular, we
rewrite Winograd-style co-reference resolution
Recent work in training large language models (LLMs) to follow natural
language instructions has opened up exciting opportunities for natural language
interface design. Building on the prior success o
Intrigued by the claims of emergent reasoning capabilities in LLMs trained on
general web corpora, in this paper, we set out to investigate their planning
capabilities. We aim to evaluate (1) how good
Large language models (llm) have been built to acquire co-occurrence-based knowledge from language corpora, but the robustness of their world knowledge has been questioned.
Using curated sets of minimal sentence pairs (n=1215), we tested whether llms are more likely to generate plausible eventdescriptions compared to their implausible counterparts.