klotz: llm*

Bookmarks on this page are managed by an admin user.

0 bookmark(s) - Sort by: Date / Title ↓ / - Bookmarks from other users for this tag

  1. - Demonstrates how to improve two pretrained models' proficiency in the Dafny verified programming language.
    - Uses 178 programming problems from the MBPP dataset for prompting GPT-4 and PaLM-2 to generate methods in Dafny.
    - Three types of prompts were used: a direct contextless prompt, one that includes a signature of the method and test cases, and a third one that decomposes the problem into steps and includes dynamically chosen similar examples.
    - GPT-4 was able to generate verified (and human-evaluated) Dafny methods in 58% of the cases with the third prompt.
    - Contributes a collection of 153 MBPP problems implemented and formally verified in Dafny, 50 written by authors and 103 synthesized by GPT-4.
  2. Resource-efficient LLMs and Multimodal Models

    A useful survey of resource-efficient LLMs and multimodal foundations models.

    Provides a comprehensive analysis and insights into ML efficiency research, including architectures, algorithms, and practical system designs and implementations.
  3. Large language models exhibit surprising emergent generalization properties, yet also struggle on many simple reasoning tasks such as arithmetic and parity. This raises the question of if and when Transformer models can learn the true algorithm for solving a task. We study the scope of Transformers' abilities in the specific setting of length generalization on algorithmic tasks. Here, we propose a unifying framework to understand when and how Transformers can exhibit strong length generalization on a given task. Specifically, we leverage RASP (Weiss et al., 2021) -- a programming language designed for the computational model of a Transformer -- and introduce the RASP-Generalization Conjecture: Transformers tend to length generalize on a task if the task can be solved by a short RASP program which works for all input lengths. This simple conjecture remarkably captures most known instances of length generalization on algorithmic tasks. Moreover, we leverage our insights to drastically improve generalization performance on traditionally hard tasks (such as parity and addition). On the theoretical side, we give a simple example where the "min-degree-interpolator" model of learning from Abbe et al. (2023) does not correctly predict Transformers' out-of-distribution behavior, but our conjecture does. Overall, our work provides a novel perspective on the mechanisms of compositional generalization and the algorithmic capabilities of Transformers.
    2024-01-07 Tags: , , , by klotz
  4. Introduction of HuggingFace’s New Fine-tuned Model Zephyr-7B-α
    2023-10-28 Tags: , , , by klotz
  5. 2023-11-16 Tags: , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llm

About - Propulsed by SemanticScuttle