klotz: llm* + rope*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. Microsoft researchers introduce LongRoPE2, a method to extend large language model context windows to 128K tokens while maintaining over 97% short-context accuracy, addressing key limitations in positional embeddings.

  2. This article explains the Long RoPE methodology used to expand the context lengths in LLMs without significant performance degradation. It discusses the importance of context length in LLMs and the limitations of previous positional encoding methods. The article then introduces Rotational Positional Encoding (RoPE) and its limitations, and explains how Long RoPE extends RoPE to larger contexts.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llm + rope

About - Propulsed by SemanticScuttle