Tags: flash attention* + grouped-query-attention* + llm* + multi-query-attention* + quantization* + key-value cache*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. Improving the memory and computational efficiency of Large Language Models (LLMs) for handling long input sequences, including retrieval augmented questions answering, summarization, and chat tasks. It covers various techniques, such as lower precision computing, Flash Attention algorithm, positional embedding methods, and key-value caching strategies. These methods help reduce memory consumption and increase inference speeds while maintaining high accuracy levels in LLM applications. Furthermore, it highlights some advanced approaches like Multi-Query-Attention (MQA) and Grouped-Query-Attention (GQA), which further enhance computational and memory efficiency without compromising performance.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "flash attention+grouped-query-attention+llm+multi-query-attention+quantization+key-value cache"

About - Propulsed by SemanticScuttle