Tags: explainability* + llm* + mistral 7b*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. This paper explores whether some language model representations may be inherently multi-dimensional, contrasting the linear representation hypothesis. The authors develop a method using sparse autoencoders to find multi-dimensional features in GPT-2 and Mistral 7B. They find interpretable examples such as circular features representing days of the week and months of the year, which are used to solve computational problems involving modular arithmetic.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "explainability+llm+mistral 7b"

About - Propulsed by SemanticScuttle