Tags: transformers*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. This article explains BERT, a language model designed to understand text rather than generate it. It discusses the transformer architecture BERT is based on and provides a step-by-step guide to building and training a BERT model for sentiment analysis.
  2. This tutorial covers fine-tuning BERT for sentiment analysis using Hugging Face Transformers. Learn to prepare data, set up environment, train and evaluate the model, and make predictions.
  3. Discusses the trends in Large Language Models (LLMs) architecture, including the rise of more GPU, more weights, more tokens, energy-efficient implementations, the role of LLM routers, and the need for better evaluation metrics, faster fine-tuning, and self-tuning.
  4. This article explains the Long RoPE methodology used to expand the context lengths in LLMs without significant performance degradation. It discusses the importance of context length in LLMs and the limitations of previous positional encoding methods. The article then introduces Rotational Positional Encoding (RoPE) and its limitations, and explains how Long RoPE extends RoPE to larger contexts.
  5. python -c "from transformers import pipeline; print(pipeline('sentiment-analysis')('I love you'))"
  6. Exploring the architecture of OpenAI’s Generative Pre-trained Transformers.
    2023-12-10 Tags: , , by klotz
  7. Delving into transformer networks
  8. 2023-06-29 Tags: , , , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "transformers"

About - Propulsed by SemanticScuttle