0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag
This guide demonstrates how to execute end-to-end LLM workflows for developing and productionizing LLMs at scale. It covers data preprocessing, fine-tuning, evaluation, and serving.
This post discusses a study that finds that refusal behavior in language models is mediated by a single direction in the residual stream of the model. The study presents an intervention that bypasses refusal by ablating this direction, and shows that adding in this direction induces refusal. The study is part of a scholars program and provides more details in a forthcoming paper.
This article announces a comprehensive course on fine-tuning large language models (LLMs) offered on the freeCodeCamp.org YouTube channel. The course, developed by Krish Naik, covers topics such as QLORA, LORA, quantization with LLama2, gradient, and Google Gemma Model, among others. The course aims to help learners deepen their understanding of machine learning and artificial intelligence.
In this tutorial, learn how to improve the performance of large language models (LLMs) by utilizing a proxy tuning approach, which enables more efficient fine-tuning and better integration with the AI model.
Generate instruction datasets for fine-tuning Large Language Models (LLMs) using lightweight libraries and documents.
efficient method for fine-tuning LLM using LoRA and QLoRA, making it possible to train them even on consumer hardware
DocLLM is a lightweight extension to traditional LLMs for reasoning over visual documents, considering both textual semantics and spatial layout. It avoids expensive image encoders and focuses on bounding box information. It outperforms SotA LLMs on 14 out of 16 datasets across all tasks and generalizes well to previously unseen datasets.
Keywords:
First / Previous / Next / Last
/ Page 2 of 0