ShellGPT is a powerful command-line productivity tool driven by large language models like GPT-4. It is designed to streamline the development workflow by generating shell commands, code snippets, and documentation directly within the terminal, reducing the need for external searches. The tool supports multiple operating systems including Linux, macOS, and Windows, and is compatible with various shells such as Bash, Zsh, and PowerShell. Beyond simple queries, it offers advanced features like shell integration for automated command execution, a REPL mode for interactive chatting, and the ability to implement custom function calls. Users can also leverage local LLM backends like Ollama for a free, privacy-focused alternative to OpenAI's API.
This article details how to use Ollama to run large language models locally, protecting sensitive data by keeping it on your machine. It covers installation, usage with Python, LangChain, and LangGraph, and provides a practical example with FinanceGPT, while also discussing the tradeoffs of using local LLMs.
This article details the process of running a personal AI assistant on a low-cost microcontroller. It covers the use of Ollama for running large language models (LLMs) locally and MimicLaw for optimizing the model for resource-constrained devices. The author shares their experience with porting and running the models, along with the challenges and solutions encountered.
This article discusses how to effectively prompt local Large Language Models (LLMs) like those run with LM Studio or Ollama. It explains that local LLMs behave differently than cloud-based models and require more explicit and structured prompts for optimal results. The article provides guidance on how to craft better prompts, including using clear language, breaking down tasks into steps, and providing examples.
This post reviews two LLM options in Emacs - Ellama and gptel - and how to set them up, including adding models from OpenRouter and Ollama.
This article details the setup and initial testing of Goose, an open-source agent framework, paired with Ollama and the Qwen3-coder model, as a free alternative to Claude Code. It covers the installation process, initial performance observations, and a comparison to cloud-based solutions.
This article provides a comprehensive guide on implementing the Model Context Protocol (MCP) with Ollama and Llama 3, covering practical implementation steps and use cases.
The ollama 0.14-rc2 release introduces experimental functionality allowing LLMs to use tools like bash and web searching on your system, with safeguards like interactive approval and command allow/denylists.
This article details how to build a 100% local MCP (Model Context Protocol) client using LlamaIndex, Ollama, and LightningAI. It provides a code walkthrough and explanation of the process, including setting up an SQLite MCP server and a locally served LLM.
A tutorial on building a private, offline Retrieval Augmented Generation (RAG) system using Ollama for embeddings and language generation, and FAISS for vector storage, ensuring data privacy and control.
1. **Document Loader:** Extracts text from various file formats (PDF, Markdown, HTML) while preserving metadata like source and page numbers for accurate citations.
2. **Text Chunker:** Splits documents into smaller text segments (chunks) to manage token limits and improve retrieval accuracy. It uses overlapping and sentence boundary detection to maintain context.
3. **Embedder:** Converts text chunks into numerical vectors (embeddings) using the `nomic-embed-text` model via Ollama, which runs locally without internet access.
4. **Vector Database:** Stores the embeddings using FAISS (Facebook AI Similarity Search) for fast similarity search. It uses cosine similarity for accurate retrieval and saves the database to disk for quick loading in future sessions.
5. **Large Language Model (LLM):** Generates answers using the `llama3.2` model via Ollama, also running locally. It takes the retrieved context and the user's question to produce a response with citations.
6. **RAG System Orchestrator:** Coordinates the entire workflow, managing the ingestion of documents (loading, chunking, embedding, storing) and the querying process (retrieving relevant chunks, generating answers).