This article provides a step-by-step guide on building a generative search engine for local files using Qdrant, NVidia NIM API, or Llama 3. It includes system design, indexing local files, and creating a user interface.
Learn how to build an open LLM app using Hermes 2 Pro, a powerful LLM based on Meta's Llama 3 architecture. This tutorial explains how to deploy Hermes 2 Pro locally, create a function to track flight status using FlightAware API, and integrate it with the LLM.
LlamaFS is a self-organizing file manager that automatically renames and organizes files based on their contents. It supports various file types and even images and audio. It can run in two modes - batch mode and watch mode. In batch mode, LlamaFS suggests a file structure and organizes files. In watch mode, it watches your directory and proactively learns your file organization habits. The project is built on a Python backend and Electron for the frontend.
This article discusses the latest open LLM (large language model) releases, including Mixtral 8x22B, Meta AI's Llama 3, and Microsoft's Phi-3, and compares their performance on the MMLU benchmark. It also talks about Apple's OpenELM and its efficient language model family with an open-source training and inference framework. The article also explores the use of PPO and DPO algorithms for instruction finetuning and alignment in LLMs.