Meta releases Llama 3.2, which features small and medium-sized vision LLMs (11B and 90B) alongside lightweight text-only models (1B and 3B). It also introduces the Llama Stack Distribution.
A step-by-step guide on building llamafiles from Llama 3.2 GGUFs, including scripting and Dockerization.
Learn how to deploy a private instance of Llama 3.2 with a Retriever-Augmented Generation (RAG) API using Lightning AI Studios, enabling you to leverage large language models in a secure and customizable environment.