This repository contains the Llama Stack API specifications as well as API Providers and Llama Stack Distributions. The Llama Stack aims to standardize the building blocks needed for generative AI applications across various development stages.
It includes API specifications and providers for the Llama Stack, which aims to standardize components needed for developing generative AI applications. The stack includes APIs for Inference, Safety, Memory, Agentic System, Evaluation, Post Training, Synthetic Data Generation, and Reward Scoring. Providers offer actual implementations for these APIs, either through open-source libraries or remote REST services.
llama-cpp-python offers a web server which aims to act as a drop-in replacement for the OpenAI API. This allows you to use llama.cpp compatible models with any OpenAI compatible client (language libraries, services, etc).
The author has also automated their weeknotes by using an Observable notebook, which generates the "releases this week" and "TILs this week" sections.
The notebook fetches TILs from the author's Datasette, grabs releases from GitHub, and assembles a markdown string for the new post.
* `llm` CLI tool for running prompts against large language models
* Automation of weeknotes using an Observable notebook
* Notebook generates "releases this week" and "TILs this week" sections
* Tool stores prompts and responses in a SQLite database