This repository contains the Llama Stack API specifications as well as API Providers and Llama Stack Distributions. The Llama Stack aims to standardize the building blocks needed for generative AI applications across various development stages.
It includes API specifications and providers for the Llama Stack, which aims to standardize components needed for developing generative AI applications. The stack includes APIs for Inference, Safety, Memory, Agentic System, Evaluation, Post Training, Synthetic Data Generation, and Reward Scoring. Providers offer actual implementations for these APIs, either through open-source libraries or remote REST services.
Dune is a shell designed for powerful scripting, combining elements of bash and Lisp, offering normal shell operations and functional programming abstractions for sysadmin tasks.
Datasette is introduced as a functional interactive frontend to tabulated data, either in CSV format or a database schema, catering to data journalists, museum curators, archivists, local governments, and researchers.
The author explores creating tables and inserting data into a SQLite database, then targets the database with Datasette to showcase how errors in data can be identified and corrected.
ASCVIT V1 aims to make data analysis easier by automating statistical calculations, visualizations, and interpretations.
Includes descriptive statistics, hypothesis tests, regression, time series analysis, clustering, and LLM-powered data interpretation.
- Accepts CSV or Excel files. Provides a data overview including summary statistics, variable types, and data points.
- Histograms, boxplots, pairplots, correlation matrices.
- t-tests, ANOVA, chi-square test.
- Linear, logistic, and multivariate regression.
- Time series analysis.
- k-means, hierarchical clustering, DBSCAN.
Integrates with an LLM (large language model) via Ollama for automated interpretation of statistical results.
InstructLab is an open-source project that facilitates contributions to Large Language Models (LLMs) by enabling community members to add 'skills' or 'knowledge' to existing models. InstructLab uses a model-agnostic technology to allow model creators to integrate new skills without retraining the entire model.
ntfy is a simple HTTP-based pub-sub notification service that allows you to send notifications to your phone or desktop via scripts or using a REST API. It's open source and free software.
A list of 13 open-source software for building and managing production-ready AI applications. The tools cover various aspects of AI development, including LLM tool integration, vector databases, RAG pipelines, model training and deployment, LLM routing, data pipelines, AI agent monitoring, LLM observability, and AI app development.
1. Composio - Seamless integration of tools with LLMs.
2. Weaviate - AI-native vector database for AI apps.
3. Haystack - Framework for building efficient RAG pipelines.
4. LitGPT - Pretrain, fine-tune, and deploy models at scale.
5. DsPy - Framework for programming LLMs.
6. Portkey's Gateway - Reliably route to 200+ LLMs with one API.
7. AirByte - Reliable and extensible open-source data pipeline.
8. AgentOps - Agents observability and monitoring.
9. ArizeAI's Phoenix - LLM observability and evaluation.
10. vLLM - Easy, fast, and cheap LLM serving for everyone.
11. Vercel AI SDK - Easily build AI-powered products.
12. LangGraph - Build language agents as graphs.
13. Taipy - Build AI apps in Python.
This article lists five of the best open-source email clients for Linux, with a focus on Geary as the author's preferred choice. The article provides details on each client, including its strengths, weaknesses, and suitability for different user needs.
A ruby script calculates VRAM requirements for large language models (LLMs) based on model, bits per weight, and context length. It can determine required VRAM, maximum context length, or best bpw given available VRAM.
How to use Kubernetes to manage and streamline AI workflows, leveraging the power of open source tools and the Kubernetes AI Toolchain Operator.