This document details how to use function calling with Mistral AI models to connect to external tools and build more complex applications, outlining a four-step process: User query & tool specification, Model argument generation, User function execution, and Model final answer generation.
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). It provides a simple yet robust interface using llama-cpp-python, allowing users to chat with LLM models, execute structured function calls and get structured output.