Tags: function calling* + mcp* + llm* + python*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. LLM 0.26 introduces tool support, allowing LLMs to access and utilize Python functions as tools. The article details how to install, configure, and use these tools with various LLMs like OpenAI, Anthropic, Gemini, and Ollama models, including examples with plugins and ad-hoc functions. It also discusses the implications for building 'agents' and future development plans.
  2. This tutorial demonstrates how to integrate Google’s Gemini 2.0 with an in-process Model Context Protocol (MCP) server using FastMCP, creating tools for weather information and integrating them into Gemini's function calling workflow.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "function calling+mcp+llm+python"

About - Propulsed by SemanticScuttle