A-Evolve, a new framework developed by Amazon researchers, aims to revolutionize the development of agentic AI systems. It addresses the current bottleneck of manual tuning by introducing an automated evolution process. Described as a potential "PyTorch moment" for agentic AI, A-Evolve moves away from hand-tuned prompts towards a scalable system where agents improve their code and logic iteratively.
The framework centers around an ‘Agent Workspace’ with components like manifest files, prompts, skills, tools, and memory. A five-stage loop—Solve, Observe, Evolve, Gate, and Reload—ensures stable improvements. A-Evolve is modular, allowing for "Bring Your Own" approaches to agents, environments, and algorithms, and has demonstrated State-of-the-Art performance on benchmarks like MCP-Atlas and SWE-bench Verified.
This article introduces agentic TRACE, an open-source framework designed to build LLM-powered data analysis agents that eliminate data hallucinations. TRACE shifts the LLM's role from analyst to orchestrator, ensuring all computations are deterministic and data-driven. The framework achieves this by having the LLM work with metadata instead of raw data, relying on the database as the source of truth, and providing a complete audit trail. Example use cases demonstrate the system's ability to deliver verifiable results on inexpensive models like Gemini 3.1 Flash Lite. The author provides a quick start guide and encourages contributions to the project.
The Model Context Protocol (MCP) is becoming a key component in the agentic AI space, enabling models to interact with external tools and data. The project's 2026 roadmap focuses on addressing challenges for production deployment. Key priorities include improving scalability by evolving the transport and session model, clarifying agent communication and task lifecycle management, maturing governance structures for wider community contribution, and preparing for enterprise requirements like audit trails and authentication. The roadmap also highlights ongoing exploration of areas like event-driven updates and security.
Open-source coding agents like OpenCode, Cline, and Aider are reshaping the AI dev tools market. And OpenCode's new $10/month tier signals falling LLM costs. These agents act as a layer between developers and LLMs, interpreting tasks, navigating repositories, and coordinating model calls. They offer flexibility, allowing developers to connect their own providers and API keys, and are becoming increasingly popular as a way to manage the economics of running large language models. The emergence of these tools indicates a shift in value towards the agent layer itself, with subscriptions becoming a standard packaging method.
GenAI-based coding assistants are evolving towards agent-based tools that require contextual information. This paper presents a preliminary study investigating the adoption of AI context files (like AGENTS.md) in 466 open-source software projects, analyzing the information provided, its presentation, and evolution over time. The findings reveal a lack of established content structure and significant variation in context provision, highlighting opportunities for studying how structural and presentational modifications can improve generated content quality.
In this tutorial, we build a hierarchical planner agent using an open-source instruct model. We design a structured multi-agent architecture comprising a planner agent, an executor agent, and an aggregator agent, where each component plays a specialized role in solving complex tasks. We use the planner agent to decompose high-level goals into actionable steps, the executor agent to execute those steps using reasoning or Python tool execution, and the aggregator agent to synthesize results into a coherent final response. By integrating tool usage, structured planning, and iterative execution, we create a fully autonomous agent system that demonstrates how modern AI agents reason, plan, and act in a scalable and modular manner.
Alibaba has released CoPaw, an open-source framework designed to provide a standardized workstation for deploying and managing personal AI agents. It addresses the shift from LLM inference to autonomous agentic systems, focusing on the environment in which models operate. CoPaw utilizes AgentScope, AgentScope Runtime, and ReMe to handle agent logic, execution, and persistent memory, enabling long-term experience and multi-channel connectivity.
Vercel has released Skills.sh, an open-source tool designed to provide AI agents with a standardized way to execute reusable actions, or skills, through the command line. The project introduces what Vercel describes as an open agent skills ecosystem, where developers can define, share, and run discrete operations that agents can invoke as part of their workflows.
A curated collection of Awesome LLM apps built with RAG, AI Agents, Multi-agent Teams, MCP, Voice Agents, and more. This repository features LLM apps that use models from OpenAI, Anthropic, Google, xAI and open-source models like Qwen or Llama.
Kagent is an open-source agentic AI framework for Kubernetes that aims to provide autonomous problem solving and remediation for cloud-native infrastructure, moving beyond traditional automation to a more intelligent and self-healing system.