agent-core

agent-coreは、LLM(大規模言語モデル)を使用したエージェントAIワークフローを簡単に作成できるフレームワークです。MCPとの統合が容易で、Pythonパッケージとして利用可能です。ユーザーは20行未満のコードでAIエージェントを作成でき、会話の履歴管理も自動で行われます。

GitHubスター

0

ユーザー評価

未評価

フォーク

0

イシュー

1

閲覧数

3

お気に入り

0

README
agent-core

A framework for working with LLMs and creating agentic AI workflows compatible with MCP meant to replace my earlier ai-tools repo and power the v2v-chatbot-api, future AI projects, and work across both JavaScript and Python

Features
  • 🎯 AI agents in fewer than 20 lines of code
  • 🛠️ Super easy MCP integration
  • 🤖 Near universal support (OpenAI, Anthropic, Google Gemini, Ollama)
  • 📝 Automatic conversation management with history
  • 📦 Available as a Python package for easy integration
Step 1: Installation
Use as a Library in Your Project

Add to your project's requirements.txt:

git+https://github.com/BSchoolland/agent-core.git

Then install:

pip install -r requirements.txt
Step 2: Usage

Take a look at the examples

from agent_core import Agent
import asyncio

async def main():
    agent = await Agent.create(
        model='gemini-2.0-flash', # any valid model name you have access to
        mcp_servers=['examples/mcp_server.py']
    )
    await agent.run("Generate 2 random numbers between 1-100, add them together and save the result in a file called 'result.txt'")
    await agent.close()

if __name__ == '__main__':
asyncio.run(main())
Step 3: Profit

Claim your agents can replace all workers in 6 months and raise millions of dollars using pure AI hype.

Local model setup (optional)

Local models are managed through the Ollama CLI. See ollama.com for installation instructions.

A model must be installed through the Ollama CLI before it can be used with agent-core. You can install a model by running ollama pull <model-name> once Ollama is installed.

Development settup
1. Clone the Repository
git clone https://github.com/BSchoolland/agent-core.git
cd agent-core
2. Environment Setup

Copy .env.example to a new file called .env and replace any of the three API keys with your actual key (which you'll need to obtain from the provider).

cp .env.example .env

For this example I'll use Gemini because as of June 2025 their API has a free tier and obtaining an API key is super simple provided you have a Google account.

3. Installation

Set up a virtual environment

python3 -m venv venv
source venv/bin/activate

Install dependencies

pip install -r requirements.txt

# For CLI usage (optional)
pip install -e .
Documentation
  • CLI Documentation - Detailed CLI usage and examples
  • Agent Core Documentation - Framework internals and API reference
  • Examples - Code examples for library usage and sample MCP server
  • FastMCP - Suggested framework for building MCP servers. Note that you don't have to use this, you can use any MCP server you want. I just think this is nice and easy.
In depth documentation

So far, I've gone over only the basics. The below lists all the options and features these classes support.

Agent

Parameters:

  • model: The model to use for the agent. Can be any model from OpenAI, Anthropic, Google, or huggingface through Ollama.
  • provider: The provider to use for the agent. Can be openai, anthropic, google, or ollama. If this is not specified, the framework will attempt to automatically detect the provider based on the model name.
  • max_steps: The maximum number of act steps the agent will take before automatically failing. This defaults to 10 to prevent infinite loops, and does not include planning or reasoning steps.
  • mcp_servers: A list of MCP servers to use for the agent. Can be any MCP server you want.
  • type: The type of agent to create the types are as follows:
    • react: Follows the pattern reason->act->reason->act->... until the goal is complete. This is the default setting and is great for most tasks.
    • planner: Follows the pattern plan->act->act->act->... until the goal is complete. This is ideal for relatively straight forward tasks that the agent can make a complete plan for before starting.
    • hybrid: Follows the pattern plan->reason->act->reason->act->... until the goal is complete. Think of this as a hybrid of the react and planner types. It takes the longest to run and requires the most tokens but can be better than plain react for well defined yet complex tasks.
    • simple: Follows the pattern act->act->act... until the goal is complete. This tends to be less accurate but is faster and cheaper for simple tasks.

Methods:

  • run(goal: str): Run the agent with the given goal.
  • close(): Close the agent.
Conversation

Parameters:

  • model: The model to use for the conversation. Can be any model from OpenAI, Anthropic, Google, or huggingface through Ollama.
  • provider: The provider to use for the conversation. Can be openai, anthropic, google, or ollama. If this is not specified, the framework will attempt to automatically detect the provider based on the model name.

Methods:

  • generate_response(message: str): Generate a response to the given message.
  • close(): Close the conversation.
Supported Models

The framework automatically detects which models are actually available from each provider:

  • OpenAI: All available models (gpt-4o, gpt-4o-mini, o1, etc.)
  • Anthropic: All available Claude models (claude-3-5-sonnet, claude-3-haiku, etc.)
  • Google: All available Gemini models (gemini-2.5-pro, gemini-2.5-flash, etc.)
  • Ollama: All locally installed models available through the Ollama CLI, from Meta, Google, DeepSeek, etc.

Use agentcore list to see exactly which models are available to your system right now.

CLI Usage

The CLI provides easy access to both chat and agent functionality:

# Interactive chat with a model
agentcore chat gpt-4o-mini

# TODO: run agents with MCP servers through the terminal

# List available models
agentcore list

See the CLI Documentation for detailed usage instructions.

Roadmap
  • Multi-MCP server support, right now it's a one at a time thing
  • Release as an npm package using a js wrapper
  • Publish to PyPI for easier installation
作者情報
Benjamin Schoolland

Software developer with experience in fullstack web development and a strong foundation across multiple programming languages.

@BizzNEST

11

フォロワー

42

リポジトリ

0

Gist

27

貢献数

トップ貢献者

スレッド