consciousness/poc-memory
Kent Overstreet a29b6d4c5d Add direct API backend for agent execution
When api_base_url is configured, agents call the LLM directly via
OpenAI-compatible API (vllm, llama.cpp, etc.) instead of shelling
out to claude CLI. Implements the full tool loop: send prompt, if
tool_calls execute them and send results back, repeat until text.

This enables running agents against local/remote models like
Qwen-27B on a RunPod B200, with no dependency on claude CLI.

Config fields: api_base_url, api_key, api_model.
Falls back to claude CLI when api_base_url is not set.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-18 23:05:14 -04:00
..
.claude feedback: not-relevant/not-useful commands, edge strength adjustment 2026-03-14 12:14:17 -04:00
agents Move poc-agent into workspace, improve agent prompts 2026-03-18 22:45:01 -04:00
defaults split into workspace: poc-memory and poc-daemon subcrates 2026-03-08 20:43:59 -04:00
schema transcript progress: capnp append-only log replaces stub nodes 2026-03-16 17:40:32 -04:00
src Add direct API backend for agent execution 2026-03-18 23:05:14 -04:00
build.rs split into workspace: poc-memory and poc-daemon subcrates 2026-03-08 20:43:59 -04:00
Cargo.toml Add direct API backend for agent execution 2026-03-18 23:05:14 -04:00
config.example.jsonl split into workspace: poc-memory and poc-daemon subcrates 2026-03-08 20:43:59 -04:00