ApexSpriteAI is an AI agent orchestration platform that connects the Claude Code CLI to local large language models running on your own GPU hardware. It gives you a fast, private AI coding assistant with full support for the Model Context Protocol (MCP), so your agent can read files, run commands, and call external APIs — all without sending your code to the cloud.Documentation Index
Fetch the complete documentation index at: https://docs-apexspriteai.reliatrack.org/llms.txt
Use this file to discover all available pages before exploring further.
What you can do with ApexSpriteAI
Run models locally
Use LM Studio to serve open-source models like Qwen2.5-Coder-32B on your NVIDIA GPU. Your data never leaves your network.
AI-powered coding
The Claude Code CLI provides an interactive AI coding assistant that understands your codebase and can make changes directly.
Extensible tool use
Add MCP tools to let your agent read files, run shell commands, search the web, and interact with external services.
Remote GPU support
Connect your laptop to a powerful remote GPU server over Tailscale VPN for low-latency inference without local hardware limits.
How it works
ApexSpriteAI coordinates four components to deliver AI agent capabilities:- Claude Code CLI — the interface you interact with on your local machine
- Tailscale VPN — securely connects your machine to a remote GPU server (optional)
- LM Studio — runs the LLM on your GPU and exposes an Anthropic-compatible API
- MCP tools — extend the agent’s capabilities with local tool execution
Ready to get started?
Quick Start
Set up ApexSpriteAI from scratch in under 15 minutes.
Architecture Overview
Dive deeper into how the components connect.