Get fleet-rlm running locally in under five minutes.Documentation Index
Fetch the complete documentation index at: https://docs.qredence.ai/llms.txt
Use this file to discover all available pages before exploring further.
Prerequisites
- Python 3.10 or later
- uv package manager
- Daytona API key — sandbox execution backend
- An LLM API key — any LiteLLM-supported provider (OpenAI, Anthropic, Gemini, etc.)
1. Add fleet-rlm to a uv project
pnpm or a separate frontend build step.
2. Configure environment variables
Create a.env file in your working directory:
.env
3. Launch the Web UI
http://127.0.0.1:8000. You should see four product surfaces:
- Workbench — adaptive chat and runtime execution
- Volumes — runtime-backed file browsing
- Optimization — DSPy evaluation and GEPA workflows
- Settings — runtime configuration and diagnostics
4. Try a recursive task
In the Workbench, send a prompt that exceeds a single context window — for example, asking the agent to summarize a long document or analyze a repository. The agent will:- Decide the task is too large for one ReAct context.
- Call
delegate_to_rlmto spin up an isolated child Daytona sandbox. - Run a bounded
dspy.RLMinside the child, recursing further withsub_rlmif needed. - Stream intermediate trace events back to the UI and bubble the final answer into the chat.
Other surfaces
Terminal chat
API server only
Next steps
Installation guide
Install from PyPI or from source for contributors.
Core concepts
ReAct orchestration, recursive RLM, runtime surfaces.
Recursive RLM
How
delegate_to_rlm and child sandbox isolation work.Troubleshooting
Common installation and runtime issues.