GenieHive/docs/roadmap.md

35 lines
1.3 KiB
Markdown

# GenieHive Roadmap
## Completed Foundations
- control-plane registry with SQLite persistence
- node registration and heartbeat
- role catalog and route resolution
- client-facing `GET /v1/models`
- client-facing `POST /v1/chat/completions`
- client-facing `POST /v1/embeddings`
- first control-plus-node demo flow
## Immediate Next Milestones
1. Run and document the first live LLM demo against real upstream servers.
2. Validate the `GET /v1/models` metadata as a Codex-friendly offload catalog for lower-complexity tasks.
3. Add `POST /v1/audio/transcriptions`.
4. Add a richer node metrics model for queue depth, current load, and observed performance over time.
5. Add a stronger operator/client distinction in the public metadata and auth surfaces.
## LLM Demo Note
The project is now ready for a first live LLM demo using GenieHive as:
- master: control plane
- peer: one or more node agents with pre-existing local LLM servers
- client: a small demo agent or Codex configured against GenieHive
The current live-demo priority is chat-first. Embeddings are also wired in GenieHive, but upstream compatibility differs across local servers, so the safest first demo matrix is:
- Ollama for chat and embeddings
- vLLM for chat and embeddings
- llama.cpp for chat
- llamafile for chat