1.3 KiB
1.3 KiB
GenieHive Roadmap
Completed Foundations
- control-plane registry with SQLite persistence
- node registration and heartbeat
- role catalog and route resolution
- client-facing
GET /v1/models - client-facing
POST /v1/chat/completions - client-facing
POST /v1/embeddings - first control-plus-node demo flow
Immediate Next Milestones
- Run and document the first live LLM demo against real upstream servers.
- Validate the
GET /v1/modelsmetadata as a Codex-friendly offload catalog for lower-complexity tasks. - Add
POST /v1/audio/transcriptions. - Add a richer node metrics model for queue depth, current load, and observed performance over time.
- Add a stronger operator/client distinction in the public metadata and auth surfaces.
LLM Demo Note
The project is now ready for a first live LLM demo using GenieHive as:
- master: control plane
- peer: one or more node agents with pre-existing local LLM servers
- client: a small demo agent or Codex configured against GenieHive
The current live-demo priority is chat-first. Embeddings are also wired in GenieHive, but upstream compatibility differs across local servers, so the safest first demo matrix is:
- Ollama for chat and embeddings
- vLLM for chat and embeddings
- llama.cpp for chat
- llamafile for chat