Add evidence trail learner workbench

This commit is contained in:
wesley 2026-04-25 13:40:25 +00:00 committed by welsberr
parent 59d45c2942
commit 074999fbe1
5 changed files with 1133 additions and 6 deletions

2
.gitignore vendored
View File

@ -13,3 +13,5 @@ build/
configs/config.yaml
tmp-*
codex*
ops/
webui/node_modules/

View File

@ -37,6 +37,15 @@ Didactopus is built around a few core ideas:
- explanations, critique, and next-step advice should preserve learner trust
- local and low-cost deployment matter for access
It should also operate under a scientific-virtues outlook. In practice that
means Didactopus should reinforce habits such as:
- curiosity about the question rather than premature closure
- honesty about what is observed versus what is inferred
- skepticism toward weakly supported claims, including model-generated claims
- attentiveness to source quality, caveats, and uncertainty
- willingness to revise when better evidence changes the picture
In practice, that means Didactopus tries to help with:
- topic structure
@ -198,6 +207,14 @@ It is also meant to support the pedagogy around learning:
- exporting evidence and capability artifacts
- supporting multilingual and accessible outputs
Operationally, the scientific-virtues framing means Didactopus should:
- separate observation from interpretation in learner-facing flows
- reward justified revision rather than answer persistence
- surface uncertainty explicitly instead of smoothing it away
- push learners toward source comparison and evidence quality checks
- avoid presenting confident unsupported synthesis as settled knowledge
This is why the repository contains review workspaces, validation flows, knowledge graphs, and capability export machinery rather than only a chat interface.
## Grounded AI Learner And Skill Production

View File

@ -22,6 +22,8 @@ Near-term scope:
- make mentor, practice, and evaluator turns consistently source-grounded
- improve trust-preserving feedback behavior
- extend the session flow beyond one short interaction
- make scientific virtues operational in the session loop by separating observation from interpretation, preserving uncertainty, and rewarding justified revision
- replace stubbed provider output in learner-facing pilot flows with configured real model backends where available
Current code anchors:
@ -102,6 +104,25 @@ Target features:
- active practice task
- evaluator feedback
- recommended next step
- first external pilot should use the `evidence-trail` evo-edu pack as a learner-workbench test case
Current progress:
- the first external pilot pack now exists at `domain-packs/evidence-trail/`
- `pack_to_frontend` output is generated and copied into `webui/public/packs/evidence-trail-pack.json`
- the web UI now has a learner-workbench launcher and `Evidence Trail` pilot mode in addition to the review workbench
- the learner pilot exposes question, observation, interpretation, uncertainty, and revision-trigger fields directly in the UI
- scientific virtues are now reflected in the UI framing and in backend learner-session prompt construction
- the backend now exposes `POST /api/learner-workbench/session`
- end-to-end verification succeeded locally: the API starts, the endpoint returns structured concept/session output, and the frontend/backend contract is working
Immediate next steps:
- replace current stubbed mentor/practice/evaluator text with a configured real provider path
- enrich the `Evidence Trail` pack with grounded source fragments so returned guidance is based on more than pack metadata
- persist learner-session state instead of treating each call as a stateless step
- connect learner progress, evidence, and revision history to the standard backend session model
- define deployment notes for running the learner workbench against the local API outside development mode
Current pilot state:
@ -208,13 +229,18 @@ Examples:
- Keep accessibility and low-cost deployment in scope from the start, not as cleanup work.
- Preserve provenance and license compliance as first-class constraints.
- Advance the current roadmap without assuming abundant compute, fluent English, expert supervision, or mature learners.
- Treat scientific virtues as operational principles: encourage curiosity, honesty about evidence, skepticism toward weak claims, attentiveness to caveats, and revision when the evidence changes.
- Separate observation from interpretation in learner-facing guidance so the system does not blur grounded support with model inference.
- Frame revision as progress rather than as failure, especially in mentor and evaluator feedback.
## Suggested Implementation Sequence
1. Strengthen `didactopus.learner_session` into the standard session backend.
2. Fold the learner-workbench pilot into that backend without losing its stronger study-state framing.
3. Build a small model-benchmark harness around the unified learner backend.
4. Add accessible learner HTML and text-first outputs.
5. Add local TTS and STT support to the same session flow.
6. Expand adaptive practice and diagnostics.
7. Improve review, impact analysis, and incremental update support.
3. Replace stubbed learner-workbench provider output with a configured real model backend.
4. Ground the `evidence-trail` pilot in richer source fragments and persisted learner state.
5. Build a small model-benchmark harness around the unified learner backend.
6. Add accessible learner HTML and text-first outputs.
7. Add local TTS and STT support to the same session flow.
8. Expand adaptive practice and diagnostics.
9. Improve review, impact analysis, and incremental update support.

1082
webui/package-lock.json generated Normal file

File diff suppressed because it is too large Load Diff