KnowledgeRefinery/docs/operational-notes.md

67 lines
2.7 KiB
Markdown
Raw Permalink Normal View History

# Operational Notes
## Data Locations
Each workspace has its own data directory under `~/.knowledge-refinery/workspaces/<id>/`.
| Item | Path |
|------|------|
| Workspace root | `~/.knowledge-refinery/workspaces/<id>/` |
| SQLite DB (metadata + vectors + graph) | `~/.knowledge-refinery/workspaces/<id>/refinery.db` |
| PID file | `~/.knowledge-refinery/workspaces/<id>/daemon.pid` |
| Workspace config | `~/.knowledge-refinery/workspaces.json` |
## Resetting
To start fresh, remove the data directory:
```bash
rm -rf ~/.knowledge-refinery
```
## Monitoring
The Go daemon logs to stdout with structured messages. Key log patterns:
- `[pipeline] Stage: ...` - Pipeline stage progress
- `[embedder] Embedded batch: X chunks` - Embedding progress
- `[error]` - Errors during processing
### Live Pipeline Monitoring
During pipeline execution, real-time progress is available via the enriched `/ingest/status` endpoint. The daemon maintains:
- **Live progress dict**: Per-stage status (pending/running/done) with progress percentages
- **Counters**: chunk_count, annotation_count, concept_count, edge_count
- **Activity log**: 200-entry ring buffer; the last 50 events are returned via the API
The SwiftUI app polls at 1.5-second intervals and renders a Pipeline Progress Panel with stage checkmarks, animated counters, and an auto-scrolling activity log. Polling auto-stops when the pipeline reaches idle/done state. The universe visualization auto-refreshes every 5 seconds during processing using `mergeUniverse()` for incremental node injection.
## API Endpoints
| Method | Path | Description |
|--------|------|-------------|
| GET | /health | Health check |
| POST | /volumes/add | Add watched directory |
| GET | /volumes/list | List watched directories |
| DELETE | /volumes/remove | Remove watched directory |
| POST | /ingest/start | Start pipeline |
| GET | /ingest/status | Pipeline status |
| POST | /search | Vector search |
| GET | /evidence/{asset_id} | Get asset info |
| GET | /evidence/chunk/{chunk_id} | Get chunk details |
| GET | /evidence/assets/all | List all assets |
| GET | /universe/snapshot?lod=macro | Universe snapshot |
| POST | /universe/focus | Focus on node |
| POST | /concepts/refine | Refine concept |
| GET | /concepts/list | List concepts |
| GET | /concepts/{id} | Concept detail |
## Performance Considerations
- Large files (>500MB) are skipped by default
- Embedding batch size defaults to 32
- SQLite uses WAL mode for concurrent reads
- Pipeline runs in a background goroutine
- Incremental processing skips unchanged files (content hash comparison)
- Vector search: brute-force cosine similarity, all vectors loaded in memory (~150MB for 50K vectors)
- Go daemon starts in <100ms, uses ~30MB base memory