consciousness/training
Kent Overstreet 047da10123 training: add preflight checks + progress logging to trainer
Review pass before running on b200. 27B model + 100+ story corpus
means any misconfiguration costs real time; better to fail before
model load and give visible progress during forwards.

* Pre-load-model validation: stories-dir and paired-dir exist,
  corpus has >= min_positives emotions.
* Per-batch progress log every 5 batches with elapsed + ETA.
* Relative depth printed for target layers (e.g. "layer 40 (51%)").
* Skip empty .txt files with a warning rather than feeding the
  tokenizer an empty string.
* Assert non-empty strings in _collect_activations.

Co-Authored-By: Proof of Concept <poc@bcachefs.org>
2026-04-18 01:06:07 -04:00
..
amygdala_stories training/amygdala_stories: scaffold + initial batch of 15 stories 2026-04-18 01:06:07 -04:00
amygdala_training training: add preflight checks + progress logging to trainer 2026-04-18 01:06:07 -04:00
apollo_plugin training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00
research research: latent reasoning integration plans for Qwen 3.5 27B 2026-04-12 15:50:09 -04:00
DESIGN.md training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00
pyproject.toml training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00