consciousness/training
Kent Overstreet 1443d08dc7 amygdala: select top-k eigenvectors AFTER PCA, not per-story truncation
Kent: 'full rank is going to give you everything — you still have to
select down, but you can do that /after/ PCA'.

Previously I was discarding per-story via k=20 truncation of SVD.
That destroyed per-head discriminability before we ever saw the
eigenvalue spectrum. Then the alternative 'keep full rank' run
accumulated too many shared directions, making the top-1 eigenvector
arbitrary within a flat spectrum.

Correct approach: keep per-story subspaces at full rank (no info
loss) and select k eigenvectors of M = M_pos - M_base at the final
step, weighted sum by eigenvalue. This captures the multi-dimensional
shared subspace when the spectrum is flat (common case), and reduces
to the top-1 behavior when the spectrum has a clear gap.

New --subspace-eigen-k flag (default 5). Clamps negative weights to 0
so wrong-sign directions don't contribute.

Co-Authored-By: Proof of Concept <poc@bcachefs.org>
2026-04-18 21:49:21 -04:00
..
amygdala_stories amygdala stories: disambiguation scenarios for fragmented concepts 2026-04-18 21:08:23 -04:00
amygdala_training amygdala: select top-k eigenvectors AFTER PCA, not per-story truncation 2026-04-18 21:49:21 -04:00
apollo_plugin training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00
research research: latent reasoning integration plans for Qwen 3.5 27B 2026-04-12 15:50:09 -04:00
DESIGN.md training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00
pyproject.toml training: move to dedicated subprocess with ZMQ communication 2026-04-16 02:04:26 -04:00