Predict cortical surface activity from text, audio, and video.
Watch NForge simulate a full inference pipeline — from model loading to cortical surface predictions.
The brain doesn't process language, sound, and video in isolation — it integrates them into a unified perceptual experience. NForge models this process by predicting human fMRI responses to naturalistic multimodal stimuli.
Built on Meta's TRIBE v2 with significant architectural improvements and four new capabilities.
Visualise which temporal windows most strongly drove each brain region. Hook into transformer self-attention layers and project scores onto the HCP MMP1.0 parcellation.
Run sliding-window predictions from a live feature stream without pre-loading the full clip. Thread-safe, configurable context window and step size.
Per-vertex importance scores showing how much text, audio, and video contributed to each prediction. Supports ablation and integrated gradient methods.
Few-shot adaptation to unseen subjects via ridge regression or nearest-neighbour matching. No full retraining needed.
Optional backbone compilation for faster training and inference. Compiles the encoder and combiner while preserving dynamic subject layer indexing.
Professionally organized with clean subpackages: core, data, training, inference, viz, and configs. Full test coverage with pytest.
End-to-end architecture from raw stimuli to cortical surface predictions.
Install NForge with pip. Choose the extras you need.
NForge supports training on multiple neuroimaging datasets out of the box.
| Dataset | Subjects | Stimuli | TR (s) |
|---|---|---|---|
| Algonauts2025Bold | 4 | TV sitcom "Friends" + movies | 1.49 |
| Wen2017 | 3 | Short videos (11.7s) | ~2 |
| Lahner2024Bold | 10 | Short videos (6.2s) | ~2 |
| Lebel2023Bold | 8 | Spoken narrative (6-18s) | ~2 |
NForge enables new research directions in computational neuroscience and beyond.
Run experiments locally or on a SLURM cluster.