File size: 1,331 Bytes
4cf119e 4dfe94f 731438b 4dfe94f 4cf119e 330fff3 4dfe94f 4cf119e 4dfe94f 731438b 4cf119e 4dfe94f 731438b 4dfe94f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 | ---
title: TRIBE V2 — Brain Response Prediction
emoji: 🧠
colorFrom: indigo
colorTo: yellow
sdk: gradio
sdk_version: "6.11.0"
python_version: "3.12"
app_file: app.py
pinned: false
license: cc-by-nc-4.0
hardware: zero-a10g
---
# TRIBE V2 — Brain Response Prediction
Predicts fMRI brain responses to **video, audio, and text** using Meta's TRIBE V2 foundation model.
## Features
- **Text Scorer** — Paste a script/hook, get brain engagement scores (~30s)
- **Video Scorer** — Upload a video for full multimodal analysis (~2-5 min)
- **A/B Tester** — Compare two text versions head-to-head
- **API** — Programmatic JSON access
## How It Works
TRIBE V2 combines LLaMA 3.2-3B (text), V-JEPA2 (video), and Wav2Vec-BERT (audio) to predict cortical surface activations across 20,484 brain vertices. Scores are derived from region-of-interest analysis using the Destrieux atlas.
## Scores
- **Attention Capture** — Will they stop scrolling?
- **Emotional Valence** — Does it trigger feelings?
- **Language Processing** — Is the message clear?
- **Visual Imagery** — Are visuals compelling?
- **Viral Potential** — Composite engagement score
## Citation
```
@article{tribe2024,
title={A Foundation Model of Vision, Audition, and Language for In-Silico Neuroscience},
author={Meta FAIR},
year={2024}
}
```
|