Skip to content
Back to writing
In progressresearchinterpretabilityclip

Visual NLA

An applied-research exploration extending Anthropic's May 2026 Natural Language Autoencoder technique from language-model activations to image embeddings.

Started
May 2026
Output
Blog · Repo · Demo

An applied-research exploration: train and probe a CLIP-based Visual Natural Language Autoencoder, extending the NLA technique (overview) from language-model activations to image embeddings.

The aim is a credible publishable artifact — a written explainer and an open-source training and probing repo. A hosted demo will follow once the model is trained.

Entries log the daily decisions, dead-ends, and small wins as the work moves from “read the paper” to “shipped artifact.”

Entries

0 total

No entries yet.