← Home

Quick answer

AI Summary: Discovers that multimodal models like CLIP naturally develop 'multimodal neurons' that abstractly link images, text, and sketches of a concept, revealing both profound cognitive similarities to the human brain and novel adversarial vulnerabilities.

Claim

Multimodal Neurons in Artificial Neural Networks

Gabriel Goh·
Nick Cammarata·
Chelsea Voss·
Shan Carter·
Michael Petrov·
Ludwig Schubert·
Alec Radford·
Chris Olah

ABSTRACT

We investigate the internal representations of the CLIP model and discover the presence of 'multimodal neurons'. These neurons fire not only for specific visual features (like a spider) but also for the text representing that concept, and even abstract sketches or comic depictions of it. We demonstrate that these networks naturally develop representations akin to the 'Halle Berry neuron' discovered in the human brain, grouping disparate visual and textual stimuli under single, highly abstract conceptual nodes. Furthermore, we reveal how these abstract representations make the model vulnerable to a novel typographic attack, where simply writing a word on an object can completely alter the model's classification.

Review Snapshot

Explore ratings

4.6
★★★★★
5 ratings
5 star
60%
4 star
40%
3 star
0%
2 star
0%
1 star
0%

Recommendation

100%

recommend this content.

Review this content

Share your opinion to help other learners triage faster.

Write a review

Invite a reviewer

Invite someone by email to share an invited review for Multimodal Neurons in Artificial Neural Networks.

Author Inquiries

Public questions about this content. Attendemia will route your question to the author. Vote on the most important ones. No guarantee of response.
Post an inquiry
Sort by: Most helpful