Michal Golovanevsky Profile
Michal Golovanevsky

@MichalGolov

Followers
30
Following
13
Statuses
15

CS PhD student @BrownCSDept | Biomedical AI | Multimodal Learning.

Providence, RI
Joined September 2022
Don't wanna be here? Send us removal request.
@MichalGolov
Michal Golovanevsky
4 months
RT @WilliamRudmanjr: NOTICE uses Symmetric Token Replacement for text corruption and Semantic Image Pairs (SIP) for image corruption. SIP r…
0
1
0
@MichalGolov
Michal Golovanevsky
4 months
RT @WilliamRudmanjr: We extend the generalizability of NOTICE by using Stable-Diffusion to generate semantic image pairs and find results a…
0
1
0
@MichalGolov
Michal Golovanevsky
4 months
RT @WilliamRudmanjr: The finding that important attention heads implement one of a small set of interpretable functions boosts transparency…
0
2
0
@MichalGolov
Michal Golovanevsky
4 months
RT @WilliamRudmanjr: How do VLMs like BLIP and LLaVA differ in how they process visual information? Using our mech-interp pipeline for VLMs…
0
2
0
@MichalGolov
Michal Golovanevsky
4 months
RT @WilliamRudmanjr: Instead, LLaVA relies on self-attention heads to manage “outlier” attention patterns in the image, focusing on regulat…
0
1
0
@MichalGolov
Michal Golovanevsky
7 months
RT @WilliamRudmanjr: The finding that important cross-attention heads implement one of a small set of interpretable functions helps boost V…
0
1
0
@MichalGolov
Michal Golovanevsky
7 months
RT @WilliamRudmanjr: By visualizing cross-attention patterns, we've discovered that these universal heads fall into three functional catego…
0
1
0
@MichalGolov
Michal Golovanevsky
7 months
RT @WilliamRudmanjr: Performing activation patching with NOTICE reveals a set of Universal Cross-Attention Heads that have a significant pa…
0
1
0
@MichalGolov
Michal Golovanevsky
7 months
RT @WilliamRudmanjr: NOTICE uses Symmetric Token Replacement for text corruption and introduces Semantic Minimal Pairs (SMP) for image corr…
0
1
0
@MichalGolov
Michal Golovanevsky
7 months
RT @WilliamRudmanjr: Mechanistic interpretability has advanced our understanding of LLMs, but what about multimodal models? Introducing NOT…
0
6
0
@MichalGolov
Michal Golovanevsky
2 years
Lastly, we showcased the importance of structured clinical data. Using ablation analysis, we determined that structured clinical data (demographics, memory scores, balance testing, etc.) helps performance the most when combined with other modalities.
Tweet media one
0
0
1
@MichalGolov
Michal Golovanevsky
2 years
Through our experiments, we found that cross-modal attention in combination with self-attention contributes the most to performance, followed by cross-modal attention alone. This demonstrates the value of capturing cross-modal interactions through attention.
Tweet media one
0
0
1
@MichalGolov
Michal Golovanevsky
2 years
My work with @CarstenEickhoff and Ritambhara Singh, used attention-based deep learning to classify patients into control, moderate cognitive impairment, and Alzheimer’s disease. Our model achieved state-of-the-art performance on the ADNI dataset, at 96.88% accuracy.
Tweet media one
0
0
1