fionakryan Profile Banner
Fiona Ryan Profile
Fiona Ryan

@fionakryan

Followers
946
Following
1K
Statuses
229

Computer Science PhD student @GeorgiaTech and NSF Graduate Research Fellow interested in understanding human behavior with computer vision

Atlanta, GA
Joined August 2018
Don't wanna be here? Send us removal request.
@fionakryan
Fiona Ryan
3 days
@rebeccayelin soooo cool!!
1
0
1
@fionakryan
Fiona Ryan
4 days
RT @dimadamen: 🛑📢 HD-EPIC: A Highly-Detailed Egocentric Video Dataset New collected videos��
0
26
0
@fionakryan
Fiona Ryan
5 days
RT @ego4_d: Ego-Exo4D is the world's largest source of egocentric body + hand pose estimates and eye gaze data exists across the dataset. W…
0
2
0
@fionakryan
Fiona Ryan
21 days
RT @maxxu05: My paper RelCon: Relative Contrastive Learning for a Motion Foundation Model for Wearable Data, from my @Apple internship, ha…
0
9
0
@fionakryan
Fiona Ryan
1 month
nice Gaze-LLE demo from @fffiloni!
@fffiloni
Sylvain Filoni
1 month
Gaze-LLE @gradio demo is out on @huggingface, running with ZeroGPU 🤗 I provide a Dockerfile for local installation too if needed 😉 Space link: 👀
Tweet media one
0
0
6
@fionakryan
Fiona Ryan
1 month
@fffiloni @huggingface @moondreamai @vikhyatk moondream is so fast, super cool work! @vikhyatk
0
0
1
@fionakryan
Fiona Ryan
1 month
@BazaritoDorito @fffiloni @Gradio @huggingface @RemiCadene absolutely! there's uncertainty in all predictions, and our model is designed to produce a heatmap over different targets. in the GazeFollow test set, which we benchmark on, the images have multiple raters and in ambiguous cases the annotators often select different targets too.
Tweet media one
0
0
1
@fionakryan
Fiona Ryan
1 month
@BazaritoDorito @fffiloni @Gradio @huggingface @RemiCadene the model is trained on images where human annotators estimated the gaze targets, including some cases where the eyes aren't visible. the model likely uses other cues like the general head direction and faces being a common gaze target to make this prediction
1
0
2
@fionakryan
Fiona Ryan
1 month
RT @dimadamen: Ego4D and Ego-Exo4D @ego4_d Consortium held a short online meeting last month sharing some of the members' works around usin…
0
3
0
@fionakryan
Fiona Ryan
1 month
@jacob_g_martin1 another thing is that these datasets are annotated by humans who are simply guessing where the person is looking. some have multiple annotators to capture uncertainty, but many only have one. would love to hear any thoughts you have on how to evaluate models for this task!
1
0
1
@fionakryan
Fiona Ryan
1 month
@jacob_g_martin1 the coordinates of the prediction and ground truth are normalized so that the height and width are equal to 1 prior to L2 calculation (i.e. x = x / width, y = y/ height). we do this to follow the established metric calculation protocol for the benchmarks we compare on
1
0
0
@fionakryan
Fiona Ryan
2 months
@Abhinav95_ thank you!🤗
0
0
1
@fionakryan
Fiona Ryan
2 months
@NeuralBets_ @tom_doerr I'm less familiar with the benchmarks in this area, and there are commercial solutions like Tobii that could be the best. this article compares some open source options like OpenGaze and WebGazer
0
0
4
@fionakryan
Fiona Ryan
2 months
@Sourabh85426135 check out tab 3. in the paper - there are many recent models for this task with nice results! ours has the benefit of using a frozen encoder + very few learnable params. however for on-device applications, the ViT encoder itself may pose challenges compared to smaller conv models
1
0
1
@fionakryan
Fiona Ryan
2 months
RT @TelepathicPug: @fionakryan This is great. I'm having too much fun.
Tweet media one
0
1
0
@fionakryan
Fiona Ryan
2 months
@RigoTech1 @sangminlee777 @judyfhoffman @RehgJim that's my hypothesis, along with the results being quite close to the inter-rater agreement of the annotations. sweeping LoRA hyperparams (like which params to tune) could potentially get more gains. our main focus was on using the encoders out of the box.
1
0
0
@fionakryan
Fiona Ryan
2 months
@tom_doerr I didn't, but there are several works on webcam-based eye tracking that are designed specifically to predict the coordinate on the screen the viewer is looking at!
2
0
27
@fionakryan
Fiona Ryan
2 months
@vikhyatk thanks for sharing!
0
0
2
@fionakryan
Fiona Ryan
2 months
Big thank you to the Gaze-LLE team: Ajay Bati, @sangminlee777, Daniel Bolya, @judyfhoffman , @RehgJim for working on this project with me!
2
2
33