Tejas Vaidhya Profile
Tejas Vaidhya

@imtejas13

Followers
356
Following
3K
Statuses
563

Building @NolanoOrg | Curious about everything | MSc student @Mila_Quebec. Previously: @iitkgp, @ETH_en, @ZFellows_

Montréal, Québec
Joined April 2020
Don't wanna be here? Send us removal request.
@imtejas13
Tejas Vaidhya
18 days
@ArnabMondal96 @_AyushKaushal @irinarish @iclr_conf - Most of those reviews are vague and uninformative. Additionally, some of them request unreasonable training of the 4B+ model until convergence, which will not happen until 11 trillion tokens.
0
0
0
@imtejas13
Tejas Vaidhya
1 month
@willccbb Go for it.
0
0
1
@imtejas13
Tejas Vaidhya
2 months
RT @benjamintherien: Learned optimizers can’t generalize to large unseen tasks…. Until now! Excited to present μLO: Compute-Efficient Meta-…
0
33
0
@imtejas13
Tejas Vaidhya
2 months
Lesson learned: When you make a mistake, reading the policy can always reveal a way to fix it.
0
0
0
@imtejas13
Tejas Vaidhya
2 months
RT @irinarish: @BensenHsu @sama Indeed, memory is becoming a bottleneck, but there is lots of work on quantization/compression and also tra…
0
2
0
@imtejas13
Tejas Vaidhya
3 months
@ArnabMondal96 @iclr_conf Isn't it truly beautiful?
0
0
1
@imtejas13
Tejas Vaidhya
4 months
@srush_nlp Thanks
0
0
1
@imtejas13
Tejas Vaidhya
4 months
RT @ArnabMondal96: Looking for PhD interns for Summer 2025 with strong publication records. Experience with video-language foundation model…
0
69
0
@imtejas13
Tejas Vaidhya
5 months
1
0
1
@imtejas13
Tejas Vaidhya
5 months
There is more to it. The scaling curve you forgot to talk about. I will be archiving the full report soon.
@sdianahu
Diana
5 months
Deepsilicon runs neural nets with 5x less RAM and ~20x faster. They are building SW and custom silicon for it. What’s interesting is that they have proved it with SW, and you can even try it. On why we funded them 1/7
0
0
4
@imtejas13
Tejas Vaidhya
5 months
@ArnabMondal96 What about Anil Kapoor?
0
0
0
@imtejas13
Tejas Vaidhya
5 months
RT @GCResearchTeam: Spectra by @NolanoOrg is an open suite of 54 LLMs spanning FP16 training, ternary training, and post-training quantisat…
0
2
0
@imtejas13
Tejas Vaidhya
6 months
@arna_ghosh @irinarish The missing tools in the toolkit of modern deep learning.
0
0
1
@imtejas13
Tejas Vaidhya
6 months
@NolanoOrg
Nolano.ai
7 months
🚀 SpectraSuite of Ternary and FP16 LLMs 🚀 We’re thrilled to release the Spectra Suite of open ternary (TriLMs) and FP16 (FloatLMs) language models from 99M to 3.9B parameters. At billion+ parameter scale, TriLMs upto 10x smaller can match the performance of FloatLMs. 1/5
Tweet media one
0
0
1
@imtejas13
Tejas Vaidhya
6 months
@paul_cal We are planning to release a bigger model soon.
0
0
1
@imtejas13
Tejas Vaidhya
6 months
@OrniasDMF @irinarish Scaling in progress.
1
0
1
@imtejas13
Tejas Vaidhya
6 months
RT @BlancheMinerva: Very cool paper that shows impressive performance with ternary LLMs. Discovering new papers that use @AiEleuther's GPT-…
0
3
0