flat Profile Banner
Stephen Panaro Profile
Stephen Panaro

@flat

Followers
503
Following
1K
Statuses
819

making coffee and other things. @BrewTimerApp

boston
Joined May 2013
Don't wanna be here? Send us removal request.
@flat
Stephen Panaro
9 years
“We won’t run it in digital because we’re purists and maniacs.”
2
0
5
@flat
Stephen Panaro
5 days
@MaxWinebach sent 🍿
0
0
0
@flat
Stephen Panaro
5 days
Despite the general messiness, it wrote one line of code in a different and simpler way than my solution. Neat.
0
0
0
@flat
Stephen Panaro
10 days
@fleetwood___ Oh yeah. If this looks good for normal (scale/bias) quants, I am hoping it helps non-uniform/LUT methods too.
0
0
1
@flat
Stephen Panaro
13 days
Possible silver lining: code could be a jumpstart if you want to try quantizing ModernBERT with SpinQuant.
0
0
0
@flat
Stephen Panaro
14 days
@ivanchanavinah Dope! Text extraction is pretty good for the few I tried but formulas could use some work.
0
0
1
@flat
Stephen Panaro
14 days
@dessatel Haha if tool == “LLM wrapper” I’m ok with that
0
0
1
@flat
Stephen Panaro
16 days
Tried a lot of prompt variations with o1-mini/preview so I do think it is just beyond them.
0
0
2
@flat
Stephen Panaro
22 days
Might test this out. Seems low effort/fairly interesting. GPTQ, AWQ, llama.cpp, maybe HQQ. Anything others?
@flat
Stephen Panaro
26 days
Seems like QuaRot/SpinQuant should be a prerequisite if you’re quantizing LLMs in any way. Having smaller weights/activations can only help. Right?
0
0
3
@flat
Stephen Panaro
25 days
Read about it here:
1
0
8
@flat
Stephen Panaro
26 days
@dessatel Yeah! So many cool things in the paper. Lossless 8-bit RTN too.
0
0
1
@flat
Stephen Panaro
26 days
SpinQuant: QuaRot:
0
0
0
@flat
Stephen Panaro
1 month
@CulStory @answerdotai Awesome. after the holidays I’ll take a look and see if I can get rotate working again.
0
0
0