flat Profile Banner
Stephen Panaro Profile
Stephen Panaro

@flat

Followers
507
Following
1K
Statuses
820

making coffee and other things. @BrewTimerApp

boston
Joined May 2013
Don't wanna be here? Send us removal request.
@flat
Stephen Panaro
9 years
“We won’t run it in digital because we’re purists and maniacs.”
2
0
5
@flat
Stephen Panaro
1 day
@dessatel @handleym99 Those two papers have definitely made me discount perplexity more. Most semi-recent quant papers I’ve read show a variety of benchmarks which I think is better signal. This is also new and interesting, different impact for context vs weight knowledge:
0
0
2
@flat
Stephen Panaro
9 days
@MaxWinebach sent 🍿
0
0
0
@flat
Stephen Panaro
9 days
Despite the general messiness, it wrote one line of code in a different and simpler way than my solution. Neat.
0
0
0
@flat
Stephen Panaro
15 days
@fleetwood___ Oh yeah. If this looks good for normal (scale/bias) quants, I am hoping it helps non-uniform/LUT methods too.
0
0
1
@flat
Stephen Panaro
17 days
Possible silver lining: code could be a jumpstart if you want to try quantizing ModernBERT with SpinQuant.
0
0
0
@flat
Stephen Panaro
18 days
@ivanchanavinah Dope! Text extraction is pretty good for the few I tried but formulas could use some work.
0
0
1
@flat
Stephen Panaro
19 days
@dessatel Haha if tool == “LLM wrapper” I’m ok with that
0
0
1
@flat
Stephen Panaro
20 days
Tried a lot of prompt variations with o1-mini/preview so I do think it is just beyond them.
0
0
2
@flat
Stephen Panaro
27 days
Might test this out. Seems low effort/fairly interesting. GPTQ, AWQ, llama.cpp, maybe HQQ. Anything others?
@flat
Stephen Panaro
1 month
Seems like QuaRot/SpinQuant should be a prerequisite if you’re quantizing LLMs in any way. Having smaller weights/activations can only help. Right?
0
0
3
@flat
Stephen Panaro
30 days
Read about it here:
1
0
8
@flat
Stephen Panaro
1 month
@dessatel Yeah! So many cool things in the paper. Lossless 8-bit RTN too.
0
0
1
@flat
Stephen Panaro
1 month
SpinQuant: QuaRot:
0
0
0