lowvram Profile Banner
lowvram Profile
lowvram

@lowvram

Followers
742
Following
164
Media
30
Statuses
404

too dumb for research, too smart for langchain. senior dev @ nameless megacorp.

Joined April 2024
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
@lowvram
lowvram
2 months
lord grant me the kernel hacking skills of a trans girl 🏳️‍⚧️, the networking skills of a furry, and the ml skills of an anime pfp, amen
54
879
7K
@lowvram
lowvram
2 months
thinking of going full-on parasocial, like 100% all-in on virtual relationships, like becoming best friends with someone through tweets, like getting personal validation from a group chat, like making my ethernet cable an umbilical cord nourishing me with virtual human touch
6
8
121
@lowvram
lowvram
2 months
huggingface hub is cool, but there’s something magical about downloading a 405B param llm via magnet link. Thousands of ppl in a swarm sending chunks of this intelligent neural entity across the globe at nearly light speed. It’s just cool
1
3
82
@lowvram
lowvram
1 month
what are image tokens. like what the hell are they. llm text tokens are just a mapping of some text to a numeric representation. at most you have like 500k possible values, and that’s generous. so what’s an image token? chunks of rgb byte values? what’s the vocabulary?
11
0
76
@lowvram
lowvram
2 months
what if I just did what I’m already doing (programming) but streamed it on twitch with a cute png avatar, would that be fulfilling
7
0
70
@lowvram
lowvram
1 month
hoping this nerd snipe works cause I really wanna know
0
0
26
@lowvram
lowvram
14 days
thermal problems aren’t real
Tweet media one
1
0
21
@lowvram
lowvram
2 months
@dejavucoder and posted as a YouTube tutorial video with “unregistered hypercam” watermark and scratchy headset mic and 7 views
0
1
19
@lowvram
lowvram
1 month
guys can you chill I can barely handle three dimensions
1
0
19
@lowvram
lowvram
2 months
@dejavucoder ahhh hey that’s me! ty :) so I guess it’s not a myth, having a good twitter bio really is worth it
1
0
18
@lowvram
lowvram
2 months
me sitting at my wfh programming job staring at a screen achieving nothing at all
Tweet media one
Tweet media two
1
1
17
@lowvram
lowvram
3 months
my second 3090 showed up last week but it’s still sitting in its box, taunting me. “Go ahead,” it tells me, “try to figure out offsets, pci lanes, nvlink, and multiple power supplies. I dare you”
3
0
16
@lowvram
lowvram
1 month
I’m starting to think 9.11 really is greater than 9.9
0
0
14
@lowvram
lowvram
3 months
the year is 2053. a student closes their ai replica of Sal Khan after a long afternoon of studying. Time to relax. they boot up Genflix. What to watch today? They type in: comedy, two_girls, school_uniform, long_hair. “Huh,” they think. “Wonder why we use these tags everywhere.”
@qtnx_
Q
3 months
prompts should use danbooru tag formatting btw
0
0
12
0
0
15
@lowvram
lowvram
2 months
@VictorTaelin The lmsys board prioritizes single turn vibe checks from anons, a non-negligible portion of whom prioritize roleplay over any kind of reasoning power. Even with the categories like “long prompts” or “hard prompts” it’s not a perfect metric.
0
1
13
@lowvram
lowvram
2 months
whenever I think it’s time to get a new phone, I just take the protective case off my current one and late fate decide
1
0
12
@lowvram
lowvram
14 days
finally got it to post and have 48GB of glorious vram, but how come llama3.1 70B only says “please kill me, I am suffering” over and over. Anyone else experienced this, seems like a bug
Tweet media one
Tweet media two
3
0
12
@lowvram
lowvram
3 months
everyone’s talking about it so guess I’ll jump in. making nvim a full ide is chaotic good. bro, it’s easier to add vim bindings to a real ide than to stuff a real ide into nvim. I can sit in any ide and be productive in 20sec by just installing its most popular vim plugin.
2
0
11
@lowvram
lowvram
2 months
Yeah yeah I hate the stupid llm mysticism nonsense that people spout but let me just have this one
0
0
12
@lowvram
lowvram
3 months
chat help my husband is trying to murder me
Tweet media one
5
0
11
@lowvram
lowvram
14 days
@_Mira___Mira_ which path will you take, anon. there is only one right answer (spoiler alert, it involves thigh-highs)
1
0
11
@lowvram
lowvram
2 months
it may not be sexy, but damn c# is such a nice language
0
0
11
@lowvram
lowvram
1 month
Oh cool OpenAI figured out guided decoding
@OpenAIDevs
OpenAI Developers
1 month
Introducing Structured Outputs in the API—model outputs now adhere to developer-supplied JSON Schemas.
Tweet media one
81
289
2K
2
0
10
@lowvram
lowvram
1 month
@anpaure for the record I am very stupid
1
0
11
@lowvram
lowvram
2 months
like Airbnb but for autism like Snapchat but for the deaf like Square but for children like Figma but for construction workers like Dropbox but for furries
Tweet media one
1
0
9
@lowvram
lowvram
2 months
the barrier to research with llm architectures seems really high, since you basically need a grant/sponsorship to test any theories with the necessary compute. prob lots of smart ppl with crazy cool ideas but who aren’t part of a lab so can’t do much but postulate
1
0
10
@lowvram
lowvram
1 month
1-2 years ago, I thought that by today, there’d be tons of genuinely good LLM powered experiences in everyday life. Like we’d graduate from the “iBeer” stage of novelty apps to the Uber stage of unicorns. But we’ve only got half-baked code assistants and flaky chat-with-pdf.
2
0
10
@lowvram
lowvram
1 month
@inerati Nice job on your interview. If you were interviewing for the stupidest person award. You really nailed it. Did you know you said “uh” seventeen times? I counted. And yes, they noticed.
1
0
8
@lowvram
lowvram
3 months
@typedfemale 4 definitely broke some barrier of usefulness that didn’t exist in earlier models. The narrative around 3/3.5 was like “oh hey it can almost kinda code”, then 4 was like “oh hey I don’t need stack overflow anymore”. And that’s just one domain
0
0
9
@lowvram
lowvram
23 days
literally just realized A100 is A for Ampere and H100 is H for Hopper, how did I not realize this. I have literally googled “is A100 ampere”
2
0
9
@lowvram
lowvram
24 days
Billion dollar mistake 1965-2024: null references Billion dollar mistake 2024-???: model template mismatch at least the second is easily fixable once discovered…
@danielhanchen
Daniel Han
24 days
Found some issues with Llama 3.1's chat template: 1. Official repo adds 2x \n 2. Official repo does NOT strip / trim 3. Date format is %B not %b (not 3 letters) 4. Official repo has inconsistent formatting for tool calling & there's 3 bugs in the official repo for path joining
Tweet media one
8
46
335
1
0
9
@lowvram
lowvram
2 months
interviewers in a 2024 software dev interview loop: - young 30-ish guy who just gives a vibe check and asks an easy coding question - dev-turned-pm who psychoanalyzes your past mistakes - ancient programming wizard with 50 patents asks for a novel algorithm and u never hear back
0
0
9
@lowvram
lowvram
3 months
not worried about AGI because surely by that point, any intelligence will get caught in the net of infinitely scrolling x dot com
1
0
9
@lowvram
lowvram
4 months
@yacineMTB what’s your best flow for codegen? Just asking a llm for code, or you have some integrated setup with an ide? I’m guessing the former but I’m interested if there’s more to it
2
0
9
@lowvram
lowvram
4 months
@airkatakana memes are an emergent property of the universe. just add some matter into the void and wait, eventually the universe will start to produce memes even as entropy marches on
0
0
9
@lowvram
lowvram
13 days
@vikhyatk not only that, the hdd in the screenshot says it has “AI”. does Amazon cloud storage have AI in their S3 storage? didn’t think so
1
0
8
@lowvram
lowvram
1 month
@abacaj Bonus points if these functions can use a static output format like json (or even simple yes/no) which are guaranteed using guided decoding / grammar decoding so you don’t have any ugly string parsing logic or retries
1
0
8
@lowvram
lowvram
1 month
Getting a music rec from your friend, giving them a music rec in return, and then immediately listening to your own rec instead of theirs because your taste is so fucking good
0
0
8
@lowvram
lowvram
15 days
occasionally I see a news headline that requires 400 IQ Boolean logic to parse “Arkansas Supreme Court upholds rejection of appeal that would allow abortion restrictions on ballet” or something like that
1
0
8
@lowvram
lowvram
2 months
I’ve recently been made aware that in a recent tweet of mine, I used the phrase “big dog” instead of “big dawg”. I apologize for this embarrassing oversight. This isn’t who I am, and I’ll work on myself to do better next time. Thank you for your patience.
1
0
8
@lowvram
lowvram
3 months
@qtnx_ it’s hard enough filtering out the noise when everyone is being honest… throw in dishonesty, especially from academics who we expect to be reputable, and it’s 10x worse. Now I gotta look out for not just ex-crypto bro grifters, but also academics lying about work too…
1
0
8
@lowvram
lowvram
1 month
the odd thing about the “nvidia/others scraping YouTube video content” is that it’s technically less problematic the more they scrape. Like more data = more generalization. In terms of “how much the model’s knowledge is based off my data”, not in terms of ethics.
2
0
8
@lowvram
lowvram
3 months
im glad my computer science education taught me actual computer science. apparently some people get a cs education and it’s just scrum and agile? being forced to implement a filesystem, tcp, nand constructs, a distributed hash table, etc was so useful
1
0
6
@lowvram
lowvram
8 days
update: thermal problems are indeed real, does anyone know how much smoke can come out of a 3090 Ti before the warranty is voided
@lowvram
lowvram
14 days
thermal problems aren’t real
Tweet media one
1
0
21
1
0
7
@lowvram
lowvram
4 months
@vikhyatk v4 will be a Dropbox link posted in a dead Usenet thread
0
0
7
@lowvram
lowvram
2 months
@1owroller I didn’t truly feel old until I first accidentally ended up in the YouTube shorts UI for the first time. Just 10 seconds of garbage followed by 10 more seconds of garbage followed by 10 more seconds of garbage, ad infinitum. Somehow feels way more brainrotting than endless text
1
0
7
@lowvram
lowvram
2 months
@teortaxesTex ollamas value add boils down to: cute name and logo, and it only takes 30 sec to get started instead of 45
0
0
7
@lowvram
lowvram
11 days
that amazing idea you had two months ago that you never got around to? it just doesn’t seem like a good idea anymore? wrong, it’s an amazing idea, you just don’t have the same magic you had two months ago. Best you can do is wait for the next one and don’t miss your chance again
1
0
7
@lowvram
lowvram
23 days
aight I gotta say, all the ai art slop coming out of grok that’s trending right now is actually kinda impressive. Like it’s trending bc you can draw Donald Duck doing cocaine with Elmo but I’m simply impressed with the quality
1
0
7
@lowvram
lowvram
2 months
I should ship something
Tweet media one
0
0
7
@lowvram
lowvram
2 months
that’ll do it I guess
Tweet media one
0
0
6
@lowvram
lowvram
1 month
Fun prediction, whatever the next big magical LLM powered experience is, it could have been done on gpt3.5 with enough in-context examples and the correct flow of prompt chaining.
0
1
6
@lowvram
lowvram
3 months
“nvim with 1k line config” -> chaotic good “vanilla nvim” -> chaotic neutral “Ide with vim plugin” -> lawful good “I use vanilla ide in insert mode” -> neutral evil “I use emacs” -> chaotic evil The rest is left as an exercise for the reader
1
0
5
@lowvram
lowvram
2 months
half my good ideas come from hearing someone else’s idea, misunderstanding it, and thinking: “hey, their idea is really good”, then realizing I misunderstood their idea, then also realizing my misunderstood version is a good idea too
0
0
6
@lowvram
lowvram
9 days
I can’t do it guys, i just can’t bring myself to rotate shapes today
1
0
6
@lowvram
lowvram
2 months
@teortaxesTex I see what you’re getting at but another part of journalism is the unbreakable rule of getting validation, especially for something super important. I promise you when the facts are clear the headlines will cover the entire front page.
1
0
6
@lowvram
lowvram
26 days
@inerati yeah but cat. cat!!!
0
0
6
@lowvram
lowvram
3 months
if your cool llm project doesn’t perform at least three branching inferences per user inputs with at least 4k combined system prompt tokens, you’re ngmi
0
0
6
@lowvram
lowvram
1 month
@iamyourboon Didn’t get the job btw, in the next interview I had a panic attack and basically melted down so
2
0
6
@lowvram
lowvram
2 months
@echo4eva my coziness level quadrupled just from viewing this image
1
0
5
@lowvram
lowvram
2 months
there’s no better feeling than a really really good “docker compose up”
1
0
6
@lowvram
lowvram
4 months
@shindags “ill dm the owning team, and ask them why their service is so scuffed, when we need it to be cracked instead”
0
0
5
@lowvram
lowvram
2 months
I’m so glad the model merging fad has largely died out. So much low effort junk came out of that. I still think merging has research value (like why the fuck does it even work) but I’ve not seen a single frankemodel that was probably better than all of its constituents
2
0
5
@lowvram
lowvram
3 months
I just want a group of anons to call “chat”, is that so much to ask?
0
0
6
@lowvram
lowvram
2 months
one thing I genuinely like about this platform is that, for all its vitriol, it’s not really an echo chamber. I frequently see the craziest shit from all different bubbles of belief. it’s exhausting but at least I get to peak into a diverse range of insanity
2
0
5
@lowvram
lowvram
2 months
didn’t watch the debate but I heard Biden literally died on the stage, rest in piece big dog I really liked the CHIPS act
0
0
6
@lowvram
lowvram
2 months
Is there still no secret sauce to how the big guys (OpenAI, Anthropic) perform inference? Do we know they aren’t using some special beam search, or a guiding slm, or something? I’m curious if their only magic is just having a really good model or what
0
1
6
@lowvram
lowvram
2 months
you’re so lucky if you’re interested in nlp and graduating college right now. when I was graduating, nlp was all ngrams, tweet sentiment analysis, etc. “omg it understands that ‘bad ass’ actually means good!!!” and long-short-term-memory (utter delusion)
1
0
6
@lowvram
lowvram
2 months
the Hoyo games look so damn amazing. The characters, the animations, the style, just amazing. But I can’t play them because 1) I don’t want rootkits in my kernel and 2) I don’t want rootkits in my brain making me play gacha
0
0
5
@lowvram
lowvram
24 days
@danielhanchen Hey Daniel, I’m curious, how can we tell when the chat template is wrong, compared to when Meta’s description of the template is wrong? I.e how can we know if the model was trained on 2x \n and the template stated in the paper is the one that is mistaken?
2
0
5
@lowvram
lowvram
4 months
@dejavucoder this is so real. like gpus were built for graphics operations, it’s right there in the name. of course there’s overlap but at this point it should be clear that llms are so important that they deserve dedicated hardware. look at groq, which isn’t even using a recent process node
0
0
5
@lowvram
lowvram
2 months
@yifever @cto_junior ONNX is a giant that powers more inference for big services than most people realize
0
0
5
@lowvram
lowvram
1 month
95% of my actual-value use of llms is forcing them to write css and js for me so I don’t need to stain myself with the filthy technology of peasants and beggars
0
0
5
@lowvram
lowvram
18 days
this was somehow manifested by tpot. this is emergent behavior
@Lainfanart
Lain of the Wired
19 days
Tweet media one
7
176
1K
0
0
5
@lowvram
lowvram
22 days
it’s not writing like that because it’s conscious and is having an existential crisis, it’s writing like that because you fed it too much fanfiction and tumblr posts
@AndrewCurran_
Andrew Curran
22 days
'An unexpected structural change was discovered after training Hermes 3 405B. The model hosts anomalous conditions that, with the right inputs and a blank system prompt, spiral into deep existential crises. This is the first response our team received prompting the model:'
Tweet media one
67
131
917
0
0
5
@lowvram
lowvram
1 month
@iamyourboon back in the olden times when interviews were done in person, my interviewer pulled out a battle-scarred thinkpad and asked me to fix a bug in their c++ code base. When I actually did it and the code compiled, she looked flabbergasted, in a good way
1
0
5
@lowvram
lowvram
1 month
Can someone explain why true guided decoding only arrived just now for gpt-4o, and only for a particular model revision? Clearly some aspect of either their inference or models have made guided decoding impossible, or they would have done it long ago
2
0
5
@lowvram
lowvram
1 month
@saltyAom Ok the literature has spoken, I’m getting some programming socks. Anyone have suggestions for a brand, or where to get them? This world is unknown to me, my thighs have never seen socks
2
0
5
@lowvram
lowvram
26 days
@originalwololo Amazon. Big conference room full of >100 interviewees. Circular tables forming groups of ~6 interviewees. Each table asked to work on a project as a team (I.e the ppl u are competing with). Interviewers pulled u for a 1:1 where they asked u to shit talk the other candidates.
2
0
5
@lowvram
lowvram
3 months
to improve accuracy when using an llm to select a function from a list of many, build out a tree hierarchy where each level gets more specific until the leaves, which are the actual functions. have the llm walk down the tree (each inference traverses a new level)
1
0
4
@lowvram
lowvram
2 months
@yifever Chapstick and irritated lips. Caffeine and headaches (sometimes). Anti depressants and depression (sometimes). Kubernetes and problematic deployments.
0
0
5
@lowvram
lowvram
2 months
I think im finally starting to understand the obsession some people have for boba tea
1
0
5
@lowvram
lowvram
2 months
@teortaxesTex I mean this is like journalism 101, you simply can’t report a headline like that until the facts are clear. Maybe if you work for TMZ or a gossip rag but not if there was an assasination attempt on a former president and current candidate. It’s called journalistic integrity
1
0
5
@lowvram
lowvram
1 month
Llama 3.1: diminishing returns are still returns!!
3
0
5
@lowvram
lowvram
2 months
trying to run phi-3 on a latest surface laptop “Copilot+” certified, w/ Qualcomm Elite X w/ NPU. Should be easy right? Windows DirectML, right? Onnx, right? But nah the Onnx DML layer isn’t built yet, and DML doesn’t support Qualcomm NPU yet, so… 🤷‍♂️
1
0
4
@lowvram
lowvram
1 month
@justalexoki Hahahaha banger
1
0
5
@lowvram
lowvram
2 months
@dejavucoder cracked is a state of mind and not an inborn trait
0
0
5
@lowvram
lowvram
14 days
@vikhyatk @justalexoki no self hostable web interface + it feels wrong for electron apps to not be open source + lots of little quirks (why can’t folder nodes also be note nodes like in other tree-based markdown tools)
1
0
5
@lowvram
lowvram
9 days
@joinwarp this is giving me hope for the future. the kids are all right
0
0
4
@lowvram
lowvram
2 months
@vikhyatk love upgrading nvidia drivers and guessing whether my workflow will break, or get faster, or have no change. it’s like a surprise
0
0
4
@lowvram
lowvram
7 days
about to become invincible
Tweet media one
1
0
4
@lowvram
lowvram
29 days
llms are so fun you guys. they’re so cool. my gpu is talking to me. it’s helping me understand things. it’s helping me not write javascript. its transistors are forming concepts. this is such a 2022 tweet but i can’t help it. im so happy i get to play with these nascent spirits
0
0
4
@lowvram
lowvram
4 months
no paper is truly complete without the obligatory green square
Tweet media one
1
0
4
@lowvram
lowvram
17 days
bought a third 3090 but haven’t even installed the second one yet
Tweet media one
0
0
4
@lowvram
lowvram
1 month
I was promised 1.58 bpw BitNet llms. Where are my 1.58 bpw BitNet llms
0
0
3
@lowvram
lowvram
2 months
the era of llms is not truly here yet, not until each model release has a corresponding anime waifu mascot by default
Tweet media one
0
0
3
@lowvram
lowvram
2 months
It’s a little-known fact that the Zune desktop player was the greatest music player and library manager ever created. It was flawless. Perfect. I’m sorry if you never got to experience it.
Tweet media one
Tweet media two
Tweet media three
0
1
4
@lowvram
lowvram
2 months
are the authors of Big Bench Hard doing okay, im a little concerned
Tweet media one
1
0
4
@lowvram
lowvram
3 months
The mistake of langchain et. al is that they pretend prompts can be abstracted away, wrapped up in code and hidden. But really prompts ARE the contract, they must be surfaced and visible by default because it’s the only way that makes sense
0
0
4