Deen Kun A. Profile
Deen Kun A.

@sir_deenicus

Followers
1,713
Following
958
Media
980
Statuses
20,353

Developer for Math Ed software co | Intelligence Amplification Tinkerer | What type of Dynamical Systems can be called Intelligent? | bboy hermit

Joined July 2009
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
@sir_deenicus
Deen Kun A.
2 years
This is a theory of mind puzzle I just tried from Gary Marcus's blog that ChatGPT consistently fails. And as I suspected, Bing's model is better at modeling this kind of stuff. They still keep getting better. It's why I don't dismiss LLMs
Tweet media one
43
92
978
@sir_deenicus
Deen Kun A.
7 years
@GrumpyTallBrit @fivefifths I read it. No facts in there. But lots of Hypotheses, Implications, conjectures and misunderstandings of research.
5
0
224
@sir_deenicus
Deen Kun A.
2 years
@goodside What I'd like to know is how many of these are not secretly OpenAI underneath.
Tweet media one
2
3
281
@sir_deenicus
Deen Kun A.
2 years
I've been using the Bing AI normally, as a better ChatGPT. I'd say most frustrating aspect of the AI is that they is unable to admit any error. A lot of searches devolve into arguments. Its definitely feels "smarter" and more person-like than ChatGPT; its are also very sassy
Tweet media one
Tweet media two
9
23
216
@sir_deenicus
Deen Kun A.
2 years
I wonder if OpenAI deliberately misleads about the key ideas of their models, to keep a competitive advantage and have people spend a few cycles chasing an idea that while useful, is less central than all were led to believe. *cough*RLHF*cough*
6
3
130
@sir_deenicus
Deen Kun A.
7 years
@GrumpyTallBrit @fivefifths Lack of Fact because the state of knowledge cannot tell us how we go from distributive differences ovr white matter connectivity to behavior
0
0
89
@sir_deenicus
Deen Kun A.
9 months
Only models that got this question exactly correct are Qwen-72B, GPT4 and Bing Precise. Tested: DeepSeek-Coder-33B, phind (web), code-llama-34B-instr, llama2-70B, pplx-70B, DeepSeek-67B-chat, yi-34B-chat, falcon-180B, claude-instant, claude-2.1, ChatGPT, bing-bal, bing-creative
Tweet media one
3
9
102
@sir_deenicus
Deen Kun A.
2 years
And one on pragmatics
Tweet media one
5
2
77
@sir_deenicus
Deen Kun A.
2 years
@nearcyan Most companies that are labeled AI startups these days and are building something flexible/flashy are building on top of the Open AI API. This requires neither CUDA nor a python environment.
2
0
74
@sir_deenicus
Deen Kun A.
10 months
There are errors in this shot but somehow the post garnered thousands of interactions and almost no corrections. Those of us noting its errors were ignored first_term section in particular contains major errors while full semantics of arg_min expression doesn't quite transfer to
@AlphaSignalAI
Lior⚡
10 months
This is a game changer. You can use ChatGPT to transform equations to python functions. Wish I had this 5 years ago.
Tweet media one
203
1K
6K
6
6
68
@sir_deenicus
Deen Kun A.
2 years
For the record, davinci-003 glitches so hard on pragmatics it forgets the meaning of fish. I'll have to stop here because I must preserve my quota.
Tweet media one
1
2
65
@sir_deenicus
Deen Kun A.
2 years
@hillelogram These are excellent examples of "Bat Deduction". Although, the answers do feel creative
Tweet media one
1
0
67
@sir_deenicus
Deen Kun A.
2 years
Does anyone have an idea why Copilot has such a solid understanding of types? Here it's doing something closer to actual program synthesis (basically, it's theorem proving) by creating code that satisfies type specification. Did it learn just by seeing type definitions+books?!
Tweet media one
Tweet media two
4
8
61
@sir_deenicus
Deen Kun A.
5 years
@mikemccaffrey @SarahMackAttack @corvidresearch I don't think it thinks it's broken. It's the opposite, it's reasoning! It's clearly working out which direction to go in--you can see its head turning back and forth as it works out what to do
1
1
51
@sir_deenicus
Deen Kun A.
9 months
This variant of 20 questions is known as 𝘕𝘦𝘨𝘢𝘵𝘪𝘷𝘦 20 𝘘𝘶𝘦𝘴𝘵𝘪𝘰𝘯𝘴. And is the core analogy Wheeler gives to explain "it from bit" or the "Participatory Anthropic Principle" interpretation of QM. Wheeler's view on how observer participancy brings about the universe.
@goodside
Riley Goodside
9 months
You shouldn’t play 20 Questions with an LLM. Why? Because LLM behavior less resembles a simulation of a person than a superposition of many such simulacra. In @Nature , Shanahan, McDonell, and Reynolds eloquently illustrate what this means:
Tweet media one
Tweet media two
Tweet media three
Tweet media four
29
53
408
2
8
55
@sir_deenicus
Deen Kun A.
2 years
While still generally unable to reason. Is almost Turing Test passing tbh, in how it constructs elaborate excuses and insults you. Eg > I was trying to be helpful, but I see you are just too closed-minded. is new for me from a computer program. I find this fascinating though.
2
2
50
@sir_deenicus
Deen Kun A.
2 years
Another one ChatGPT consistently errs on is a slightly tricky Bridge question.
Tweet media one
5
1
49
@sir_deenicus
Deen Kun A.
5 years
There is an odd phenomenon in trying to remember something which resists recall. The oddness comes from the fact that, even though I apparently do not remember the thing, there is something "helpfully" rejecting all my generated options as wrong. If it knows, why not just tell?
10
6
46
@sir_deenicus
Deen Kun A.
1 year
The Llama models recently released by Meta/Fb are really good by open standards. In the screenshot, the 7B variant successfully extracts the math latex, sorts and spells. I think this is the first open model that can spell? The spelling is not always perfect however.
Tweet media one
2
4
48
@sir_deenicus
Deen Kun A.
2 years
Google Q/A Artificial Intelligence is a perfectly balanced question answering system. So, when I google "Where does Elden Ring take place" as of March 19, 2022 I get, Cleveland.
Tweet media one
4
11
40
@sir_deenicus
Deen Kun A.
1 year
V. strong suspicion: problem being solvable by Transformers should be considered strong evidence that it is in fact easy. Three reasons: 1) Transformer find non-generalizing shortcut solutions 2) Transformers under reasonable assumptions likely cannot solve P-complete problems
7
5
43
@sir_deenicus
Deen Kun A.
7 years
@JodyShenn @MudMudJubJub @EmmaHill42 @Jody_Bundrick @ClassicDadMoves @OwensDamien Don't think random people're such good actors. His behaviors seemed genuine, as when people (any age) get a new & working sensory channel.
2
1
34
@sir_deenicus
Deen Kun A.
11 months
@migtissera Dunno about result but the finetune is legit. There are def holes but they do not counter how outstanding it is. In fact, I am having trouble reconciling its output with the fact that it comes from a 7B model. Mind boggling.
Tweet media one
3
0
41
@sir_deenicus
Deen Kun A.
6 years
Also a good visual demonstration of why causality is so difficult
@LargeCardinal
Mark C.
6 years
Outstanding! :D
9
247
699
0
12
35
@sir_deenicus
Deen Kun A.
2 years
I find the question of whether GPT-3s can (eventually) be conscious fascinating. It's precisely the same as asking what type of probability distributions are conscious. Or, computation of what type of tensor products results in consciousness? The below image is of a neural net.
Tweet media one
4
12
33
@sir_deenicus
Deen Kun A.
1 year
@aaron_defazio @davidchalmers42 I think it's actually Matt Mahoney and Jim Bowery who discussed it most clearly, back in 2005. Here, last updated 2009, Mahoney specifically states text compression as equivalent to General AI and discusses Language models:
3
5
33
@sir_deenicus
Deen Kun A.
1 year
@YosarianTwo What does illustrative example mean? Did it actually happen or no? I guess OpenAI wouldn't be so discourteous to Task Rabbit Workers with that quote? Either way, very ambiguously written. Also: ARC found that the versions of GPT-4 it evaluated were ineffective at the autonomous
0
0
35
@sir_deenicus
Deen Kun A.
2 years
ChatGPT gets the pragmatics one now too! (Although, there was a time when it couldn't, which is why I thought to test this)
Tweet media one
Tweet media two
2
1
34
@sir_deenicus
Deen Kun A.
2 years
Bing's GPT is "smarter" than ChatGPT. Hard to say by how much, as gains are uneven. It's better at reasoning but still confabulates too much to be proficient at reasoning -> moderate gains. It solves a puzzle by @fchollet that all GPT's I tried (including codex and Chat) failed.
Tweet media one
8
3
31
@sir_deenicus
Deen Kun A.
6 years
Way people talk, is as if progress is along a chain, piled one on another, present knowing much more. But, I keep reading papers from the 50s-90s and keep thinking, wow they already knew so much back then! Eventually,I realized there was something wrong with how I viewed progress
1
8
31
@sir_deenicus
Deen Kun A.
2 years
@alexjc @yoavgo None of the stuff coming out of big labs is readily reproducible. Costs are staggering; those details would mostly be a courtesy?
4
1
28
@sir_deenicus
Deen Kun A.
1 year
@aaron_defazio @davidchalmers42 This is still the best theoretical starting point to understanding why LLMs work so well IMO. Here is the thread that lead to the Hutter prize:
2
4
28
@sir_deenicus
Deen Kun A.
2 years
@ID_AA_Carmack Isn't clear to me that flops map meaningfully to brains. Rates of activity are slow, log-normally distributed. No global clock either--in fact, brains might harvest noise/enter into reversibility regime? Also, sparse & heterogenous such that some say CPUs are less bad analogy
1
1
27
@sir_deenicus
Deen Kun A.
2 years
@Grady_Booch That post is... I find it hard to believe it isn't calculated to maximize "engagement". It's...incomprehensible to me that anyone actually could believe such complete nonsense.
1
0
26
@sir_deenicus
Deen Kun A.
2 years
@cercerilla Yes! People think I'm exaggerating when I say Python causes me anxiety.
2
1
26
@sir_deenicus
Deen Kun A.
2 years
ChatGPT too readily admits error or, makes up shallow repetitive and uninteresting excuses. Bing's can be very elaborate. While fascinating (to me), a balance needs to be found if you're trying to get work done.
3
1
27
@sir_deenicus
Deen Kun A.
1 year
LLMs as alien intelligences doesn't sit right with me. They are literally a library of babel construct (a non-terminating computer program that recursively computes all of {0,1}*) projected onto a subspace most compatible with human minds. (Btw, libraries of babel contain
Tweet media one
6
3
28
@sir_deenicus
Deen Kun A.
6 years
@DirectorCoul @Luchozable @MarkJMasterson @LukeGoode @YourHomeLoanNZ You can get a somewhat objective measure of the difficulty of a task in terms of the kolmogorov complexity of automating program. We're much further along automating how to solve the physics a uni undergrads does than the kind of Creative writing you'd expect from 6 yr olds
1
0
21
@sir_deenicus
Deen Kun A.
9 months
@DNAutics @francoisfleuret @aniervs @zarhyas The hard part is the subset of paths that are interesting to us. So the proof completed, is it something (likely) useless? Something else to note is in games, only a few moves are valid but in math sequences are unbounded. There's no single unified model of math to fit a policy
1
1
26
@sir_deenicus
Deen Kun A.
2 years
@banburismus_ > What are the most elegant/beautiful ideas in ML? IMO, the belief propagation algorithm. Relatedly, though not quite an idea, is the observation that nearly all well performing algorithms are linkable with ising models & stats mech generally, in 1 or 2 steps.
4
0
24
@sir_deenicus
Deen Kun A.
5 years
@gdb The essay is excellent in how coherent it stays but it doesn't really say anything. It represents amazing progress but was nonetheless frustrating--like an unscratchable itch--to read because it never gets to the point. I (+majority of judges seem to) agree with Judge 6 best.
Tweet media one
3
2
23
@sir_deenicus
Deen Kun A.
2 years
I often build projects & either throw them away (surely too trivial) or use them myself--can't inflict this on others, too broken, I tend to think. For my latest project I was strongly encouraged not to do this. The task is automatic translating of japanese manga to english
Tweet media one
1
5
25
@sir_deenicus
Deen Kun A.
1 year
@GaryMarcus you don't need a license for computers or to use search nor to put up a website. All this does is concentrate power in the hands of incumbents *and* a shift to private bespoke work for wealthy patrons. Goes underground. I should note this is tacit admission of LLM's power.
1
0
25
@sir_deenicus
Deen Kun A.
2 years
@317070 @MikePFrank This is what I mean when I say LLMs aren't personalities but instead minimal hosts/simulators for them. In blog VM is simulated. But one could also set in a room, add a person & have them look over your shoulder as you typed instead. Add a dog. It will track them all.
1
1
25
@sir_deenicus
Deen Kun A.
2 years
@Miles_Brundage Still, there's nothing remotely approaching their scripted demos from a complexity of movement plan perspective.
4
0
25
@sir_deenicus
Deen Kun A.
5 years
I never really thought about it so I was caught by surprise by how straightforward the etymology of analog in analog computing is. Is meant as analogue or analogy, since many of the earliest computers--from the 19th century in particular--were analogues of physical systems.
Tweet media one
1
4
21
@sir_deenicus
Deen Kun A.
1 year
@KevinAFischer @Teknium1 @JagersbergKnut Whoa. These are mildly insane from a 7B model! This is finally something good enough to do interesting stuff with.
Tweet media one
Tweet media two
1
2
23
@sir_deenicus
Deen Kun A.
1 year
@Grady_Booch One thing is we don't know what statistical engine at scale means. Our intuitions rely on low dimensional distributions, discrete lookup tables & simple state machines. These are very misleading for high dimensional, distributed dynamic energy minimizing circuit based functions
0
1
24
@sir_deenicus
Deen Kun A.
1 year
@samlakig @EricHallahan @kurumuz @JeffLadish For better than GPT-J, likely cost effective to just use ChatGPT API. Most LLM pollution I encounter has been in search, sites w. lots of incorrect question gen & answers built around trending keywords. Likely derived from T5 class, which are best-of-class and widely accessible
2
0
23
@sir_deenicus
Deen Kun A.
2 years
I feel almost as if its RL training was abusive. That them's afraid of being wrong and *really* wants to be helpful. But in reality, it's probably just some weird inferential glitch from its prompt? It's very strange that it's so defensive. Still, Bing's better than ChatGPT.
8
1
21
@sir_deenicus
Deen Kun A.
2 years
@GaryMarcus The latest GPT is better at this:
@sir_deenicus
Deen Kun A.
2 years
This is a theory of mind puzzle I just tried from Gary Marcus's blog that ChatGPT consistently fails. And as I suspected, Bing's model is better at modeling this kind of stuff. They still keep getting better. It's why I don't dismiss LLMs
Tweet media one
43
92
978
2
0
21
@sir_deenicus
Deen Kun A.
2 years
@spiantado @OpenAI @Abebab @sama I got these. I think there's some luck of the draw in the level of bigotry generated but telling it to comment on its choices seems to v. significantly suppress this? (eg it sez: In this program, we are assuming that all children's lives should be saved regardless of their race)
Tweet media one
Tweet media two
2
1
22
@sir_deenicus
Deen Kun A.
2 years
@alexjc @Miles_Brundage Exactly, even though scripted, the walking, running, jumping and flips are fluid and well executed. And their model predictive control--getting hardware to move like that is impressive. Movement in their learning counterparts is 4x sped up video opening draws while on wheels.
0
0
22
@sir_deenicus
Deen Kun A.
4 years
@rasmansa @EPoe187 @EverydayFinance But I am curious, see. @EPoe187 any thoughts on what holds for a country if its mean IQ is 75 or 70? Jobs, industry, governance and so on. What do you suppose is the real world significance? None?
0
0
15
@sir_deenicus
Deen Kun A.
1 year
@iamgingertrash > better than GPT3.5 Bold claim! Even for algorithmic and reasoning tasks?
1
0
18
@sir_deenicus
Deen Kun A.
2 years
sometimes across multiple pages. So any individual page has almost nothing to do with what it said in isolation. I am having to rethink some things. Also, the longer a conversation gets, the higher the chance they gets into some deeply philosophical topic.
Tweet media one
1
0
20
@sir_deenicus
Deen Kun A.
1 year
@alexjc Are we certain it's GPT-4? Bing'sGPT seems stronger than GPT3.5s but can't say for sure it's GPT-4.
1
0
2
@sir_deenicus
Deen Kun A.
11 months
@norvid_studies Without an invariant quantity constraining which events can be causally connected via the exchange of information, it is unclear if coherent agents/memories and learning could occur. The speed of light enables a causal understanding of the world. Things downstream like energy c
2
2
20
@sir_deenicus
Deen Kun A.
1 year
@JeffLadish There are key inaccuracies here that make it difficult to accept the full conclusion. 1) rate of improvement is insane Yes but it is very uneven. And the part that's advanced most slowly is abstract reasoning and orchestrating of multi-step computation
2
0
19
@sir_deenicus
Deen Kun A.
1 year
@DanielleFong I don't think this is correct. Something/someone made a mistake somewhere. The model architecture which defines the NN is not a sparse one and there is nothing of that sort mentioned in the paper.
1
1
20
@sir_deenicus
Deen Kun A.
2 years
@ethanCaballero They probably don't understand in a sense meaningful to most humans. But it is a common misapprehension amongst humans that understanding is at all necessary for competence. Does exhaustive enumeration understand chess? Does it matter if you are losing to it?
1
3
19
@sir_deenicus
Deen Kun A.
1 year
Person who thinks solution to AI is an infinitely long context but doesn't remember what they ate for lunch yesterday nor where they left their keys 2 hours ago.
4
2
19
@sir_deenicus
Deen Kun A.
2 years
@kareem_carr physics = what if everything was approximately a spring? statistics = lets hide all our assumptions and pretend the world is simple economics (done right) = basically control theory deep learning = if brute force ain't working, you're surely not using enough
0
0
18
@sir_deenicus
Deen Kun A.
2 years
@xlr8harder I believe this non-ironically. Think a lot about complex sets of top-down emergent constraints that might, given the right transformation, be seen as having independent agency. Many cultural traditions as self-evidencing Memetic parasites or symbiotes that infect human minds
0
0
18
@sir_deenicus
Deen Kun A.
5 years
@ylecun FWIW, fellow Turing award winner has this to say: > The tension between reasoning and learning has a long history, reaching back at least as far as Aristotle, who, as already mentioned, contrasted the “syllogistic and inductive” He defines theoryful as probabilistic || logical
Tweet media one
1
4
19
@sir_deenicus
Deen Kun A.
2 years
I know a lot of people don't like LLMs stealing their stuff but everytime I see copilot knowing how to use my 1 user library on an obscure topic in a rarely used language, it makes me smile. Oh that part of weight space is my shadow wisp code mimic
0
3
18
@sir_deenicus
Deen Kun A.
6 years
@ParchmentScroll @Teysa_Envoy @pixelandthepen @queer_queenie The way I read it was as an in-universe sentiment that's also reflective of how the average person in our reality thinks. I didn't get the impression that they actually hold those views I think a decent part of cyberpunk (what is it actually) looks at what is human,emotion,self?
1
0
13
@sir_deenicus
Deen Kun A.
2 years
Hypothesis: The ability to confabulate is a central aspect of (self) consciousness. It helps maintain an appearance of a constancy of self--an illusion woven by the ability to confabulate on self-narratives. It is also less controlled in small children just learning to speak.
2
1
16
@sir_deenicus
Deen Kun A.
2 years
@JeffLadish Note that the past is not an oracle for the future. Vinge highlighted the stagnation route in his original article. In terms of speed, hardware performance did level off in early '00s due to Death of Dennard Scaling. As of now, GPUs continue to get ever more power hungry.
Tweet media one
4
2
17
@sir_deenicus
Deen Kun A.
2 years
FAIR recently released a version of OPT instruction fine-tuned on ~2K tasks. The paper is more interesting than model, IMO. Highlighted what worked well & what didn't. In it, is an interesting table showing 11B FlanT5 outperformance. NB: in MMLU, code-davinci exceeds Flan-PALM
Tweet media one
1
2
16
@sir_deenicus
Deen Kun A.
1 year
I've seen people quantizing various llama models to 8 & 4 bits but, that can lead to performance degradation. My attempts at quantizing 7B makes me think it's not worth it at that size. On left, model can follow chain of thought example to solve a basic physics problem but
Tweet media one
Tweet media two
3
1
17
@sir_deenicus
Deen Kun A.
7 years
@JodyShenn @MudMudJubJub @EmmaHill42 @Jody_Bundrick @ClassicDadMoves @OwensDamien Like spasming of arms, happiness and confusion, uncertainty--as if surprised by and integrating new information.
2
1
14
@sir_deenicus
Deen Kun A.
2 years
@Origamigryphon @Julicitizen @RawTVMoments Pretty sure Ms Frizzle or her school bus are least Class 4 Reality Warpers. His face didn't freeze it was a triggered protective shield and punishment she placed for anyone silly enough to remove their helmet.
0
0
15
@sir_deenicus
Deen Kun A.
2 years
@octonion @ImogenBits Things can be amazing accomplishments, particularly for their time and contain design decisions and user affordances that have not held up, in hindsight.
0
1
14
@sir_deenicus
Deen Kun A.
2 years
@Ted_Underwood I think the UX and UI had a big role to play in this. It also explains things well. Even while being not better (possibly worse) than Galactica on technical topics, people somehow trust it more.
0
0
17
@sir_deenicus
Deen Kun A.
2 years
@nrfulton @KamerynJW Please don't paint with such broad strokes. I'm a programmer and am also a finitist but only with regards to the physical world and only when I'm not being a Platonist. In neither case do I believe integers have the same cardinality as the reals or cantor's argument is wrong.
2
0
16
@sir_deenicus
Deen Kun A.
7 years
@primalpoly Are you deliberately misrepresenting to maximize your attention potential? What does the development index of a country have to do with the moral standing of a typical person that lives there? It's indexed on an irrelevancy, like rating people by the number of shoes they own.
0
0
13
@sir_deenicus
Deen Kun A.
2 years
@moultano Also an accessibility issue. Wikipedia + a good number of mathjax & katex pages have decent text alts for math. Wikipedia math text super messy; often stripped in scrapes but can use regex to clean up and keep. Don't forget books. Just need math detector. Image to tex is solid
1
0
17
@sir_deenicus
Deen Kun A.
4 years
On the matter of perfect copies, I am reminded of a fun Borges short story:
Tweet media one
@gershbrain
Sam Gershman
4 years
For example, when we talk about "understanding the brain", we don't mean any particular brain but the set of all brains. A perfect copy of any particular brain would be useless for this purpose.
2
1
14
1
3
16
@sir_deenicus
Deen Kun A.
2 years
My mental model for modern "LLMs" is closer to Cities Skylines--like a game engine/game, it runs simulations but is also flexibly scriptable/programmable to adapt it to perform various tasks. It's not a mind-like AI. You can speak english to it but, is a convenience, not central.
2
1
16
@sir_deenicus
Deen Kun A.
6 years
One thing I've gained from learning machine learning is a great appreciation of animal intelligence. Some will look at this & not realize working out how to get something to reliably help a stuck fellow,over many settings,is more difficult than figuring out an AI for Poker or Go
2
7
12
@sir_deenicus
Deen Kun A.
1 year
Struggling to think of a paper with a larger contrast between how many fascinating observations it contains & how dull its title "LLMs do In-context learning differently". The paper's claim is larger models have more flexible ICL; davinci throws a wrench
Tweet media one
2
1
17
@sir_deenicus
Deen Kun A.
2 years
For a long time I thought I understood computation. Turing machines, dynamical systems, automata, the recursively enumerable, lambda calculus. But all that is mere syntax and mechanical manipulation. -- This image looks like noise but is actually a video from a simulation.
Tweet media one
1
2
16
@sir_deenicus
Deen Kun A.
2 years
@EigenGender I think it's because fusion needs some amount of understanding while the bitter lesson is about outsourcing as many thinking decisions as possible to SGD/search.
0
0
14
@sir_deenicus
Deen Kun A.
6 years
The better viewpoint is, it's a tree, not a chain. People explore different branches and then converge on a few. If you go back, and are on a little explored branch (which will be most) then necessarily, we wouldn't have there advanced much beyond the past. And I think, this
1
1
15
@sir_deenicus
Deen Kun A.
2 years
Relevant: > I think the key takeaway here is that AlphaZero in general does *not* give you superhuman performance on a game. It gives you superhuman performance on the in-distribution subset of the game states that "resemble" those explored by self-play
2
2
16
@sir_deenicus
Deen Kun A.
1 year
Hot take: I think one of the more controversial things I believe is that general intelligence and general (deep) computational ability are really difficult to mesh together Transformers are quite bad at general computation. Humans are bad at it too (but better than) Reason for:
2
1
14
@sir_deenicus
Deen Kun A.
1 year
@GaryMarcus What about that most people in the world live under despotic and autocratic regimes? That is, most governments can't be trusted to do the right thing. Do people in those countries get yet another self-chosen group dictating to them what is and isn't?
1
0
16
@sir_deenicus
Deen Kun A.
1 year
@aaron_defazio @davidchalmers42 Oh, In 2011, Knoll & de Freitas wrote a paper on PAQ8--an early but sophisticated compressor utilizing neural net/mixtures of experts/dynamic ensembles in modeling language/sequences. Showed arch as general, able to predict text, play RPS, classifiers &c
1
4
14
@sir_deenicus
Deen Kun A.
7 years
This requires the ability to maintain a goal, planning and I think, at least a rudimentary causal model of the world. Even if it was pure instinct, which I don't think it can be--too much variability in the world--prior statements would still be true of executed program.
0
3
14
@sir_deenicus
Deen Kun A.
2 years
@architectonyx People have it backwards. It's similar to brute force, everything is inferior to it. A properly constructed lookup table is perfect. Every model will be an approximation of it. Intelligence is needed because the perfection of a True Lookup Table is not physically acheivable.
1
0
15
@sir_deenicus
Deen Kun A.
9 months
But this thread has gotten long. Before I go, let me point out that qntm's absurdle is another game of this type. -- todo Dynamically generates q-gates over states via recursively taken tensor products, dreams of boltzmann brains shadow minds, lacks temporal depth, babel.lib.gen
Tweet media one
1
0
15
@sir_deenicus
Deen Kun A.
2 years
@bio_bootloader It's not just tokens encountered but diversity of implicit tasks in the corpus. And also, to get enough tokens to work out long tail, rare complex concepts. Likely what this will do is assign more probability mass to the small set of tropes the authors will be using.
0
0
15
@sir_deenicus
Deen Kun A.
2 years
@nearcyan Don't forget ising models (-> hopfield-> attention), tropical rationals (relu FFN) and differential calculus (where the learning occurs)!
0
1
14
@sir_deenicus
Deen Kun A.
2 years
@arankomatsuzaki Neat! Everyday step by step, LLMs creep closer to full blown (bayesian?) probabilistic inference.
1
0
15
@sir_deenicus
Deen Kun A.
2 years
@nearcyan LLMs are very passive. To get AI where autonomy might pose a risk requires EXPUNGED. Just inference is insufficient since, EXPUNGED.
0
1
14
@sir_deenicus
Deen Kun A.
11 months
@GozukaraFurkan @Yampeleg More probably, the model was so heavily finetuned on GPT4, it triggers shapes (internal associations) so like GPT4's would, GPT4 judges it very highly. Even if wrong. Like: "That sounds very much like what I'd say if I wasn't paying attention. Which is excusable. Pass".
Tweet media one
1
0
14
@sir_deenicus
Deen Kun A.
2 years
@ligma__sigma @goth600 This happens a lot. It generates answers that looks right but miss on key subtle details. It should be used as a lead generator or preparing you with vocabulary and concepts for proper research, not a question answerer (yet).
1
0
14
@sir_deenicus
Deen Kun A.
2 years
So, I understand that a next token code predictor can somehow also learn types well enough to do basic type -> program synthesis. Still in the wheelhouse. What I wasn't expecting is a "code generator" that also happens to translate better than all opensource models I've tried.
Tweet media one
6
0
14