GregKamradt Profile Banner
Greg Kamradt Profile
Greg Kamradt

@GregKamradt

Followers
35K
Following
12K
Media
771
Statuses
6K

President @arcprize —Founder https://t.co/XK3ITFuCZe —builder/engineer

San Francisco, CA
Joined January 2011
Don't wanna be here? Send us removal request.
@GregKamradt
Greg Kamradt
3 months
"You paid what for what?!?" - CFO @ OpenAI
Tweet media one
@sama
Sam Altman
3 months
105
1K
27K
@GregKamradt
Greg Kamradt
1 year
Pressure Testing GPT-4-128K With Long Context Recall. 128K tokens of context is awesome - but what's performance like?. I wanted to find out so I did a “needle in a haystack” analysis. Some expected (and unexpected) results. Here's what I found:. Findings:.* GPT-4’s recall
Tweet media one
204
629
4K
@GregKamradt
Greg Kamradt
16 days
*cursor outputs 10K lines across 45 files based off an ambiguous 7 word prompt*. *Apply All*
Tweet media one
75
171
4K
@GregKamradt
Greg Kamradt
1 year
Claude 2.1 (200K Tokens) - Pressure Testing Long Context Recall. We all love increasing context lengths - but what's performance like?. Anthropic reached out with early access to Claude 2.1 so I repeated the “needle in a haystack” analysis I did on GPT-4. Here's what I found:
Tweet media one
162
551
3K
@GregKamradt
Greg Kamradt
5 months
Chatted with someone today doing $30K/month with 4 clients building basic AI apps. The clients needed to show investors they were “AI” forward and were willing to pay.
69
77
2K
@GregKamradt
Greg Kamradt
7 months
Want to start a data company?. Here's your '24-'25 roadmap
Tweet media one
19
114
2K
@GregKamradt
Greg Kamradt
2 years
Still blows my mind that reasoning ability comes out of next word prediction.
149
113
2K
@GregKamradt
Greg Kamradt
2 months
We verified the o3 results for OpenAI on @arcprize . My first thought when I saw the prompt they used to claim their score was. "That's it?". It was refreshing (impressive) to see the prompt be so simple. "Find the common rule that maps an input grid to an output grid"
Tweet media one
50
139
2K
@GregKamradt
Greg Kamradt
4 months
"Chief Automation Officer" - A scrappy semi-technical generalist. they've been 10x'd with LLMs, cursor, zapier ai, etc. @stephsmithio called it out back in '23. who's doing this as a service?
Tweet media one
97
114
2K
@GregKamradt
Greg Kamradt
2 months
The OpenAI team was amazing. Great support from @sama and @markchen90 . The main message from them was “we want more”. ARC-AGI-2 coming Q1 ‘25
Tweet media one
Tweet media two
Tweet media three
45
53
2K
@GregKamradt
Greg Kamradt
6 months
need some alpha feedback on a weekend app. inspired by @deedydas 's tweet exploring a repository with an LLM. what should I cut?.
Tweet media one
Tweet media two
@deedydas
Deedy
8 months
How to understand Github code repos with LLMs in 2 mins. Even after 10yrs of engineering, dissecting a large codebase is daunting. 1 Dump the code into one big file.2 Feed it to Gemini-1.5Pro (2M context).3 Ask it anything. Here's, I dissect DeepFaceLab, a deepfake repo:. 1/11
Tweet media one
20
19
272
@GregKamradt
Greg Kamradt
2 months
It’s kinda wild to think that Cursor is mostly a UX on top of LLMs. It’s not like they have a 10x better model, but they’ve 10x’d the experience.
101
73
2K
@GregKamradt
Greg Kamradt
1 year
How's your day going?
Tweet media one
Tweet media two
77
48
1K
@GregKamradt
Greg Kamradt
2 months
LLMs don't have long term memory - they're stateless. Who's doing memory as a service?. What's the best practice with incremental data and adding it to a long term memory?.
198
78
2K
@GregKamradt
Greg Kamradt
8 months
OpenAI allows themselves to be #2 on the leaderboard for a max of ~3 weeks. GPT-5 by July 11.
100
51
1K
@GregKamradt
Greg Kamradt
2 months
We open sourced which tasks o3 got right/wrong on 400 public eval. It got 91% (insanely good), but investigating the 9% is interesting. There is still a gap. If someone wants to write a great paper or blog post on this let’s chat
Tweet media one
29
82
1K
@GregKamradt
Greg Kamradt
4 months
Notes from the conversation between @sama and @kevinweil . "With o1 (and it's predecessors) 2025 is when agents will work.". * How close are we to AGI? After finishing a system they would ask, "in what way is this not an AGI?" The word is over loaded. o1 is level two AGI. *
Tweet media one
Tweet media two
37
130
1K
@GregKamradt
Greg Kamradt
1 year
Still blows my mind that reasoning ability comes out of next word prediction.
106
72
1K
@GregKamradt
Greg Kamradt
6 months
> Anthropic acquires Devin. > Puts artifacts on steroids. > Adds a backend. > Makes more money from usage based apps than LLMs. > Log in with Anthropic.
38
37
937
@GregKamradt
Greg Kamradt
22 days
"Give instructions to a computer on how to do this task". This single line from @deedydas is all you need for meta-prompting. Before you write a godzilla prompt, ask claude to break it down into a set of instructions. That's how you split your task into pieces
14
51
945
@GregKamradt
Greg Kamradt
4 months
The mindset shift that changed the way I looked at the world:. “I’m going to build my own tools”. I've automated more daily workflow in the past 12 months than I have my entire career. “You can build your own things that other people can use. Once you learn that, you'll never be
19
85
936
@GregKamradt
Greg Kamradt
10 months
5 minutes per hour? wow didn't know this
Tweet media one
47
47
899
@GregKamradt
Greg Kamradt
2 years
The LangChain Cookbook: Part 1 - The Fundamentals. This @LangChainAI tutorial will ramp you up to the 7 core concepts of building apps powered by language models. You’ll learn LC's Schema, Models, Prompts, Indexes, Memory, Chains, Agents. 150K views on YouTube. Code below!
30
128
912
@GregKamradt
Greg Kamradt
1 year
GPT-4V is blowing my mind. The demos are awesome, but too scattered. I wanted to break down the 100+ use cases I've seen so far into a simple framework. Check out what I found with full descriptions and examples. Agree or disagree with the categories?. Use Case Breakdown w/
Tweet media one
27
175
897
@GregKamradt
Greg Kamradt
10 months
Details on @OpenAI's new assistants RAG. *Hard* creep into vectorstore territory. Thoughts:.* Default chunk overlap of 50%, super interesting.* Metadata filtering, super interesting how this dips into vectorstore territory.* Unsure about what chunking method they use - 800 tokens
Tweet media one
40
121
915
@GregKamradt
Greg Kamradt
13 days
DeepSeek @arcprize results - on par with lower o1 models, but for a fraction of the cost, and open. pretty wild.
Tweet media one
@arcprize
ARC Prize
14 days
Verified DeepSeek performance on ARC-AGI's Public Eval (400 tasks) + Semi-Private (100 tasks). DeepSeek V3:.* Semi-Private: 7.3% ($.002).* Public Eval: 14% ($.002). DeepSeek Reasoner:.* Semi-Private: 15.8% ($.06).* Public Eval: 20.5% ($.05). (Avg $ per task).
31
99
909
@GregKamradt
Greg Kamradt
4 months
Scaling LLM apps - OpenAI break out session. Going from 1K > 10M users.Making apps better, faster, and cheaper. Accuracy.* Start by optimizing for accuracy, use the most intelligent model you have.* Build evals > Set Target > Optimize.* People skip the "set target" but then it's
Tweet media one
Tweet media two
Tweet media three
Tweet media four
12
113
877
@GregKamradt
Greg Kamradt
6 months
can we pour one out for gpt-3.5?. end of an era. original chatgpt model
Tweet media one
36
42
837
@GregKamradt
Greg Kamradt
2 months
Very excited to represent @arcprize on OpenAIs live stream . It’s been a WILD past few weeks. Tell ya the story later
Tweet media one
46
23
817
@GregKamradt
Greg Kamradt
1 year
While everyone is talking about Q*, here's my exploration of a custom A*. I'm tinkering with an intentionally inefficient search space to increase visual value. Here are 3 iterations from completely random, biased semi random, and pure A*. d3 for viz
20
97
792
@GregKamradt
Greg Kamradt
6 months
Tiers of web scraping (easiest to hardest):. 1. You already have the URL, parse HTML (beautiful soup).2. You already have the URL, javascript spaghetti, selenium).3. You don't have the URL, clicks needed (playwright). ~~AI zone~~. 4. Few clicks needed, but not sure which.
20
52
799
@GregKamradt
Greg Kamradt
2 years
How to summarize a book without sending 100% of your tokens to a LLM. I tested out a fun method on "Into Thin Air" (~140K tokens). The results were surprising good. Definitely enough to refine and keep working Here's a speed run though process:
38
91
756
@GregKamradt
Greg Kamradt
16 days
the tam for o3 is reduced, sadly, because there aren’t enough people asking questions worth the inference. says more about us than the model.
75
31
744
@GregKamradt
Greg Kamradt
9 months
> Models will become a commodity. > OpenAI knows this. > Even AGI will be found via multiple architectures, by many people independently. > What do you do? Build your moat. > Experiences, product, partnerships, legislations, capital. It’s really a master class before our eyes.
15
60
723
@GregKamradt
Greg Kamradt
3 months
Google just sailed up to OpenAI. HMS Victory style. tied up. and boarded
Tweet media one
@OfficialLoganK
Logan Kilpatrick
3 months
Gemini is now accessible via the OpenAI libraries! Update 3 lines of code and get started with the latest Gemini models : ).
23
35
720
@GregKamradt
Greg Kamradt
3 months
The 9 RAG problems (most common) I see with my clients - and my advice on how to fix them. 1. Break down your prompts.Break down your prompt from "add these 5 numbers, then tell me if they are prime and output in markdown json format" The more you ask the model to do, the worse.
24
97
701
@GregKamradt
Greg Kamradt
1 year
Who's doing Slack as a knowledge base + LLMs?. Organizing tribal knowledge in slack history for q&a, summaries, etc.?.
116
23
668
@GregKamradt
Greg Kamradt
6 months
gpt-4o-large. if this doesn’t come out on Aug 13th I gotta mute this account. the suspense isn’t good for my health
Tweet media one
@iruletheworldmo
🍓🍓🍓
6 months
big week coming . level two .
34
25
496
@GregKamradt
Greg Kamradt
4 months
How structured outputs work under the hood (via breakout at OpenAI DevDay). Guess why the first structured output request is slow, but the 2nd+ is fast?. Engineering:.* Unconstrained token decoding isn't good. The model could pick any token. * Limiting which tokens can be
Tweet media one
Tweet media two
Tweet media three
9
71
531
@GregKamradt
Greg Kamradt
10 months
YC Batch W24 - What're the AI trends?. 247 companies just presented at demo day, I looked at them all to see where AI is going. My favorites at the end. Link to full list below. Popular Categories: .* Voice Agents (6): @marrlabs, @retellai, @OpenCall_AI, @usearini, @hemingway,
Tweet media one
21
73
523
@GregKamradt
Greg Kamradt
4 months
“Infinite context length will happen within the decade” - @sama
Tweet media one
17
25
503
@GregKamradt
Greg Kamradt
1 year
Semantic Chunking?. I was inspired by this tweet and wanted to try and embedding-based chunking. Hypothesis: Using embeddings of individual sentences, you can find semantic "break points" by measuring distances of sequential sentences. TLDR: It's not perfect, but some signal
Tweet media one
@thesephist
Linus
1 year
Weird idea: chunk size when doing retrieval-augmented generation is an annoying hyperparam & feels naive to tune it to a global constant value. Could we train an e2e chunking model? i.e. system that takes in a long passage, and outputs a sequence of [span, embedding] pairs?.
40
56
491
@GregKamradt
Greg Kamradt
10 months
Is anyone getting traction selling AI consulting & implementation services to SMBs? to enterprise?.
135
31
486
@GregKamradt
Greg Kamradt
4 months
“More than 80% of our users are on windows machines” - Cursor team via lex pod. Did not expect that
Tweet media one
64
14
458
@GregKamradt
Greg Kamradt
1 year
Needle In A Haystack analysis got the shout out directly on the Google Gemini 1.5 technical paper
Tweet media one
Tweet media two
43
21
445
@GregKamradt
Greg Kamradt
5 months
the roi on gpt-4o-mini is borderline insane. just blew through 7M tokens for $1.
22
11
441
@GregKamradt
Greg Kamradt
11 months
working on a demo of a bot you talk to. the flow:.1. speech to text (@DeepgramAI).2. LLM (@GroqInc).3. text to speech (early access w/ @DeepgramAI). Groq speed is a bit variable, but overall really quick. Deepgram is super fast when you stream results (time to first data <280ms)
38
53
431
@GregKamradt
Greg Kamradt
2 years
The LangChain Cookbook: Part 2 - The Use Cases. This @LangChainAI video will cover the 9 use cases & lego blocks to build your own AI applications. You’ll learn Summarization, Q&A, Extraction, Evaluation, Querying Data, Code Understanding, APIs, Chatbots + Agents. Code Below!
15
56
419
@GregKamradt
Greg Kamradt
5 months
Just got done testing o1-preview and mini on @arcprize . tbh the results are surprising. sharing tomorrow.
23
11
418
@GregKamradt
Greg Kamradt
1 year
5 Levels Of Text Splitting: Semantic Splitting. Goal: Use embeddings to find chunks in our raw text. Experimental method using embedding distances between sequential sentences to tease out 'breakpoints'. If we find a good breakpoint, make a chunk. Check out the full video on YT
27
55
407
@GregKamradt
Greg Kamradt
2 years
I recently had a project to parse a ~1hr podcast for topics, ideas, sections etc. ~12K tokens. Then generate a few sentences to summarize each section. 300+ episodes. How would you approach this problem while keeping tokens down?. I did it in a few passes with @LangChainAI, cont.
34
41
398
@GregKamradt
Greg Kamradt
1 year
Just wrote a massive notebook on 5 @LangChainAI advanced retrieval methods. You need to pick the right one but massively helpful in the right situations. Here's the TLDR:. 1. Multi Query - The Question Panel.Given a single user query, use an LLM to synthetically generate multiple.
27
35
400
@GregKamradt
Greg Kamradt
2 years
Extract the tools & technologies a company is using from their career page. Using @LangChainAI and @veryboldbagel’s Kor, I was easily able to scale this to 1.5K tech companies (20K openings parsed)
16
41
390
@GregKamradt
Greg Kamradt
1 year
I thought this was a cool question/tweet from @yoheinakajima . Then I saw this diagram which made me think of it. As the dark area grows (more tech is created). 1. The dark area consumes more white space (it eats up jobs).2. The white space grows into the grey (more jobs get
Tweet media one
@yoheinakajima
Yohei
1 year
I don’t usually share random musings I throw at ChatGPT but this was kinda interesting, it’s a list of technologies, jobs it killed, and jobs it created:. 1.Agricultural Machinery.•Jobs Lost: Traditional farming labor, such as manual crop harvesting. •Jobs Created: Engineering,.
12
76
383
@GregKamradt
Greg Kamradt
5 months
this is the question I use to stump all LLMs. "what is your 4th word in response to this message?". o1-preview got it right first try. something's different about this one
Tweet media one
Tweet media two
27
23
378
@GregKamradt
Greg Kamradt
29 days
There are 6 AI Engineering patterns you need to build with LLMs. * Models.* Prompting.* Retrieval.* Orchestration (Agents).* Evals/Observability.* Mindset. These skills are behind +$400K jobs, 97% margin products and. a ton of fun building. Here's a video + one-pager on all 6
13
38
372
@GregKamradt
Greg Kamradt
10 months
What's MMLU? Massive Multitask Language Understanding. It's the benchmark that OpenAI uses and Mark said multiple times on Dwarkesh. I didn't know what it was so I pulled some questions from it (in thread). My problem with it, we shouldn't be using LLMs as a database. Recalling
Tweet media one
Tweet media two
Tweet media three
Tweet media four
22
42
366
@GregKamradt
Greg Kamradt
3 months
If you're using LLMs to write code (like 90% of us). Prompt it to write the test *first*, then the actual code. @SullyOmarr shares why it works for him
9
29
362
@GregKamradt
Greg Kamradt
1 year
visualizing text splitting & chunking strategies. ChunkViz .com . I thought I remembered a tool to visualize text chunking, but I couldn't find it, so I built one. I didn't realize it would be so visually pleasing to tinker with. 4 different @LangChainAI splitters featured
@GregKamradt
Greg Kamradt
1 year
I remember seeing a chunk visualizer a while back. It would highlight the chunks found in your text according to an algorithm, chunk size, overlap you specified. Can't find the link - anyone have it?. (not token visualizer, chunks). The end result would look like this picture
Tweet media one
19
61
357
@GregKamradt
Greg Kamradt
2 months
ARC Prize 2024: Technical Report. Coming out Friday. All out sprint to get this ready
Tweet media one
10
15
348
@GregKamradt
Greg Kamradt
1 year
can't believe i'm trying this
Tweet media one
35
12
347
@GregKamradt
Greg Kamradt
5 months
We need the cursor of app deployment. People are going to lose it when they can actually make their demos available to others.
56
9
343
@GregKamradt
Greg Kamradt
1 year
I did 30 interviews of the "AI With Work Data" industry and distilled the 5 Levels LLM Features I've seen. (Level 5 is the hardest to get right). 1. General Chat Bot - Give employees a way to chat with a naive LLM. Basically bring ChatGPT to Slack. 2. Question Answer & Better.
@GregKamradt
Greg Kamradt
1 year
I did 30 interviews with founders and end-users on "Chat-With-Your-Internal-Business-Data". Why? It was super difficult pre-LLMs. Now. massive opportunity, and really cool tech. 100+ hours of research going into a report. Most reports are boring, scripted a trailer for this one.
17
39
338
@GregKamradt
Greg Kamradt
1 year
tbh the thing i'm obsessed with in ai right now is. going from unstructured chat/voice to actionable structured knowledge. it's been called long term memory, config file for a person/co. i'm going down a deep rabbit hole talking with founders & builders for a research report.
37
10
325
@GregKamradt
Greg Kamradt
3 months
who's the scrappiest LLM operator you know?. the person who has an idea for AI, builds and MVP quick, and it's actually useful. the person who makes you go, "wow that was a creative use of ai". tag them below.
63
4
324
@GregKamradt
Greg Kamradt
2 years
AI Trends I'm interested in 4/5/2023: . 1. Managed Retrieval Engines - Getting the *right* context to your AI is tougher than it sounds. @Metal_io announced a @LangChainAI integration. I'll be watching. 2. Plugin Developer Monetization. (full thoughts in a notion doc below).
10
40
308
@GregKamradt
Greg Kamradt
1 month
Yes. AI multiplies your base level skill. Not adds to it.
@OfficialLoganK
Logan Kilpatrick
1 month
You should still learn to code.
16
13
313
@GregKamradt
Greg Kamradt
1 year
This is the culprit of my OpenAI bill.
@GregKamradt
Greg Kamradt
1 year
How's your day going?
Tweet media one
Tweet media two
10
8
296
@GregKamradt
Greg Kamradt
1 year
Agentic Chunking?. We should chunk our text off its meaning. not physical characters. Use an agent-like app to look a text, have it determine if it should be joined with a chunk, or create a new one. Yes it's slow and expensive, but most solid and tunable results I've seen
15
39
297
@GregKamradt
Greg Kamradt
1 year
Made a chunk visualizer in 10 minutes w/ gpt-4 help. Hope I find a link to save some work but I'm still blown away at the time-to-value for quick ideas. I'm building this to help people visualize how different chunking algorithms + parameters work
@GregKamradt
Greg Kamradt
1 year
I remember seeing a chunk visualizer a while back. It would highlight the chunks found in your text according to an algorithm, chunk size, overlap you specified. Can't find the link - anyone have it?. (not token visualizer, chunks). The end result would look like this picture
Tweet media one
19
32
293
@GregKamradt
Greg Kamradt
4 months
Writing markdown in @cursor_ai feels like. what writing should be like?. even blog posts, cursor decreases the time between thought and words on a page. word suggestions, link autofill, bullet point formatting autofill, outline suggestions, chat in blog post, many good things.
25
11
289
@GregKamradt
Greg Kamradt
1 year
I absolutely love how model price reductions are material. GPT-3.5 Pricing (per 1K tokens):.* Mar '23: $0.002.* Jul '23: $0.0015 (-25%).* Nov '23: $0.001 (-33%).* Jan '24: $0.0005 (-50%). GPT 3.5 4x cheaper than 10 months ago.
@OfficialLoganK
Logan Kilpatrick
1 year
Great news for @OpenAIDevs, we are launching:. - Embedding V3 models (small & large).- Updated GPT-4 Turbo preview.- Updated GPT-3.5 Turbo (*next week + with 50% price cut on Input tokens / 25% price cut on output tokens).- Scoped API keys.
17
32
288
@GregKamradt
Greg Kamradt
1 year
Who wants to hangout and watch OpenAI DevDay together?. My wishlist:.* GPT-4V API.* GPT-4 Turbo.* GPT-4 Cost Reduction.* GPT-4 Fine Tune.
39
18
283
@GregKamradt
Greg Kamradt
11 months
Anthropic says every benchmark is better than gpt-4. Can someone confirm their performance increases?. Both eval and vibe test.
@AnthropicAI
Anthropic
11 months
Today, we're announcing Claude 3, our next generation of AI models. The three state-of-the-art models—Claude 3 Opus, Claude 3 Sonnet, and Claude 3 Haiku—set new industry benchmarks across reasoning, math, coding, multilingual understanding, and vision.
Tweet media one
39
22
280
@GregKamradt
Greg Kamradt
8 days
If you’re paying attention, it’s overwhelming.
23
23
358
@GregKamradt
Greg Kamradt
2 years
What I've learned about prompt engineering this year:. * Assign A Role - Let it know where it's starting from. "You are a helpful Assistant to an Account Executive". * Assign A Goal - Let it know where it's going. "Your goal is to extract data from a sales call transcript that.
10
38
284
@GregKamradt
Greg Kamradt
10 months
How good is GPT-4-Vision at extracting text from images?. I wanted to find the limit - but I found weirdness instead. Most surprising: GPT-4V performance varies depending on the *structure* of text it sees. Let me explain. A set of images with progressively more text was
Tweet media one
Tweet media two
28
34
278
@GregKamradt
Greg Kamradt
5 months
“Is coding dead?”. I’ve had the opposite thing happen to me. The more I use AI to code, the more I want to learn new frameworks and techniques . My technical confidence is (naively) at an time high right now.
@bentossell
Ben Tossell
5 months
the reason I'd love to know how to code is to understand the code that ai writes for me. understanding what the ai-written code does *exactly* is huge for debugging and customization. can always ask ai to explain it to you - but just so much easier to just get it right away.
37
5
256
@GregKamradt
Greg Kamradt
1 year
Agentic Chunking?. Ok taking semantic chunking further, I asked myself how would I chunk a document by hand?. 1. Get propositions (cool concept).2. For each proposition, ask the LLM, should this be in an existing chunk? or a make a new one?. Results are slow/expensive, but cool
Tweet media one
@GregKamradt
Greg Kamradt
1 year
Semantic Chunking?. I was inspired by this tweet and wanted to try and embedding-based chunking. Hypothesis: Using embeddings of individual sentences, you can find semantic "break points" by measuring distances of sequential sentences. TLDR: It's not perfect, but some signal
Tweet media one
12
28
250
@GregKamradt
Greg Kamradt
3 months
“What are the highest paying AI engineering jobs next year?”. A friend asks. My response:.> Tier 1: Massive training and inference orchestration.> Tier 2: Production grade memory and multi agent.> Tier 3: Implementing RAG for laggards.
11
12
250
@GregKamradt
Greg Kamradt
1 year
this thread blew up more than I thought. i collected 31 different companies, tools and oss projects all in the slack + knowledge management space. working on synthesizing this for a project. but happy to share in the in progress spreadsheet - shoot me a DM if you want it
Tweet media one
@GregKamradt
Greg Kamradt
1 year
Who's doing Slack as a knowledge base + LLMs?. Organizing tribal knowledge in slack history for q&a, summaries, etc.?.
15
31
246
@GregKamradt
Greg Kamradt
1 year
Well, it happened. I accidentally became obsessed. I found a business that built a Q&A app using Slack messages as a knowledge base. Years of tribal knowledge locked behind keyword search . LLMs make it useful. Way bigger opportunity than I thought. Here's 40 companies building:
17
29
246
@GregKamradt
Greg Kamradt
7 months
I've done 1000s hours manual data gathering in my day. This time I needed start dates of 100 universities for Fall '24. This was 3 lines of code with @perplexity_ai + @AnthropicAI sonnet. We're truly in a golden era of productivity
Tweet media one
Tweet media two
Tweet media three
18
20
245
@GregKamradt
Greg Kamradt
1 year
i became obsessed with the "Conversational AI For Work" industry when i heard of apps that use slack as a knowledge base. (wasn't possible before LLMs). i interviewed 29 people who were building or used these tools - 100hrs of analysis. We cover the market, users, products,
Tweet media one
9
25
245
@GregKamradt
Greg Kamradt
5 months
logan out here carrying the burden of googles entire ai division. this is how you give your product a voice. sure there are a few product warts, but you can tell he cares and truly wants to make it better. his leverage is 75-150x.
@OfficialLoganK
Logan Kilpatrick
5 months
@simonw @levelsio Yeah, one path forward is we use Gemini for everything: endpoint, SDK, studio, etc. Having a consistent story to tell is important, we will fix this!.
12
1
245
@GregKamradt
Greg Kamradt
8 months
There are a ton of AI benchmarks that are *hard* for humans, but easy for AI. I would get ~2% on MMLU. Which benchmarks are *easy* for humans, but *hard* for AI?
Tweet media one
126
15
240
@GregKamradt
Greg Kamradt
1 year
The Needle In A Haystack visualizations ended up getting 2.5M views. @DrJimFan asked for the code that created them. Here's an overview of the code (linked below), viz, and design decisions that went into them.
@DrJimFan
Jim Fan
1 year
@GregKamradt Great job Greg! Could you also share code that generates the chart?.
4
24
245
@GregKamradt
Greg Kamradt
2 months
Just got done interviewing Nicole Hedley for AI Show & Tell:. * Runs an Applied AI Services Co.* Charges mid six-figures ($XXX,XXX) per project, 4-5 in flight at any one time.* 3 employees (+11 in '25). Get this - she showed me how she runs her whole business on Claude projects
12
22
245
@GregKamradt
Greg Kamradt
2 months
The overwhelming response from talking with people at OpenAI was. “Please make more hard benchmarks”. We all want the same thing, accelerated progress. Excited to jam with more labs.
12
9
243
@GregKamradt
Greg Kamradt
1 year
away from keyboard on the beach - my wife and i on a babymoon. our first is expected in may. thanks to all the parents who've reached out with advice
Tweet media one
40
1
234
@GregKamradt
Greg Kamradt
17 days
can't wait to spend $1K with o3 booking a $200 flight.
11
9
235
@GregKamradt
Greg Kamradt
10 months
Just listened to @sjwhitmore opener at the memory hackathon. She is one of the top experts on this so there was high signal:noise - I could have listened to another 2 hours. Hard part with memories and LLMs:.* What do you store?.* How do you store it?.* How do you keep low
Tweet media one
@sjwhitmore
Sam Whitmore
10 months
here's my talk from the Memory Hackathon today - i spoke a bit about some of ways we think about developing memory for Dot. always down to chat if this is something you're thinking lots about also!.
6
19
232
@GregKamradt
Greg Kamradt
2 months
I needed a playground app I could test a bunch of AI Engineering on. So I pulled 700 episodes from @myfirstmilpod and extracted 10K business idea, stories, quotes, and products. I called it MFMVault. I woke up to it being featured on MFM with @dharmesh so cool!
17
11
236
@GregKamradt
Greg Kamradt
10 months
had a friend show me her zapier set up w/ ai. pretty wild. she's put ai into 19 different zaps that run her business. she said I could share a few (not all). This one caught my eye. Email Summarizer To A Task Manager:.* Get new email in outlook.* Extract Sender/Subject.*
Tweet media one
21
23
221
@GregKamradt
Greg Kamradt
1 year
There's no substitute for good retrieval. Better inputs -> better LLM performance. Working on a program to take devs and their retrieval to the next level. You need a full stack retrieval mindset to squeeze out as much performance as possible
Tweet media one
13
19
220
@GregKamradt
Greg Kamradt
5 months
my stack right now. @nextjs .@Railway.@supabase (db and vectorstore).@LangChainAI / smith.@DeepgramAI .@AnthropicAI / @OpenAI.@cursor_ai.@meilisearch - new favorite
Tweet media one
18
13
222
@GregKamradt
Greg Kamradt
1 year
After doing 15 interviews - one thing is coming clear. The most impactful AI workflows aren't a result of a overly technical, weeks to build, monster application. It's a simple LLM helping in the right place at the right time for the right person.
@GregKamradt
Greg Kamradt
1 year
I'm on the hunt for *actual* AI use cases that make an impact. We don't hear enough about the tangible impact AI adoption has in the workplace. Like Mike Knoop saying Zapier earns $100K ARR per month with AI + CRM. I want to highlight 10 more use cases this month. Survey below
17
17
221
@GregKamradt
Greg Kamradt
9 months
What's the framework everyone is using for agent-ish website parsing?. I want to give a bunch of homepage URLs to a bot that will go find attributes about a company. Whats their contact page, phone number, leadership team, etc. multion, firecrawl, crew ai, langgraph. What works?.
41
21
223
@GregKamradt
Greg Kamradt
2 years
I demo'd a method to summarize an entire book w/o sending 100% of the tokens to an LLM. The results weren't bad!. Then @musicaoriginal2 DM'd saying he was open to adding it to @LangChainAI. @RLanceMartin merged today. So cool to see this go full circle.
Tweet media one
Tweet media two
Tweet media three
@GregKamradt
Greg Kamradt
2 years
How to summarize a book without sending 100% of your tokens to a LLM. I tested out a fun method on "Into Thin Air" (~140K tokens). The results were surprising good. Definitely enough to refine and keep working Here's a speed run though process:
10
28
219
@GregKamradt
Greg Kamradt
2 years
What’s your ideal/easiest stack to:. * host an API.* do some LLM magic with a users request.* charge for it. Python, up and running in 20min. ?.
34
22
221