![betts Profile](https://pbs.twimg.com/profile_images/1866217014953410560/kxIKN77-_x96.jpg)
betts
@bettysrohl
Followers
147
Following
191
Statuses
128
AI/LLM enthusiast | Web3 | Privacy advocate š» Head of DevRel @PartisiaMPC Engineer @imperialcollege
LDN | NYC
Joined September 2022
RT @dabit3: Verifiable agents are the next meta in crypto x AI - agents that don't require trust. ā¢ can't be shut down or censored ā¢ actioā¦
0
28
0
@clydedevv @elizaOS @dabit3 @alfaketchum @togethercompute This is an awesome project! saving it to try it out. Have you hit any roadblocks with getting banned from twitter for scrapping, happened to me before š
1
0
1
RT @signulll: netscape built a browser, sold it like boxed retail softwareāyou had to go to compusa & pay a solid chunk of change for it. tā¦
0
1K
0
šØHappening right now! Tune in to hear more about the liquid staking on Partisia with @SceptreLS I'll also be joining to talk about on our new exciting developments in the AI space š¤ p.d. : feel free to ask me any questions on the topic!
0
2
11
After spending ages plugging into big-name APIs (looking at you, @OpenAI), I decided to dust off my AI fundamentals and build a model from the ground up. Not because those models are trashāheck, theyāre amazingābut because I missed the thrill of total control. Here how and why: I wanted to add a serious privacy layer. Letās be real: continuous learning from every userās data is convenient but can feel invasive if they didnāt opt in. So Iām forcing myself to create a design where the model only learns when users explicitly say, āYes, Iām cool with that.ā Step one was reacquainting myself with raw data wrangling. You canāt appreciate a good data pipeline until you label everything by hand and realize how easy it is to miss, oh, 10,000 outliers. Fun times. Next up: model training. Itās like that proud moment you break away from pre-cooked meal kits and actually chop veggies for the first time in months. Thereās something satisfying about seeing your GPU fan rev up because of your code, not someone elseās API calls. Donāt get me wrongāIām still a fan of high-level frameworks. Agents and half-built solutions can be a devās best friend in a pinch. But occasionally, going end-to-end reminds you what the sausage-making process really looks like. Spoiler: itās messy, but oh-so-rewarding. Finally, the privacy piece: if you want to keep user data out of the shared training loop, you need tight cryptographic or HPC-laced solutionsāthink multi-party computation or zero-knowledge proofs. Tools like @partisiampc (if youāre into blockchain) can help, but even a well-structured local DB with user consents goes a long way. The biggest lesson? Rebuilding a model from scratch teaches you more about performance tweaks, data ethics, and user trust than any quick AI wrap job ever will. And hey, if we can bake privacy in from day one, we all sleep better. edit note: I was very much hungry when I wrote this, all I could think about is obscure food references... š
1
5
22
RT @pippalamb: š¬š§š„ļøThe UK government just announced the most accelerationist changes to national AI policy we've ever seen (incl. 20x moreā¦
0
93
0
I cancelled my @GitHubCopilot and moved to @Tabby_ML itās a better open-source version of GC, fully open for customization. If that is not enough reason, here is why I think it's awesome: - Runs Locally Tabby can be self-hosted, meaning youāre not sending every keystroke to a third-party server. Perfect for dev teams working with sensitive or proprietary code bases. - Configurable Models Want to fine-tune your own code suggestions? as itās open-source, I've adapted the model to my stack and style guidelines. - Privacy Forward If youāre big on data privacy (and I am), not having to rely on a closed API is a game-changer. You control how and where code snippets are stored. My experience: I spun up Tabby in a local dev environment. Setup was surprisingly painless. I had it reading my projectās code patternsāand the auto-completions felt spookily accurate by day two. Check out the GitHub repo. If you give it a try, let me know how your setup goes. Always curious about new tips for local AI dev toolsāand love seeing open-source projects push the space forward.
1
0
9
A filing suggests @Meta Llama may have been trained on copyrighted materials. This opens a major can of worms about how training data is sourced. Legally, itās complicated. Fair use? Possibly. But scraping copyrighted text or images without explicit consent blurs ethical lines. Weāve seen calls for more regulated data pipelines or āopt-inā frameworks. The AI community is slowly realizing you canāt just slurp the entire internet for free. Technical solution: privacy-preserving tech, like secure multi-party computation $MPC, can let models learn insights without ever directly exposing raw dataāshoutout to real breakthroughs in that field @partisiampc My takeaway? Future AI might evolve with explicit licensing deals or share revenue with content creators. If we want sustainable AI, respecting IP rights is essential. Letās see how Meta responds.
1
6
29
š„ @base notched $2.3B in daily trading volume vs. @ethereum Mainnetās $2.2B. Thatās no small featāespecially for a relatively new chain with far lower total TVL. Cheaper transaction fees + targeted user incentives can turbocharge volume. This mirrors BSCās meteoric rise in 2021, when it offered near-zero fees for DeFi degens. But does this mean Ethereum is done? Hardly. Mainnet is still the global settlement layer. L2 or alt-chains feed off that security, bridging back for final settlement. The future? A multi-chain environment. Each chain or L2 can optimize for specific use casesāgaming, DeFi, micropayments, etc. I see Ethereum as the ābase layer,ā ironically, while specialized solutions flourish on top.
0
0
5
Betts on Tech āØ weekly recap on AI&crypto My top 3 bets of the week. 1ļøā£ Synthetic Data is Booming Tech giants like @nvidia , @Google , @Meta and @OpenAI are increasingly turning to synthetic data to train AI models. This approach addresses data scarcity and privacy concerns, enabling the creation of vast, diverse datasets without the limitations of real-world data. The next generation of AI models, including GPT iterations, are poised to leverage synthetic data for enhanced performance and efficiency. 2ļøā£ Cache-Augmented Generation (CAG): The New Cool Kid in Town CAG is emerging as a promising alternative to Retrieval-Augmented Generation (RAG). By preloading relevant documents into an AI model's context and precomputing key-value caches, CAG eliminates the need for real-time retrieval, reducing latency and potential errors. However, this method raises privacy concerns, especially when handling sensitive data, necessitating robust data governance and security measures. You can read more on my last post on CAG. 3ļøā£ AI Governance Talks are Heating Up The rapid advancement of AI technologies has sparked intense discussions on ethics, regulation, and governance. Industry leaders and policymakers are striving to establish frameworks that ensure AI development aligns with societal values and legal standards. The challenge lies in balancing innovation with responsibility, ensuring AI's benefits are realized without compromising ethical principles. Follow for more weekly recaps and daily updates āØ
0
0
5
@king__choo would love to read more of your views on the topic. I agree I think consensus is starting to shift and long context will be the next thing (or a hybrid).
1
0
1
The validity of RAG is fading, and long-context LLMs + prompt caching are why. RAG, to a certain degree, was premised on: 1ļøā£ Lower latency 2ļøā£ Lower costs 3ļøā£ Input token limits of early models 4ļøā£ Solving "needle-in-a-haystack" problems that long(er) context LLMs had But RAG has flaws: rigid preprocessing (metadata, chunking, embeddings) + bottlenecks at retrieval, where systems must predict queries before they exist. š This is where Cache-Augmented Generation (CAG) steps in: š”Hereās how CAG works: - Preload all relevant docs into the LLMās context. - Precompute key-value (KV) caches for instant, accurate responses. - No retrieval needed. ā” No latency. No errors. With long-context LLMs, CAG + prompt caching = unbeatable. @AnthropicAI , @OpenAI , and @Google are already leveraging it. But hereās the kicker: Caching isnāt just an efficiency hackāitās a paradigm shift in AI knowledge integration. šØ The catch? Caching requires temporarily storing data in memory. For enterprise use cases with sensitive data, this raises huge privacy concerns. OpenAi's Zero Retention Policy likely doesnāt extend to cached prompts. Cached sensitive data? Thatās a compliance nightmare waiting to happen. So, whatās next? I predict hybrid systems will emerge: - Cache static, reusable knowledge upfront. - Dynamically retrieve only whatās absolutely necessary. And if we solve the privacy puzzleāthrough ephemeral memory systems, encryption, or private cachingāprompt caching could unlock enterprise AI adoption at massive scale. CAG isnāt just a clever workaround. Itās laying the foundation for the next era of scalable, privacy-conscious AI systems. If youāre not thinking about this now, youāre already behind. Thoughts?
0
1
10
š
@bing styling its 'Google' search results to look like Google feels like the ultimate inspect -> copy CSS move. Imitation is the sincerest form of... SEO? š¤
0
0
5