harryworld Profile Banner
Harry Ng ᯅ Profile
Harry Ng ᯅ

@harryworld

Followers
1K
Following
4K
Statuses
2K

Father. Slow-nomad. Indie. 🍎 Technologies Instructor. Formerly @GoodnotesApp @SortedHQ @GA 🇭🇰 🇹🇼

Joined March 2009
Don't wanna be here? Send us removal request.
@harryworld
Harry Ng ᯅ
3 months
我是為了交朋友才演講的 #iPlayground
Tweet media one
2
1
30
@harryworld
Harry Ng ᯅ
3 hours
It takes a lot of courage to rebuild a product from scratch, and I’m looking for support to push myself forward
0
0
4
@harryworld
Harry Ng ᯅ
17 hours
2月底至3月初,分別要去上海、杭州、深圳走走
0
0
2
@harryworld
Harry Ng ᯅ
2 days
The role of PMs hasn’t changed. In the past, one may rely engineering teams to deliver. Now the engineering is a bunch of AI tools or agents. If you know what you’re doing, you can probably build stuff 100x faster.
@steipete
Peter Steinberger
2 days
“People don’t like hearing this but this is the reason why startups don’t hire big tech folks.” made the same learnings. As a small company you want people that get sh*t done. Folks from Google & co usually aren’t that.
0
0
3
@harryworld
Harry Ng ᯅ
2 days
Cursor agent 有時很固執 有句代碼必須加上 self. 才能通過編譯 他總是會把 self. 刪掉 真像個挑皮的小孩
0
0
2
@harryworld
Harry Ng ᯅ
2 days
@johnny____11 在 Cursor 的提問過程,要求 AI 把細節解說一遍,接下來只是看你想不想讀
1
0
3
@harryworld
Harry Ng ᯅ
2 days
@caiyue5 什麼進階用法? 開個會員,開班教學
0
0
2
@harryworld
Harry Ng ᯅ
5 days
Swift Taichung 新年特別版聚會要來了
1
0
3
@harryworld
Harry Ng ᯅ
5 days
0
0
1
@harryworld
Harry Ng ᯅ
6 days
Setting up the 2 targets for 2025 1. Teaching app programming 2. Building app product again
1
0
6
@harryworld
Harry Ng ᯅ
6 days
RL vs SFT
@_philschmid
Philipp Schmid
7 days
SFT Memorizes, RL Generalizes. New Paper from @GoogleDeepMind shows that Reinforcement Learning generalizes at cross-domain, while SFT primarily memorizes. rule-based tasks, while SFT memorizes the training rule. 👀 Experiments 1️⃣ Model & Tasks: Llama-3.2-Vision-11B; GeneralPoints (text/visual arithmetic game); V-IRL (real-world robot navigation) 2️⃣ Setup: SFT-only vs RL-only vs hybrid (SFT→RL) pipelines + RL variants: 1/3/5/10 verification iterations (”Reject Sampling”) 3️⃣ Metrics: In-distribution (ID) vs out-of-distribution (OOD) performance 4️⃣ Ablations: Applied RL directly to base Llama-3.2 without SFT initialization; Tested extreme SFT overfitting scenarios; Compared computational costs versus performance gains Insights/Learning 💡 Outcome-based rewards are key for effective RL training 🎯 SFT is necessary for RL training when the backbone model does not follow instructions 🔢 Multiple verification/Reject Sampling help improve generalization up to ~6% 🧮 Used Outcome-based/rule-based reward focusing on correctness 🧠 RL generalizes in rule-based tasks (text & visual), learning transferable principles. 📈 SFT leads to memorization and struggles with out-of-distribution scenarios.
Tweet media one
0
0
0
@harryworld
Harry Ng ᯅ
8 days
While AI can be used to teach hard skills, future teachers should spend more time to teach kids about being a human
@garrytan
Garry Tan
8 days
Prediction: One of the defining political conflicts of the next decade will be AI vs the extremely powerful teachers unions
Tweet media one
0
0
2
@harryworld
Harry Ng ᯅ
8 days
I have been rethinking about how AI can enhance daily planning in general, and saw YC RFS, seems to be a good opportunity
0
0
0
@harryworld
Harry Ng ᯅ
9 days
RT @ycombinator: From the AI breakthroughs of the last few months, a wave of new startup opportunities have been unlocked. Here are some o…
0
325
0
@harryworld
Harry Ng ᯅ
20 days
Arrived home~
0
0
0
@harryworld
Harry Ng ᯅ
21 days
This is a wrap for this 🇸🇬 trip, and I definitely enjoy the food in Singapore. Now is the time to onboard my flight ✈️ back to Taiwan. I’m missing my wife and kids ❤️
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
0
5