Rockwood_XRay Profile Banner
RayLuan_NewOaks-e/acc Profile
RayLuan_NewOaks-e/acc

@Rockwood_XRay

Followers
1K
Following
2K
Statuses
3K

CEO@NewOaks AI; Former Tiktok PM; NewOaksAI: Human-Like AI Phone Agents That Convert in 60s

https://www.newoaks.ai
Joined April 2023
Don't wanna be here? Send us removal request.
@Rockwood_XRay
RayLuan_NewOaks-e/acc
2 months
Human-Like AI Phone Calls That Convert #buildinpublic
1
0
4
@Rockwood_XRay
RayLuan_NewOaks-e/acc
7 hours
Censorship is not an excuse that Open AI and other LLMs’ prices are higher than DeepSeek. All we want is good product with low price.
@AravSrinivas
Aravind Srinivas
21 hours
A bunch of well meaning folks told me it’s hypocritical to go on an anti-China and anti-CCP narrative while benefitting from the DeepSeek models. I agree and take the feedback. DeepSeek is awesome and the researches deserve a ton of respect. America should stop resorting to virtue signaling and censorship ideas and just aim for meritocratic supremacy. We will make the Perplexity product better than the DeepSeek app and that should be the major reason a user should prefer to use Perplexity app over DeepSeek. Auxiliary reasons like censoring and data security are not what we should focus marketing on. We will work on this (finetuning and more improvements) and will have more to share soon.
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
1 day
DeepSeek Chat is officially live on NewOaks AI!
Tweet media one
0
1
3
@Rockwood_XRay
RayLuan_NewOaks-e/acc
1 day
DeepSeek chat is officially live on NewOaks AI!
Tweet media one
0
1
2
@Rockwood_XRay
RayLuan_NewOaks-e/acc
3 days
Wow
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
7 days
The conclusion of this Google paper is very clear: "SFT is responsible for memory, RL is responsible for generalization" Simple conclusion: Supervised fine-tuning (SFT) is like showing students a large number of examples and answers. Students learn by imitating the examples. Reinforcement learning (RL) is like letting students solve problems by themselves, giving rewards for correct answers and penalties for incorrect answers. Students learn through trial and error and summarizing patterns. The researchers designed two tasks to test the model: General Points card game: This is a card arithmetic game that requires using four cards to make the target number 24. The researchers tested: Rule changes: Change the numerical rules of J, Q, and K to see whether the model has learned the arithmetic rules or just memorized the solutions under specific rules. Visual Variation: Change the color of the cards to see if the model can still recognize the cards despite the visual changes. Virtual Reality Navigation (V-IRL): Navigating in a virtual city using instructions and street view images. The researchers tested: Rule variation: Using different ways of giving directions (e.g., absolute direction "north" vs. relative direction "turn left"). Visual changes: Test in cities the model was not trained on to see if the model can recognize landmarks and navigate in new visual environments. Key findings: Reinforcement Learning (RL) is the generalization champion! Across all tasks, models trained with RL excel at adapting to new rules and visual environments. They learn the underlying principles of arithmetic and navigation and are able to handle situations they have never seen before. Supervised fine-tuning (SFT) tends to memorize. Models trained with supervised fine-tuning tend to memorize the training data. They perform well on tasks similar to the training data, but performance drops dramatically when the rules or visual context change. They are essentially recalling patterns in the training data rather than truly understanding the task. Reinforcement learning (RL) improves the model’s visual recognition capabilities. Interestingly, RL training even improves the model’s ability to recognize objects in images, which helps in the virtual navigation task. This suggests that RL can improve the model’s basic visual understanding capabilities. Supervised fine-tuning (SFT) is still a good helper for reinforcement learning (RL). Although reinforcement learning is better in generalization, supervised fine-tuning is still useful. It can help the model initially understand the instructions and give responses in the right format. This makes it easier for reinforcement learning to further fine-tune the model on this basis to achieve better performance. “Thinking time” is crucial for reinforcement learning (RL). Giving the model more “thinking time” (adding verification reasoning steps in reinforcement learning training) can further improve the model’s generalization ability. Application suggestions: If you want an AI model that can truly understand and adapt to new situations (generalize), reinforcement learning (RL) is a better training method. It teaches the model how to learn and how to solve problems flexibly. If you just need an AI model to perform well on tasks that are very similar to its training data, then supervised fine-tuning (SFT) might be sufficient. But it is likely to struggle when circumstances change even slightly. Think of reinforcement learning (RL) as giving your AI a “brain” that can think for itself, while supervised fine-tuning (SFT) is like giving it a “cheat sheet.” In the long run, the “brain” is obviously more powerful.
Tweet media one
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
8 days
FREE TRIAL: Increase Engagement with Conversational AI Voice Agent: via @YouTube
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
8 days
Dario is deeply feared by DeepSeek. That is all I can see. Control=Fear
@labenz
Nathan Labenz
10 days
The word "control" appears 24 times in this essay – all 24 referring to export controls Zero mentions of the challenges of controlling powerful AIs, and the words "safe", "safety", and "alignment" don't appear at all Strange for the CEO of "an AI safety and research company"🤔
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
10 days
With all due respect, closed mind does not get you the final win. DeepSeek is complete open source LLM. Its victory will belong to Human Species, not like Close AI or Close Claude
@DarioAmodei
Dario Amodei
11 days
My thoughts on China, export controls and two possible futures
4
0
2
@Rockwood_XRay
RayLuan_NewOaks-e/acc
10 days
Do you agree?
@27khv
Brian McDonald
11 days
Telegram founder Pavel Durov says China’s rise in AI, shown by DeepSeek’s success, stems from its Soviet-style education system, which fosters fierce competition—unlike Western schools that hide grades to protect feelings. “Eliminate the losers, and you eliminate the winners.”
Tweet media one
Tweet media two
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
10 days
Tweet media one
0
0
2
@Rockwood_XRay
RayLuan_NewOaks-e/acc
10 days
It's hard to believe, but due to H100 restrictions, DeepSeek was forced to train R1 manually, with thousands of Chinese citizens holding flags to act as logic gates.
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
11 days
Picaso style snake, very cool!
@BenMosleyArt
Ben Mosley
11 days
Happy Chinese New Year everyone! #ChineseNewYear
Tweet media one
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
12 days
@ylecun Compute is still the upper bound of AI
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
12 days
This is a fake DeepSeek account, scam
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
12 days
Agreed
@BarrettYouTube
Barrett
13 days
"probably because it's just harvesting data". It's funny how people like @bindureddy think that because it's Chinese company there must be something sinister going on where the CPC is gathering data on everyone around the world. The truth is that over the last few days Deekseek has got a massive amount of Publicity, and its obviously down to the fact that their servers just cannot handle all of the increased traffic, so they have decided to prioritise their app service.. It's really simple.
0
0
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
12 days
RT @BarrettYouTube: @bindureddy "probably because it's just harvesting data". It's funny how people like @bindureddy think that because it'…
0
1
0
@Rockwood_XRay
RayLuan_NewOaks-e/acc
12 days
300% ROI within one month, that is the best customer testimonial for NewOaks AI!
Tweet media one
0
1
3
@Rockwood_XRay
RayLuan_NewOaks-e/acc
15 days
DeepSeek is the Child
@sama
Sam Altman
15 days
A revolution can be neither made nor stopped. The only thing that can be done is for one of several of its children to give it a direction by dint of victories. -Napoleon
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
15 days
Which part is the most challenging part when you do it Greg?
@gregisenberg
GREG ISENBERG
15 days
Clearest path to a $10M ARR B2B AI agent startup
Tweet media one
0
0
1
@Rockwood_XRay
RayLuan_NewOaks-e/acc
15 days
Good article
@gregisenberg
GREG ISENBERG
15 days
Clearest path to a $10M ARR B2B AI agent startup
Tweet media one
0
0
0