Zhiting Hu Profile
Zhiting Hu

@ZhitingHu

Followers
3,525
Following
367
Media
59
Statuses
439

Assist. Prof. at UC San Diego; Artificial Intelligence, Machine Learning, Natural Language Processing

San Diego, CA
Joined April 2018
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@ZhitingHu
Zhiting Hu
4 months
Super excited to introduce Pandora, a generative video World Model interactively controllable by language. #Sora and #GPT4 are both powerful. How about fusing them in a single model? 💥 Pandora gives a preview:🔭 > Build a General World Model (GWM) super efficiently by
@MaitrixOrg
Maitrix.org
4 months
🔥Introducing Pandora 🌏 🪐 a World Model that generates videos of world states with real-time language control 🎥🕹️ Simulate the world across domains in an _interactive_ way! check out more
7
75
231
9
58
290
@ZhitingHu
Zhiting Hu
1 year
🧠Human brain has internal world model for conscious reasoning 🤖Language model reasoning also need world model! Reasoning via Planning (RAP) is a new principled framework for this. It prompts LM as WM & reasons w/ MC TreeSearch Better v ChainOfThought on plan-gen, math, logic
Tweet media one
@Ber18791531
Shibo Hao
1 year
🤖LLMs (autoregressive transformers) are not designed to reason in non-linear structures like humans... 🤔Instead of using it as a straight-minded reasoner, what about regarding it as a world model🌎and planning with it? Check out our new work: Reasoning via Planning (RAP) 🧵
Tweet media one
7
76
375
14
110
504
@ZhitingHu
Zhiting Hu
10 months
🔥 #NeurIPS2023 Tutorial🔥 Language Models meet World Models @tianminshu & I are excited to give tutorial on machine reasoning by connecting LLMs🗣️ world models🌎 agent models🤖 w/ amazing panelists @jiajunwu_cs @du_yilun Ishita Dasgupta,Noah Goodman
Tweet media one
5
71
411
@ZhitingHu
Zhiting Hu
1 year
🚨LLM Reasoners 🧠 A library for LLMs to do advanced reasoning, including latest algorithms: - Reasoning-via-Planning (RAP) 🎶 - Tree-of-Thought (ToT) 🌴 - beam search, and more All in unified perspective of world models🌎 and reward🥇 More alg & results coming soon!
Tweet media one
@Ber18791531
Shibo Hao
1 year
📣 Introducing LLM Reasoners (), a library for advanced reasoning with LLMs. Simply define a reward function and an optional world model, and let LLM Reasoners take care of the rest, including Reasoning Algorithms, Visualization, LLM calling, etc.
Tweet media one
7
139
648
5
72
344
@ZhitingHu
Zhiting Hu
4 years
Personal news: I'll be joining @UCSanDiego @HDSIUCSD as an assist. professor in Fall 2021. Looking fwd to joining the amazing colleagues. Huge thanks to mentors/collaborators Eric Xing, @rsalakhu , @tommmitchell , @dannydanr , @andrewgwils , Le Song etc for the invaluable support!!
20
4
291
@ZhitingHu
Zhiting Hu
4 years
#PhD #Recruiting Please RT! I’m looking for PhD students and postdocs @UCSanDiego @HDSIUCSD , to work on #ML / #AI / #NLP , and build principles, methodologies, and systems of AI agents learning from all types of experiences. More info:
Tweet media one
4
119
241
@ZhitingHu
Zhiting Hu
2 years
🤯Machine Learning has many paradigms: (un/self)supervised, reinforcement, adversarial, knowledge-driven, active, online learning, etc. Is there an underlying 'Standard Model' that unifies & generalizes this bewildering zoo? Our @TheHDSR paper presents an attempt toward it 1/
Tweet media one
1
45
234
@ZhitingHu
Zhiting Hu
1 year
🗣️Language Models ➕ 🌎World Models When the two fascinating models start to interplay with each other, lots of exciting things happen! 🔥Here we use WM to teach LM diverse embodied knowledge and skills. Improve by 64% on 18 tasks, and let GPT-J-6B surpass chatGPT! 🔥
Tweet media one
@szxiangjn
Jiannan Xiang
1 year
🤔Humans can learn from embodied experiences in the physical world. Can Language Models also do that? 🔥Check out our new paper about enhancing Language Models with World Models! 👇 1/n
Tweet media one
5
68
355
2
48
221
@ZhitingHu
Zhiting Hu
10 months
🚩LLM360 — the 1st fully _open_ LLMs 🔥open weight 🔥open training data 🔥open code: data processing, training, eval 🔥open training trajectory: 360 checkpoints - from token 0 to 1.4T 🔥open analysis Excited to see how this'd fuel research to transparentize LLMs🔍🔍
Tweet media one
Tweet media two
@_akhaliq
AK
10 months
LLM360: Towards Fully Transparent Open-Source LLMs paper page: The recent surge in open-source Large Language Models (LLMs), such as LLaMA, Falcon, and Mistral, provides diverse options for AI practitioners and researchers. However, most LLMs have only
Tweet media one
2
95
366
9
51
215
@ZhitingHu
Zhiting Hu
5 years
XLNet Pytorch for generation is now ready in Texar-pytorch! Play with XLNet & GPT-2 here. Figs show how XLNet and GPT-2 think of each other (w/ nucleus decoding) — It looks XLNet is looking forward GPT-3 which must be supervised :)
Tweet media one
Tweet media two
@ZhitingHu
Zhiting Hu
5 years
Checkout XLNet Pytorch implementation in Texar-pytorch (alpha), for encoding, classification, regression, (generation coming soon), and composing downstream models in combination of the rich Texar ML modules & functions! @ZihangDai @rsalakhu @quocleix
3
13
110
4
65
207
@ZhitingHu
Zhiting Hu
10 months
🔥 The slides (both pptx & pdf) are now available on the tutorial website 💥Thanks all participants & panelists for making it a successful tutorial and discussion! Also check out the accompanying perspective/review paper:
Tweet media one
Tweet media two
Tweet media three
Tweet media four
@ZhitingHu
Zhiting Hu
10 months
🔥 #NeurIPS2023 Tutorial🔥 Language Models meet World Models @tianminshu & I are excited to give tutorial on machine reasoning by connecting LLMs🗣️ world models🌎 agent models🤖 w/ amazing panelists @jiajunwu_cs @du_yilun Ishita Dasgupta,Noah Goodman
Tweet media one
5
71
411
3
33
183
@ZhitingHu
Zhiting Hu
4 years
New work connects GAN with Reinforcement Learning under a variational perspective, and stabilizes GAN training w/ off-the-shelf RL techniques Strongly improves image generation, text generation, text style transfer paper code 1/
Tweet media one
1
38
160
@ZhitingHu
Zhiting Hu
3 years
News🎺: A "Standard Equation" of ML that unifies many learning paradigms & algorithms, is online! We hope it can serve as a vehicle🚗 towards "panoramic learning" -- learning AI agents w/ ALL experiences check out the initial draft feedback welcomed 1/
Tweet media one
2
32
152
@ZhitingHu
Zhiting Hu
1 year
⚒️ToolkenGPT 🔥Now any LMs can use massive tools, no finetuning, no prompt length limit 💡Calling a tool is as natural as generating a word token--treat tools as token (“toolken”) embeddings. Expand toolset by plug more toolkens 🤔Can embed millions of tools for LMs in future?
Tweet media one
@Ber18791531
Shibo Hao
1 year
🤔ChatGPT plug-in is impressive, but is "learning tools in the context" the ultimate solution? Check out our ToolkenGPT (), which can handle massive tools and understand them better with new "toolken" embeddings.
10
29
209
2
21
138
@ZhitingHu
Zhiting Hu
5 years
Checkout XLNet Pytorch implementation in Texar-pytorch (alpha), for encoding, classification, regression, (generation coming soon), and composing downstream models in combination of the rich Texar ML modules & functions! @ZihangDai @rsalakhu @quocleix
3
13
110
@ZhitingHu
Zhiting Hu
6 years
Sad to be missing #NeurIPS18 cuz visa (missed ICML18 for the same) But pls come meet my coauthors Deep Generative Models with Learnable Knowledge Constraints Hu, Yang, @rsalakhu , Liang, Qin, Dong, Xing Sorry for the dense (but structured:) poster-just wanna make it self-explained
Tweet media one
2
17
110
@ZhitingHu
Zhiting Hu
5 months
LLM Reasoners have been widely used (800 GitHub stars!🌟) in many projects since its initial release. Now v1.0 is out -- new _algorithms_, systematic _evaluation_, and _visualization_! 🤩 More SOTA reasoning algorithms are on the way, including LLM reasoning for scientific
@MaitrixOrg
Maitrix.org
5 months
Releasing 🔥LLM Reasoners v1.0🔥 🥇Popular library for advanced LLM reasoning - Reasoning-via-Planning (RAP)🎶 - Chain-of-Thoughts (CoT)⛓️ - Tree-of-Thoughts (ToT)🌴 - Grace decoding💄 - Beam search🔎 🥇Enhances #Llama3 , GPT4, LLMs on @huggingface
Tweet media one
2
61
206
0
18
105
@ZhitingHu
Zhiting Hu
10 months
Tutorial just started! A packed room
Tweet media one
Tweet media two
Tweet media three
@ZhitingHu
Zhiting Hu
10 months
🔥 #NeurIPS2023 Tutorial🔥 Language Models meet World Models @tianminshu & I are excited to give tutorial on machine reasoning by connecting LLMs🗣️ world models🌎 agent models🤖 w/ amazing panelists @jiajunwu_cs @du_yilun Ishita Dasgupta,Noah Goodman
Tweet media one
5
71
411
1
12
100
@ZhitingHu
Zhiting Hu
4 years
New work “Progressive Generation of Long Text” — a super simple “non-monotonic” use of monotonic language models (eg. GPT2, BART) for generating coherent long text (1000 tokens) paper: code: 1/4
Tweet media one
2
14
79
@ZhitingHu
Zhiting Hu
10 months
🚨Oral presentation at #NeurIPS2023 today by Shibo @Ber18791531 ToolkenGPT is a way for LLMs to interact with the world🌎 via extensive plug-n-play tools⚒️ Two key takeaways: 🥇capture tool semantics by learning embeddings 🥈call the tools simply as generating words
Tweet media one
@ZhitingHu
Zhiting Hu
1 year
⚒️ToolkenGPT 🔥Now any LMs can use massive tools, no finetuning, no prompt length limit 💡Calling a tool is as natural as generating a word token--treat tools as token (“toolken”) embeddings. Expand toolset by plug more toolkens 🤔Can embed millions of tools for LMs in future?
Tweet media one
2
21
138
1
13
72
@ZhitingHu
Zhiting Hu
8 months
It's interesting to see how Sora triggers so much discussion on world models🌎 Does Sora understand Physics? Is Sora a world model? It's not a True/False question. The real questions are: ❓How capable is video (pixels) to represent the world? ❓And how efficient is it? ❓So
@ylecun
Yann LeCun
8 months
Modeling the world for action by generating pixel is as wasteful and doomed to failure as the largely-abandoned idea of "analysis by synthesis". Decades ago, there was a big debate in ML about the relative advantages of generative methods vs discriminative methods for
120
233
2K
6
15
76
@ZhitingHu
Zhiting Hu
2 years
Phew.. What a game!
Tweet media one
1
0
71
@ZhitingHu
Zhiting Hu
4 months
Nothing is better than ending the #Spring quarter with a thank-you letter from student! ❤️🏆🏆
Tweet media one
0
0
64
@ZhitingHu
Zhiting Hu
4 years
This Sunday at #KDD2020 , I'll be giving a tutorial on "Learning from ALL Experiences: A Unifying Machine Learning Perspective" w/ Eric Xing & Qirong Ho. We'll present a systematic unified perspective of machine learning with ALL Experiences, ranging from ... @kdd_news 1/
Tweet media one
2
14
60
@ZhitingHu
Zhiting Hu
5 years
We're organizing #NeurIPS2019 workshop on Learning with Rich Experience: Integration of Learning Paradigms, w/ an amazing lineup of speakers! Deadline: Sept 11 w/ @andrewgwils @chelseabfinn @rl_agent @Lianhuiq Taylor Berg-Kirkpatrick @rsalakhu & Eric Xing
0
16
47
@ZhitingHu
Zhiting Hu
11 months
Thrilled to receive the #SoCalNLP2023 award! Shoutout to @kaiwei_chang @robinomial @jieyuzhao11 and all @socalnlp organizers for the amazing event at UCLA!! 🥚Easter egg: next year #SoCalNLP is heading to UC San Diego. Looking fwd to it already! #NLProc
@socalnlp
SoCal NLP Symposium
11 months
Next, we have @Ber18791531 et. al’s amazing work, that also won an award at #SoCalNLP2023 ! 👏🏼
Tweet media one
1
2
10
0
4
48
@ZhitingHu
Zhiting Hu
5 years
The paper “transfers” an off-the-shelf _reward_ learning algorithm to learning _data_ manipulation It’s a powerful idea--transferring solutions to problems in one context to problems in another. Used in learning structured knowledge , improving GANs/VAEs
@rsalakhu
Russ Salakhutdinov
5 years
#NeurIPS2019 paper on Learning Data Manipulation: Learning to augment and re-weight data for improved training, especially in low data regime or in presence of imbalanced labels. w/t Zhiting Hu, Bowen Tan et. al.
1
68
272
1
11
47
@ZhitingHu
Zhiting Hu
6 years
The full schedule of the ICML2018 workshop “Theoretical Foundations and Applications of Deep Generative Models” is released: . Come join us on Saturday 8:30 - 18:00 & Sunday 8:30 - 12:30 @ Room A5!
Tweet media one
Tweet media two
Tweet media three
0
14
42
@ZhitingHu
Zhiting Hu
4 months
Consistency and controllability are two foundation capabilities of a World Model.🌏 > Consistency is to generate consistent videos (or other modalities) to correctly describe the world. This requires understanding of how the world works -- physics, spatiotemporal dynamics,
@ZhitingHu
Zhiting Hu
4 months
Super excited to introduce Pandora, a generative video World Model interactively controllable by language. #Sora and #GPT4 are both powerful. How about fusing them in a single model? 💥 Pandora gives a preview:🔭 > Build a General World Model (GWM) super efficiently by
9
58
290
3
8
41
@ZhitingHu
Zhiting Hu
2 years
#EMNLP2022 Discrete pompt is more robust, interpretable, and transferrable-across-LMs than soft prompt But optimizing discrete prompt is difficult. Now you can do so efficiently and neatly with Reinforcement Learning! Checkout *RLPrompt*:
@mdeng34
Mingkai Deng
2 years
#EMNLP2022 RLPrompt uses Reinforcement Learning to optimize *discrete* prompts for any LM w/ many nice properties: * prompts transfer trivially across LMs * gradient-free for LM * strong improve vs manual/soft prompts Paper Code
Tweet media one
3
22
171
0
7
40
@ZhitingHu
Zhiting Hu
2 months
Really excited about the #ACL2024 outstanding paper award to our work on Multi-Modal Theory-of-Mind evaluation! Congrats to @chuanyang_jin and @tianminshu who led the work! Check out more: - -
@MaitrixOrg
Maitrix.org
2 months
🥳Our work on Multi-Model Theory of Mind evaluation won the #ACL2024 Outstanding Paper Award! How well can machines 🤖 form a coherent mental picture🧠of humans from vision-language observations? Can machines understand humans' goals and beliefs? Our MMToM-QA shows models
0
14
29
0
10
43
@ZhitingHu
Zhiting Hu
6 years
Our ICML18 workshop: Theoretical Foundations and Applications of Deep Generative Models.. Having great speakers coming! Please consider submitting your work, and participating to share your thoughts!
@rsalakhu
Russ Salakhutdinov
7 years
Call for Participation: ICML 2018 workshop on Theoretical Foundations and Applications of Deep Generative Models.
Tweet media one
0
22
96
0
4
36
@ZhitingHu
Zhiting Hu
10 months
Reasoing via Planning (RAP): - "Combining AlphaGo-style search and LLM" - "Add world models … natively" :)
@DrJimFan
Jim Fan
10 months
@ylecun I agree, the fear of "AGI achieved by Q*" is nonsense. Combining AlphaGo-style search and LLM is an effective way to tackle specific domains like math and coding with groundtruth signals. But we do need novel ways to add world models and embodied agent capabilities natively to
13
19
320
0
6
33
@ZhitingHu
Zhiting Hu
4 months
🔥Excited that Redcoast won the #NAACL2024 best demo runner up! Redcoast is a super easy-to-use tool ☕️ for automated distributed training of #LLMs , diffusion, reinforcement learning, meta-learning, etc. Users just write three functions (collate, loss, predict), and Redcoast
Tweet media one
@naaclmeeting
NAACL HLT 2025
4 months
Some of the award winners in this edition of #NAACL2024
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
2
47
2
7
34
@ZhitingHu
Zhiting Hu
9 months
🔮CrystalChat-7B 🔥 best balanced capabilities in language and code v.s. - Mistral-7B-Instruct - CodeLlama-7B-Instruct And don't forget its most facinating feature: as part of @llm360 , it's _fully_ open!
@llm360
LLM360
9 months
1/3 We are releasing CrystalChat 🔮 — a top-scoring 7B chat model, fully open source! As always, CrystalChat is released under Apache 2.0, along with all training data, checkpoints, and implementation details. Grab the model here:
Tweet media one
4
34
159
0
6
27
@ZhitingHu
Zhiting Hu
8 months
For more discussions -- I'll be giving a tutorial at #aaai2024 tmr (Tuesday) morning with @tianminshu : Language Models meet World Models It's an extended version of the #NeurIPS2023 Tutorial:
@ZhitingHu
Zhiting Hu
10 months
🔥 #NeurIPS2023 Tutorial🔥 Language Models meet World Models @tianminshu & I are excited to give tutorial on machine reasoning by connecting LLMs🗣️ world models🌎 agent models🤖 w/ amazing panelists @jiajunwu_cs @du_yilun Ishita Dasgupta,Noah Goodman
Tweet media one
5
71
411
0
8
29
@ZhitingHu
Zhiting Hu
3 months
Optimizing pages-long expert-level prompts automatically 👇 It's fascinating that _prompt optimization_ can be formulated as a _planning_ problem: - Treat the LLM as a world model🌎 - We want a prompt, as a plan trajectory, that thrives in this world - So we do strategic
@MaitrixOrg
Maitrix.org
3 months
"With long context LLMs comes long prompts"👇 People typically just write 1- or 2-sentence quick prompts when using an LLM for a task. How to create 1- or 2-page long prompts to boost performance? 🔥PromptAgent automatically writes long prompts for you!🔥 Without need of the
Tweet media one
0
9
24
0
8
27
@ZhitingHu
Zhiting Hu
5 years
Super excited to release Texar-PyTorch v0.1 An ML library integrating the best of TensorFlow into PyTorch - replicating many useful TF modules & designs to enhance PyTorch, incl. data, model & training. See how Texar-Pytorch builds a Conditional-GPT2 1/5
Tweet media one
1
5
26
@ZhitingHu
Zhiting Hu
5 years
Texar is nominated as a best (demo) paper candidate #acl2019nlp . Key features of Texar as a toolkit for ML & NLP/text generation: 1) Two versions (TF, PyTorch), mostly same interfaces; 2) Rich pretrained models (GPT2, BERT, XLNet…), unified interfaces, rich usage; 1/2
@ACL2019_Italy
ACL2019
5 years
We are delighted to announce the list of papers that have been nominated as candidates for ACL 2019 Best Paper Awards! Check the list at #acl2019nlp
1
34
167
1
4
25
@ZhitingHu
Zhiting Hu
8 months
Nice discussion. There could be a slightly more general perspective on "retrieval": LLM generating each token can be seen as "retrieval" -- its last layer produces a latent vector. The vector is used to compare w/ word embeddings to "retrieve" the most similar word as the next
@Francis_YAO_
Yao Fu
8 months
Over the last two days after my claim "long context will replace RAG", I have received quite a few criticisms (thanks and really appreciated!) and many of them stand a reasonable point. Here I have gathered the major counterargument, and try to address then one-by-one (feels like
35
88
511
0
4
25
@ZhitingHu
Zhiting Hu
4 years
Come stop by our poster (Wed 9am PT): - A unifying variational perspective of GANs - Stabilizing GAN training with off-the-shelf Reinforcement Learning techniques paper: code:
@ZhitingHu
Zhiting Hu
4 years
#NeurIPS2020 paper/code: "free lunch" for stabilizing your #GAN training & improving image/text generation With a unifying perspective of GANs, RL & inference, we can easily import the stabilized RL algo #PPO to stabilize GAN. Also importance sampling to stabilize discriminator
0
5
20
0
10
23
@ZhitingHu
Zhiting Hu
6 years
#Texar now supports GPT-2 language models. Using GPT-2 on Texar is as simple as creating a TransformerDecoder instance and loading pre-trained weights. Texar supports many operations, like: (1/2)
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
5
22
@ZhitingHu
Zhiting Hu
5 years
Come join us for tmr (Sunday) morning’s CVPR workshop “Towards Causal, Explainable and Universal Medical Visual Diagnosis" (8:30am-12:00pm) at Hyatt Beacon B Room (4th floor of Hyatt Regency, right next to the Convention Center).
Tweet media one
Tweet media two
1
4
22
@ZhitingHu
Zhiting Hu
4 months
Check out #K2 , a fully-open 65B LLM released by @llm360 Matching the performance of #Llama2 70B, #K2 is among the most powerful LLMs made fully transparent! Over the past 6 months, @llm360 has made open a series of LLMs across different tiers, all with open weights,
@llm360
LLM360
4 months
Please welcome K2-65B🏔️, the most performant fully-open LLM released to date. As a blueprint for open-source AGI, we release all model checkpoints, code, logs, and data. About K2: 🧠65 billion parameters 🪟Fully transparent & reproducible 🔓Apache 2.0 📈Outperforms Llama 2 70B
6
146
492
0
4
21
@ZhitingHu
Zhiting Hu
6 years
Blog article introducing Texar, an open-source general-purpose text generation toolkit: We will keep posting news, latest research, tutorials, and other information related to Texar on the Medium page:
0
7
22
@ZhitingHu
Zhiting Hu
5 years
Besides delving deeper into each learning paradigm, it’s also beneficial to look broader--studying connections bw Algs, combining best & enabling new capability (eg ingesting rich supervisions). Submit ur work in this line to #NeurIPS2019 workshop --
3
6
21
@ZhitingHu
Zhiting Hu
4 years
#NeurIPS2020 paper/code: "free lunch" for stabilizing your #GAN training & improving image/text generation With a unifying perspective of GANs, RL & inference, we can easily import the stabilized RL algo #PPO to stabilize GAN. Also importance sampling to stabilize discriminator
@ZhitingHu
Zhiting Hu
4 years
New work connects GAN with Reinforcement Learning under a variational perspective, and stabilizes GAN training w/ off-the-shelf RL techniques Strongly improves image generation, text generation, text style transfer paper code 1/
Tweet media one
1
38
160
0
5
20
@ZhitingHu
Zhiting Hu
5 years
Come join the #NeurIPS2019 workshop on Learning with Rich Experience. Note the location: West 208+209. Look fwd to the super exciting talks by @RaiaHadsell @tommmitchell JeffBilmes @pabbeel @YejinChoinka & TomGriffiths, and the contributed presentations:
Tweet media one
0
6
20
@ZhitingHu
Zhiting Hu
5 years
Our work Connecting the Dots b/w MLE and RL won the best paper on #ICLR workshop Deep RL meets Structured Prediction (happening now @ R02). Joint work w bowen, zichao, Russ @rsalakhu , Eric
@audurand
Audrey Durand
5 years
Looking forward for our #ICLR2019 workshop on Deep RL Meets Structured Prediction! Thanks to our amazing invited speakers @jhamrick , @AnimaAnandkumar , @gneubig , and @Mo_Norouzi ! Also featuring contributed talks by @ZhitingHu , @wouterkool , @zafarali , and Osbert Bastani!
3
6
42
1
1
20
@ZhitingHu
Zhiting Hu
6 years
#Texar now supports pretrained BERT models. Using BERT on Texar is as simple as creating a TransformerEncoder (fig) and loading the pretrained parameters. Then you can build downstream models for both text understanding & generation tasks. Example here:
Tweet media one
1
5
18
@ZhitingHu
Zhiting Hu
5 years
Schedule for the upcoming #CVPR2019 workshop "Towards Causal, Explainable and Universal Medical Visual Diagnosis" is out (Jun 16, Sun). Looking forward to the amazing speakers: Russ @rsalakhu , Devi @deviparikh , Dina @dina_katabi , Le Lu, & Deva Ramanan
1
5
17
@ZhitingHu
Zhiting Hu
1 year
This is our 2nd recent study of connecting the concepts of 🤖Language Models and 🌎World Models Check out the other one:
@ZhitingHu
Zhiting Hu
1 year
🗣️Language Models ➕ 🌎World Models When the two fascinating models start to interplay with each other, lots of exciting things happen! 🔥Here we use WM to teach LM diverse embodied knowledge and skills. Improve by 64% on 18 tasks, and let GPT-J-6B surpass chatGPT! 🔥
Tweet media one
2
48
221
1
3
16
@ZhitingHu
Zhiting Hu
3 years
#NAACL2021 Progressive Generation of Long Text with Pretrained LMs Generating 1000-token text by converting left-to-right language models (GPT-2, BART) into progressive generators Nice work by @LTIatCMU student Bowen! paper code
Tweet media one
@ZhitingHu
Zhiting Hu
4 years
New work “Progressive Generation of Long Text” — a super simple “non-monotonic” use of monotonic language models (eg. GPT2, BART) for generating coherent long text (1000 tokens) paper: code: 1/4
Tweet media one
2
14
79
1
4
16
@ZhitingHu
Zhiting Hu
3 months
@srush_nlp We have a new work at #ICML2024 to learn latent auto-encoding for text. It improves VAEs and other DGMs quite a bit: The idea is to augment diffusion with parameterized encoder-decoder (e.g., pretrained LLMs) Or in an alternative view, it replaces the
4
4
17
@ZhitingHu
Zhiting Hu
5 years
@rsalakhu giving talk about Incorporating Domain Knowledge into Deep Learning!
Tweet media one
@ZhitingHu
Zhiting Hu
5 years
Dr Le Lu is giving an exciting talk about Deep Learning and Big Data Exploration for Preventive and Precision Medicine in Radiology @ Hyatt Beacon B #CVPR2019 Medical Visual Diagnosis workshop
Tweet media one
0
0
1
0
3
16
@ZhitingHu
Zhiting Hu
3 months
Fascinating idea of formulating LLM divergent thinking as sampling reasoning paths _proportional_ to reward functions (instead of just _maximizing_ reward)
@Lianhuiq
Lianhui Qin
3 months
💡Divergence thinking💡 is a hallmark of human creativity and problem-solving 🤖Can LLMs also do divergent reasoning to generate diverse solutions🤔? Introducing Flow-of-Reasoning (FoR) 🌊, a data-efficient way of training LLM policy to generate diverse, high-quality reasoning
7
68
240
0
2
16
@ZhitingHu
Zhiting Hu
5 years
Our #NeurIPS2019 workshop on Learning with Rich Experience (LIRE) is now accepting late-breaking submissions! Due next Monday 09/30. @andrewgwils @chelseabfinn @rl_agent @Lianhuiq @rsalakhu
@ZhitingHu
Zhiting Hu
5 years
Besides delving deeper into each learning paradigm, it’s also beneficial to look broader--studying connections bw Algs, combining best & enabling new capability (eg ingesting rich supervisions). Submit ur work in this line to #NeurIPS2019 workshop --
3
6
21
0
6
16
@ZhitingHu
Zhiting Hu
5 years
Towards Grounded, Explainable Vision + Language Models by @deviparikh !
Tweet media one
@ZhitingHu
Zhiting Hu
5 years
@rsalakhu giving talk about Incorporating Domain Knowledge into Deep Learning!
Tweet media one
0
3
16
0
5
16
@ZhitingHu
Zhiting Hu
2 years
🚨Announcing ICLR 2023 workshop🚨on Machine Learning for IoT: Datasets, Perception & Understanding We hope to better bridge the ML and IoT communities, discuss opportunities& challenges of ML on vastly heterogeneous IoT data/scenarios #ICLR2023 #ML4IoT
Tweet media one
2
2
15
@ZhitingHu
Zhiting Hu
6 years
BERT shares the key idea and almost the same model/learning objective with our recent work <Text Infilling>. BERT emphasizes that *text representation learning* should model both left and right context. Text Infilling emphasizes that *text generation* should also model both.
Tweet media one
1
2
12
@ZhitingHu
Zhiting Hu
6 years
Super excited to announce Texar, an open source toolkit led by CMU and Petuum, aiming to support a broad set of text generation tasks like machine translation, dialog, summarization, content manipulation, etc. More details on . @mldcmu @LTIatCMU @PetuumInc
Tweet media one
1
9
13
@ZhitingHu
Zhiting Hu
5 years
Awesome! We once evaluated text generative models by training classifiers on real & model-synthetic data; observed certain improvement (vs classifiers on only real data). The real datasets were small though (). Great to see systematic studies here on images
@OriolVinyalsML
Oriol Vinyals
5 years
Evaluating generative models is hard! We propose Classification Accuracy Score from classifiers trained on generated data: -Accuracy of 43% when trained purely on BigGAN samples (vs 73%) -Naive data augmentation doesn't work (yet!) Paper: cc @SumanRavuri
Tweet media one
3
69
245
0
2
13
@ZhitingHu
Zhiting Hu
5 months
Welcome @SnowflakeDB Arctic to the family of fully open LLMs! ❄️❄️
@llm360
LLM360
5 months
❄️Congrats to @SnowflakeDB for openly releasing Arctic!❄️ Arctic is available to all with an Apache 2.0 license! Great to see LLM360 member @AurickQ and the whole Snowflake AI Research’s team's amazing contribution to the open-source LLM community!
2
3
16
2
1
13
@ZhitingHu
Zhiting Hu
5 years
DON’T MISS the exciting panel by our fantastic speakers, 17:05 @West 208+209!
Tweet media one
@ZhitingHu
Zhiting Hu
5 years
Come join the #NeurIPS2019 workshop on Learning with Rich Experience. Note the location: West 208+209. Look fwd to the super exciting talks by @RaiaHadsell @tommmitchell JeffBilmes @pabbeel @YejinChoinka & TomGriffiths, and the contributed presentations:
Tweet media one
0
6
20
0
1
11
@ZhitingHu
Zhiting Hu
2 years
A nice Youtube video explaining RLPrompt: (seems like an awesome Youtube channel summarizing interesting NLP works with quick videos)
@mdeng34
Mingkai Deng
2 years
#EMNLP2022 RLPrompt uses Reinforcement Learning to optimize *discrete* prompts for any LM w/ many nice properties: * prompts transfer trivially across LMs * gradient-free for LM * strong improve vs manual/soft prompts Paper Code
Tweet media one
3
22
171
0
4
11
@ZhitingHu
Zhiting Hu
3 years
Super interesting workshop on "Machine Learning for Data" @icmlconf For ML-automated data operations (labeling, synthesis, selection, augmentation), and challenges in quality, bias, security, privacy, … Submission deadline: June 10
@ml4data
ml4data
3 years
🚨Announcing the ML4data workshop at ICML 2021🚨 — a workshop focused on how we use ML for our most precious resource: data #icml2021 #icml21 #ml4data
Tweet media one
1
12
40
0
0
10
@ZhitingHu
Zhiting Hu
6 years
New work “Toward Unsupervised Text Content Manipulation” . Language is rich with variation—given a data record, there are diverse possible ways of saying the same data, with different word choices, expressions, transitions, tones, etc (i.e., writing style).
1
5
11
@ZhitingHu
Zhiting Hu
3 years
🚨New efficient Reinforcement Learning for text generation! For various emerging apps: * Generating _prompt_ to control pretrained LMs * Generating _adversarial_ examples * Learning w/ _negative_ text * Training from scratch Nice work by @LTIatCMU student @HanGuo97
Tweet media one
@HanGuo97
Han Guo
3 years
Excited to share our latest work with Bowen Tan @waterluffy Eric Xing @ZhitingHu ! Tldr, a new NLG formulation from soft Q-learning perspective, with app. such as learning from noisy data, text attacks, prompt generation. Paper Code
Tweet media one
5
26
118
0
1
10
@ZhitingHu
Zhiting Hu
5 years
@apsdehal "ALBERT - ensemble of ALL BERT in the world" ;)
0
0
9
@ZhitingHu
Zhiting Hu
7 months
Controllable text generation leads to richer/better attacks for LLM safety
@Lianhuiq
Lianhui Qin
7 months
📢Introducing ❄️COLD-Attack⚔️, a unified framework for controllable jailbreaking of LLMs. Thanks to the controllability, COLD-Attack enables new jailbreak scenarios that are hard to detect🧐: 1⃣revising a user query adversarially with minimal paraphrasing 2⃣inserting stealthy
Tweet media one
5
48
236
0
0
9
@ZhitingHu
Zhiting Hu
2 years
We present a standardized formalism of the objective function to characterize the learning of any model with any experience: eg. data, constraints, rules, reward, adversaries, other models, evolving interacts It results in a succinct equation of the objective, w/ 3 terms: 2/
Tweet media one
1
0
9
@ZhitingHu
Zhiting Hu
3 years
Big congrats to the authors of Pollux/AdaptDL @AurickQ and to the CASL (Composability, Automatic and Scalable Learning) open source consortium
@ericxing
Eric Xing
3 years
I am thrilled to share that our paper "Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep Learning" has won the Jay Lepreau Best Paper Award at 15th USENIX Symposium on Operating Systems Design and Implementation ( #OSDI '21).
4
3
58
0
0
8
@ZhitingHu
Zhiting Hu
2 years
Paper: pdf: It's a (substantial) extension to the earlier version we presented last year: 12/12
@ZhitingHu
Zhiting Hu
3 years
News🎺: A "Standard Equation" of ML that unifies many learning paradigms & algorithms, is online! We hope it can serve as a vehicle🚗 towards "panoramic learning" -- learning AI agents w/ ALL experiences check out the initial draft feedback welcomed 1/
Tweet media one
2
32
152
0
0
9
@ZhitingHu
Zhiting Hu
1 year
The #ICLR workshop of _Machine Learning for IoT_ is happening now!
@ZhitingHu
Zhiting Hu
2 years
🚨Announcing ICLR 2023 workshop🚨on Machine Learning for IoT: Datasets, Perception & Understanding We hope to better bridge the ML and IoT communities, discuss opportunities& challenges of ML on vastly heterogeneous IoT data/scenarios #ICLR2023 #ML4IoT
Tweet media one
2
2
15
0
1
4
@ZhitingHu
Zhiting Hu
5 months
Thank you Martin @ziqiao_ma for the nice lecture! We learned a lot.
@ziqiao_ma
Martin Ziqiao Ma
5 months
Excited to guest a lecture on “Connecting Language to the World: Towards Language Grounding to Vision and Embodied Agents” at UCSD. 😇 Thanks Prof. @ZhitingHu for hosting!
Tweet media one
2
3
61
0
1
7
@ZhitingHu
Zhiting Hu
2 years
Mingkai has been making amazing works like the unified framework for language generation evaluation Looking forward to more exciting results coming from him!
@mdeng34
Mingkai Deng
2 years
Time to go public: After 2 happy years at @mldcmu , I’ll be hopping downstairs this fall to @LTIatCMU for PhD with @ericxing , continuing the work on ML+NLP. Examples include text generation from diverse supervision signals (1/2)
5
5
83
0
1
7
@ZhitingHu
Zhiting Hu
2 years
Check out our great speaker lineup! Kirsten Grauman ( @KristenGrauman7 ), Nicholas Lane, Pradeep Natarajan ( @AmazonScience ), Thomas Plotez ( @thomasploetz ), Russ Salakhutdinov ( @rsalakhu ), Dawn Song ( @dawnsongtweets ), Eric Xing ( @ericxing ), and Heather Zheng ( @heatherzheng )
1
0
6
@ZhitingHu
Zhiting Hu
5 years
New #ACL2019 work "Target-Guided Open-Domain Conversation" . Going beyond chit-chat, we impose conversational goals on open-domain chat agents--the agent needs to chat engagingly with user and *proactively guide* the conversation to a given target subject.
1
1
6
@ZhitingHu
Zhiting Hu
10 months
@BlancheMinerva Ah you're right! I somehow had a wrong impression. Bloom/Pythia etc opened similar artifacts earlier, and LLM360 is a fresh addition to these nice efforts.
1
0
6
@ZhitingHu
Zhiting Hu
5 years
Learning w Rich Experience: Integration of Learning Paradigms. Due Sept 11 Don't miss the incredible speakers @pabbeel JeffreyBilmes @YejinChoinka TomGriffiths @RaiaHadsell @tommmitchell w @andrewgwils @chelseabfinn @rl_agent @Lianhuiq TaylorBerg-Kirkpatrick @rsalakhu EricXing
1
2
5
@ZhitingHu
Zhiting Hu
3 years
Automatic prompt tuning & generation for controllable text generation How to🤔? Make it a Reinforcement Learning problem
@HanGuo97
Han Guo
3 years
One (unexpected!) use case of the algorithm is prompt generation for controlling pretrained LMs via an RL formulation with "prompt" as action and "outcome of prompted outputs" as reward. Empirically, this seems to be more effective/efficient than recent steered decoding methods.
Tweet media one
Tweet media two
0
2
19
0
2
5
@ZhitingHu
Zhiting Hu
3 years
It's appealing to see that a succinct equation of the objective function recovers a wide range of known algorithms in distinct paradigms (supervised, unsupervised, active, reinforcement, adversarial learning, etc) The equation consists of three terms: 2/
Tweet media one
1
0
5
@ZhitingHu
Zhiting Hu
6 years
* Sample generation or completion * Greedy / (top-k) sample / Gumbel-softmax / beam-search / your-customized decoding * Training / fine-tuning in (un)conditional settings * Perplexity evaluation etc Checkout example code/demo here: . (2/2)
Tweet media one
2
5
4
@ZhitingHu
Zhiting Hu
2 years
Besides, the overall formalism is agnostic to the type of model p, meaning that you can plug in and train arbitrary 'model' for your task, eg - Transformers - pretrained models - discrete/soft prompts - symbolic knowledge graph (eg. ) - world model 8/
1
0
5
@ZhitingHu
Zhiting Hu
10 months
Correction: there are indeed earlier LLMs (e.g., GPT-J/Bloom/Pythia) at a similar level of openness, as shown in the table. LLM360 is a fresh addition to these great efforts. Look forward to more openness and transparency in this realm!
0
0
5
@ZhitingHu
Zhiting Hu
1 year
@cwizprod1 LM is a type of world model because language itself is a representation of the world. The thing is how to make the WM explicit and help reasoning. LM has limited knowledge of the world (ie., a limited WM). So another topic is to teach LM more. One of our attempts here:)
@ZhitingHu
Zhiting Hu
1 year
🗣️Language Models ➕ 🌎World Models When the two fascinating models start to interplay with each other, lots of exciting things happen! 🔥Here we use WM to teach LM diverse embodied knowledge and skills. Improve by 64% on 18 tasks, and let GPT-J-6B surpass chatGPT! 🔥
Tweet media one
2
48
221
1
0
4
@ZhitingHu
Zhiting Hu
6 years
We go beyond traditional data-to-text tasks by controlling both content and writing style of generation—we want the output to describe a given data record, meanwhile following the writing style of a reference sentence. Check out the paper for dataset/methods. Code coming soon
0
1
4
@ZhitingHu
Zhiting Hu
2 years
It turns out that, by choosing appropriate experience f, divergence D, and weights alpha/beta, the equation recovers many of the most popular algorithms precisely as special cases, ranging from MLE, active learning, to posterior regularization, to policy gradient, to ... 5/
Tweet media one
1
1
4
@ZhitingHu
Zhiting Hu
1 year
@MichaelTontchev ToT is a nice work! We discuss connect/diff in RelatedWork Both r tree reasoning. ToT use depth/breath-first search; RAP use MontoCarloTreeSearch, balancing exploration-exploitation in more pincipled way This's enabled by introducing WM & reward, etc
0
1
4
@ZhitingHu
Zhiting Hu
4 years
@gneubig @hiroakiLhayashi Nice work! Aspect-based summarization is of great practical use. We also have a (preliminary) work in #emnlp2020 : "Summarizing Text on Any Aspects" w/ knowledge weak supervisions:
0
0
4
@ZhitingHu
Zhiting Hu
5 years
w/ @XiaodanLiang , Christy Li, Hao Wang, Ricardo Henao, Lawrence Carin, Eric Xing @mldcmu @DukeU
0
0
4
@ZhitingHu
Zhiting Hu
3 years
1) Experience function to encode arbitrary types of experiences 2) Divergence term to measure model fitness 3) Uncertainty term to control complexity Different choices of the components (and weights) lead to diverse existing & new algorithms 3/
Tweet media one
1
0
3
@ZhitingHu
Zhiting Hu
5 years
On data part, Texar-PyTorch replicates best practice of for easy processing, batching, iterating + efficiency w/ buffered shuffling, caching, lazy-loading. It also replicates TFRecord to ingest arbitrary complex data dataset, eg, image+caption+label 3/5
Tweet media one
1
0
3
@ZhitingHu
Zhiting Hu
5 years
Read more about Texar-PyTorch features & how easily you can customize any of the above modules for your project, either you're an ML novice or expert More resources: 5/5
1
0
3
@ZhitingHu
Zhiting Hu
4 years
For discriminator, the perspective induces importance re-weighting that downplays low-quality fake samples, leading to significantly more stable training for discriminator (and thus generator). Fig: lower variance in G and D losses throughout training 3/
Tweet media one
1
0
3
@ZhitingHu
Zhiting Hu
3 years
The formalism also opens up exciting open questions/opportunities for future, eg: -Theoretical guarantees of improvement when adding new experience in learning -From standardization to automation: how to automatically search for new algorithms -Learning w/ dynamic experience 5/
2
0
3
@ZhitingHu
Zhiting Hu
2 years
Tweet media one
0
0
3