Xipeng Qiu Profile
Xipeng Qiu

@xpqiu

Followers
294
Following
52
Statuses
20

Natural Language Processing Machine Learning

Shanghai
Joined April 2013
Don't wanna be here? Send us removal request.
@xpqiu
Xipeng Qiu
17 days
Welcome to the SpeechGPT-2.0! #LLM #gpt4o #ACL
@Open_MOSS
OpenMOSS
17 days
🥳 Introducing SpeechGPT 2.0-preview: A GPT-4o-level, real-time spoken dialogue system! (Only Chinese for now) 🎆 Highlights: ~⚡️ Real-time speech-to-speech dialogue with latency under 200ms ~😊 Rich in emotion and diverse in style, with strong speech style generalization ~🦁 Strong role-playing capabilities 🤖️ Try it out: Online system: Github: More demos:
0
1
5
@xpqiu
Xipeng Qiu
1 month
RT @TMarczew: Unraveling the Mystery of OpenAI's o1 #o1 #ReinforcementLearning #ReverseEngineering #AI #LLM A new paper, "Scaling of Sear…
0
1
0
@xpqiu
Xipeng Qiu
2 months
RT @simulately12492: #SimulatelyPapers | December 25, 2024 📄 VLABench: A Large-Scale Benchmark for Language-Conditioned Robotics Manipulat…
0
1
0
@xpqiu
Xipeng Qiu
2 months
RT @yinzhangyue: A Technical Roadmap of o1 from a Reinforcement Learning Perspective Arxiv Link:
0
4
0
@xpqiu
Xipeng Qiu
2 months
RT @Joey_zh_: As someone who values ritual, I’ve decided to release the preview version of my new work, VLABench, on my birthday, which hap…
0
10
0
@xpqiu
Xipeng Qiu
1 year
AnyGPT: The Any-to-Any Multimodal LLM - Audio, Text, and Image! Each modality is a different foreign language.
@_akhaliq
AK
1 year
AnyGPT Unified Multimodal LLM with Discrete Sequence Modeling introduce AnyGPT, an any-to-any multimodal language model that utilizes discrete representations for the unified processing of various modalities, including speech, text, images, and music. AnyGPT can be trained stably without any alterations to the current large language model (LLM) architecture or training paradigms. Instead, it relies exclusively on data-level preprocessing, facilitating the seamless integration of new modalities into LLMs, akin to the incorporation of new languages. We build a multimodal text-centric dataset for multimodal alignment pre-training. Utilizing generative models, we synthesize the first large-scale any-to-any multimodal instruction dataset. It consists of 108k samples of multi-turn conversations that intricately interweave various modalities, thus equipping the model to handle arbitrary combinations of multimodal inputs and outputs. Experimental results demonstrate that AnyGPT is capable of facilitating any-to-any multimodal conversation while achieving performance comparable to specialized models across all modalities, proving that discrete representations can effectively and conveniently unify multiple modalities within a language model.
Tweet media one
0
2
20
@xpqiu
Xipeng Qiu
3 years
RT @ak92501: xFormers: Hackable and optimized Transformers building blocks, supporting a composable construction github:
0
45
0
@xpqiu
Xipeng Qiu
3 years
Yes, it's the ideal architecture for future AI. We also have made some attempts in this direction. Learning Sparse Sharing Architectures for Multiple Tasks
@slashML
/MachineLearning
3 years
Google Research: Introducing Pathways, a next-generation AI architecture
1
0
3
@xpqiu
Xipeng Qiu
3 years
ElasticBERT allows you to more economically use BERT and choose your appropriate layers as needed.
0
0
3
@xpqiu
Xipeng Qiu
4 years
#NLProc A unified generative framework is proposed for aspect-based sentiment analysis (ABSA) to simultaneously solve 7 ABSA subtasks. The model is simple and easy to implement, and achineves SOTA results on the 7 subtasks of three popular ABSA datasets.
Tweet media one
Tweet media two
0
2
8
@xpqiu
Xipeng Qiu
4 years
RT @jure: Sharing slides from my @textgraphs @NAACLHLT workshop keynote: Reasoning with Language and Knowledge Graphs
0
44
0
@xpqiu
Xipeng Qiu
4 years
RT @mohitban47: VALUE = "Video-And-Language Understanding Evaluation"! Strong effort led by @linjiefun & @jielei for this fun collaboration…
0
4
0
@xpqiu
Xipeng Qiu
4 years
RT @omarsar0: A comprehensive overview of Transformer variants. A must-read for students getting into the world of machine learning and NL…
0
242
0
@xpqiu
Xipeng Qiu
4 years
Survey of Transformers, Comments and suggestions welcome!
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
14
60
@xpqiu
Xipeng Qiu
5 years
RT @zibuyu9: An introductory book to graph neural networks, extended from our previous survey paper. Enjoy~😁
0
8
0
@xpqiu
Xipeng Qiu
5 years
RT @WilliamWangNLP: The most comprehensive survey I’ve seen on pretrained language models for #NLProc: https://t.co…
0
122
0
@xpqiu
Xipeng Qiu
5 years
RT @aclmeeting: Here's an update from Dan Jurafsky and the #acl2020nlp team re COVID19: #NLProc
0
39
0
@xpqiu
Xipeng Qiu
6 years
@MSFTResearch @murefil @iclr2019 Congrats! But the basic idea of this paper looks like one of our papers at EMNLP 2016 "Cached Long Short-Term Memory Neural Networks" But the citation is missing.
0
0
3
@xpqiu
Xipeng Qiu
6 years
Congrats! But the basic idea of this paper looks like one of our papers at EMNLP 2016 "Cached Long Short-Term Memory Neural Networks" But the citation is missing.
@MSFTResearch
Microsoft Research
6 years
We're excited to announce that Yikang Shen, Shawn Tan, Alessandro Sordoni @murefil and Aaron Courville received the Best Paper Award at @ICLR2019. Discover their work on Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks: #ICLR2019
Tweet media one
0
0
2
@xpqiu
Xipeng Qiu
6 years
#NLProc Our recent work accepted by NAACL 2019: 1) Star-Transformer: a new lightweight architecture by moving the fully-connected self-attention into a star-shaped structure. 2) VCWE: Visual Character-Enhanced Word Embeddings
Tweet media one
Tweet media two
0
1
3