default search action
Wei Xiong 0015
Person information
- affiliation: University of Illinois Urbana-Champaign, Department of Computer Science, Urbana, IL, USA
- affiliation (former): Hong Kong University of Science and Technology, Hong Kong
Other persons with the same name
- Wei Xiong — disambiguation page
- Wei Xiong 0001 (aka: Xiong Wei 0001) — Institute for Infocomm Research, A*STAR, Singapore
- Wei Xiong 0002 — Xinjiang University, College of Mathematics and System Sciences, Urumqi, China
- Wei Xiong 0003 — University of Florida, Department of Agricultural and Biological Engineering, Gainesville, FL, USA
- Wei Xiong 0004 — Hubei University of Technology, Hubei Collaborative Innovation Center for High-efficient Utilization of Solar Energy, Wuhan, China (and 1 more)
- Wei Xiong 0005 — Sun Yat-Sen University, Department of Philosophy, Institute of Logic and Cognition, Guangzhou, China
- Wei Xiong 0006 — Shanghai Maritime University, College of Information Engineering, China
- Wei Xiong 0007 — Durham University, UK
- Wei Xiong 0008 — University of Rochester, Department of Computer Science, NY, USA (and 1 more)
- Wei Xiong 0009 — Hong Kong University of Science and Technology, Clear Water Bay, Kowloon, Hong Kong
- Wei Xiong 0010 — National University of Defense Technology, Changsha, China
- Wei Xiong 0011 — Case Western Reserve University School of Medicine, Cleveland, Ohio, USA
- Wei Xiong 0012 — Heidelberg University, Institute for Theoretical Physics, Heidelberg, Germany
- Wei Xiong 0013 — University at Albany - SUNY, NY, USA
- Wei Xiong 0014 — The Ohio State University, Columbus, OH, USA
- Wei Xiong 0016 — Tsinghua University, Beijing, China
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c17]Haoxiang Wang, Yong Lin, Wei Xiong, Rui Yang, Shizhe Diao, Shuang Qiu, Han Zhao, Tong Zhang:
Arithmetic Control of LLMs for Diverse User Preferences: Directional Preference Alignment with Multi-Objective Rewards. ACL (1) 2024: 8642-8655 - [c16]Renjie Pi, Tianyang Han, Wei Xiong, Jipeng Zhang, Runtao Liu, Rui Pan, Tong Zhang:
Strengthening Multimodal Large Language Model with Bootstrapped Preference Optimization. ECCV (33) 2024: 382-398 - [c15]Yong Lin, Hangyu Lin, Wei Xiong, Shizhe Diao, Jianmeng Liu, Jipeng Zhang, Rui Pan, Haoxiang Wang, Wenbin Hu, Hanning Zhang, Hanze Dong, Renjie Pi, Han Zhao, Nan Jiang, Heng Ji, Yuan Yao, Tong Zhang:
Mitigating the Alignment Tax of RLHF. EMNLP 2024: 580-606 - [c14]Haoxiang Wang, Wei Xiong, Tengyang Xie, Han Zhao, Tong Zhang:
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts. EMNLP (Findings) 2024: 10582-10592 - [c13]Wei Xiong, Hanze Dong, Chenlu Ye, Ziqi Wang, Han Zhong, Heng Ji, Nan Jiang, Tong Zhang:
Iterative Preference Learning from Human Feedback: Bridging Theory and Practice for RLHF under KL-constraint. ICML 2024 - [c12]Shizhe Diao, Rui Pan, Hanze Dong, Kashun Shum, Jipeng Zhang, Wei Xiong, Tong Zhang:
LMFlow: An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. NAACL (Demonstrations) 2024: 116-127 - [i26]Chenlu Ye, Wei Xiong, Yuheng Zhang, Nan Jiang, Tong Zhang:
A Theoretical Analysis of Nash Learning from Human Feedback under General KL-Regularized Preference. CoRR abs/2402.07314 (2024) - [i25]Haoxiang Wang, Yong Lin, Wei Xiong, Rui Yang, Shizhe Diao, Shuang Qiu, Han Zhao, Tong Zhang:
Arithmetic Control of LLMs for Diverse User Preferences: Directional Preference Alignment with Multi-Objective Rewards. CoRR abs/2402.18571 (2024) - [i24]Renjie Pi, Tianyang Han, Wei Xiong, Jipeng Zhang, Runtao Liu, Rui Pan, Tong Zhang:
Strengthening Multimodal Large Language Model with Bootstrapped Preference Optimization. CoRR abs/2403.08730 (2024) - [i23]Han Zhong, Guhao Feng, Wei Xiong, Li Zhao, Di He, Jiang Bian, Liwei Wang:
DPO Meets PPO: Reinforced Token Optimization for RLHF. CoRR abs/2404.18922 (2024) - [i22]Hanze Dong, Wei Xiong, Bo Pang, Haoxiang Wang, Han Zhao, Yingbo Zhou, Nan Jiang, Doyen Sahoo, Caiming Xiong, Tong Zhang:
RLHF Workflow: From Reward Modeling to Online RLHF. CoRR abs/2405.07863 (2024) - [i21]Haoxiang Wang, Wei Xiong, Tengyang Xie, Han Zhao, Tong Zhang:
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts. CoRR abs/2406.12845 (2024) - [i20]Wei Xiong, Chengshuai Shi, Jiaming Shen, Aviv Rosenberg, Zhen Qin, Daniele Calandriello, Misha Khalman, Rishabh Joshi, Bilal Piot, Mohammad Saleh, Chi Jin, Tong Zhang, Tianqi Liu:
Building Math Agents with Multi-Turn Iterative Preference Learning. CoRR abs/2409.02392 (2024) - [i19]Xuanchang Zhang, Wei Xiong, Lichang Chen, Tianyi Zhou, Heng Huang, Tong Zhang:
From Lists to Emojis: How Format Bias Affects Model Alignment. CoRR abs/2409.11704 (2024) - [i18]Tianqi Liu, Wei Xiong, Jie Ren, Lichang Chen, Junru Wu, Rishabh Joshi, Yang Gao, Jiaming Shen, Zhen Qin, Tianhe Yu, Daniel Sohn, Anastasiia Makarova, Jeremiah Z. Liu, Yuan Liu, Bilal Piot, Abe Ittycheriah, Aviral Kumar, Mohammad Saleh:
RRM: Robust Reward Model Training Mitigates Reward Hacking. CoRR abs/2409.13156 (2024) - 2023
- [j2]Hanze Dong, Wei Xiong, Deepanshu Goyal, Yihan Zhang, Winnie Chow, Rui Pan, Shizhe Diao, Jipeng Zhang, Kashun Shum, Tong Zhang:
RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment. Trans. Mach. Learn. Res. 2023 (2023) - [j1]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Reward Teaching for Federated Multiarmed Bandits. IEEE Trans. Signal Process. 71: 4407-4422 (2023) - [c11]Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, Liwei Wang, Tong Zhang:
Nearly Minimax Optimal Offline Reinforcement Learning with Linear Function Approximation: Single-Agent MDP and Markov Game. ICLR 2023 - [c10]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Provably Efficient Offline Reinforcement Learning with Perturbed Data Sources. ICML 2023: 31353-31388 - [c9]Chenlu Ye, Wei Xiong, Quanquan Gu, Tong Zhang:
Corruption-Robust Algorithms with Uncertainty Weighting for Nonlinear Contextual Bandits and Markov Decision Processes. ICML 2023: 39834-39863 - [c8]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Reward Teaching for Federated Multi-armed Bandits. ISIT 2023: 1454-1459 - [c7]Zhihan Liu, Miao Lu, Wei Xiong, Han Zhong, Hao Hu, Shenao Zhang, Sirui Zheng, Zhuoran Yang, Zhaoran Wang:
Maximize to Explore: One Objective Function Fusing Estimation, Planning, and Exploration. NeurIPS 2023 - [i17]Hanze Dong, Wei Xiong, Deepanshu Goyal, Rui Pan, Shizhe Diao, Jipeng Zhang, Kashun Shum, Tong Zhang:
RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment. CoRR abs/2304.06767 (2023) - [i16]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Reward Teaching for Federated Multi-armed Bandits. CoRR abs/2305.02441 (2023) - [i15]Zhihan Liu, Miao Lu, Wei Xiong, Han Zhong, Hao Hu, Shenao Zhang, Sirui Zheng, Zhuoran Yang, Zhaoran Wang:
One Objective to Rule Them All: A Maximization Objective Fusing Estimation and Planning for Exploration. CoRR abs/2305.18258 (2023) - [i14]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Provably Efficient Offline Reinforcement Learning with Perturbed Data Sources. CoRR abs/2306.08364 (2023) - [i13]Shizhe Diao, Rui Pan, Hanze Dong, Kashun Shum, Jipeng Zhang, Wei Xiong, Tong Zhang:
LMFlow: An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. CoRR abs/2306.12420 (2023) - [i12]Yong Lin, Hangyu Lin, Wei Xiong, Shizhe Diao, Jianmeng Liu, Jipeng Zhang, Rui Pan, Haoxiang Wang, Wenbin Hu, Hanning Zhang, Hanze Dong, Renjie Pi, Han Zhao, Nan Jiang, Yuan Yao, Tong Zhang:
Mitigating the Alignment Tax of RLHF. CoRR abs/2309.06256 (2023) - [i11]Wei Xiong, Hanze Dong, Chenlu Ye, Han Zhong, Nan Jiang, Tong Zhang:
Gibbs Sampling from Human Feedback: A Provable KL- constrained Framework for RLHF. CoRR abs/2312.11456 (2023) - 2022
- [c6]Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, Tong Zhang:
A Self-Play Posterior Sampling Algorithm for Zero-Sum Markov Games. ICML 2022: 24496-24523 - [c5]Han Zhong, Wei Xiong, Jiyuan Tan, Liwei Wang, Tong Zhang, Zhaoran Wang, Zhuoran Yang:
Pessimistic Minimax Value Iteration: Provably Efficient Equilibrium Learning from Offline Datasets. ICML 2022: 27117-27142 - [i10]Han Zhong, Wei Xiong, Jiyuan Tan, Liwei Wang, Tong Zhang, Zhaoran Wang, Zhuoran Yang:
Pessimistic Minimax Value Iteration: Provably Efficient Equilibrium Learning from Offline Datasets. CoRR abs/2202.07511 (2022) - [i9]Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, Liwei Wang, Tong Zhang:
Nearly Minimax Optimal Offline Reinforcement Learning with Linear Function Approximation: Single-Agent MDP and Markov Game. CoRR abs/2205.15512 (2022) - [i8]Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, Tong Zhang:
A Self-Play Posterior Sampling Algorithm for Zero-Sum Markov Games. CoRR abs/2210.01907 (2022) - [i7]Han Zhong, Wei Xiong, Sirui Zheng, Liwei Wang, Zhaoran Wang, Zhuoran Yang, Tong Zhang:
GEC: A Unified Framework for Interactive Decision Making in MDP, POMDP, and Beyond. CoRR abs/2211.01962 (2022) - [i6]Chenlu Ye, Wei Xiong, Quanquan Gu, Tong Zhang:
Corruption-Robust Algorithms with Uncertainty Weighting for Nonlinear Contextual Bandits and Markov Decision Processes. CoRR abs/2212.05949 (2022) - 2021
- [c4]Chengshuai Shi, Haifeng Xu, Wei Xiong, Cong Shen:
(Almost) Free Incentivized Exploration from Decentralized Learning Agents. NeurIPS 2021: 560-571 - [c3]Pushi Zhang, Xiaoyu Chen, Li Zhao, Wei Xiong, Tao Qin, Tie-Yan Liu:
Distributional Reinforcement Learning for Multi-Dimensional Reward Functions. NeurIPS 2021: 1519-1529 - [c2]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Heterogeneous Multi-player Multi-armed Bandits: Closing the Gap and Generalization. NeurIPS 2021: 22392-22404 - [i5]Pushi Zhang, Xiaoyu Chen, Li Zhao, Wei Xiong, Tao Qin, Tie-Yan Liu:
Distributional Reinforcement Learning for Multi-Dimensional Reward Functions. CoRR abs/2110.13578 (2021) - [i4]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Heterogeneous Multi-player Multi-armed Bandits: Closing the Gap and Generalization. CoRR abs/2110.14622 (2021) - [i3]Chengshuai Shi, Haifeng Xu, Wei Xiong, Cong Shen:
(Almost) Free Incentivized Exploration from Decentralized Learning Agents. CoRR abs/2110.14628 (2021) - 2020
- [c1]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Decentralized Multi-player Multi-armed Bandits with No Collision Information. AISTATS 2020: 1519-1528 - [i2]Chengshuai Shi, Wei Xiong, Cong Shen, Jing Yang:
Decentralized Multi-player Multi-armed Bandits with No Collision Information. CoRR abs/2003.00162 (2020) - [i1]Haishan Ye, Wei Xiong, Tong Zhang:
PMGT-VR: A decentralized proximal-gradient algorithmic framework with variance reduction. CoRR abs/2012.15010 (2020)
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2025-01-09 19:26 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint