Weizhu Chen
~Weizhu_Chen1
22
论文总数
11.0
年均投稿
平均评分
接收情况16/22
会议分布
ICLR
12
NeurIPS
6
COLM
3
ICML
1
发表论文 (22 篇)
202510 篇
4
Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment
ICLR 2025Rejected
4
Relative Preference Optimization: Enhancing LLM Alignment through Contrasting Responses across Identical and Diverse Prompts
ICLR 2025Rejected
3
Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
ICLR 2025Poster
3
SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation
COLM 2025Poster
3
LongRoPE2: Near-Lossless LLM Context Window Scaling
ICML 2025Poster
4
Segmenting Text and Learning Their Rewards for Improved RLHF in Language Models
ICLR 2025Rejected
3
SwS: Self-aware Weakness-driven Problem Synthesis in Reinforcement Learning for LLM Reasoning
NeurIPS 2025Poster
4
Scaling Laws of Synthetic Data for Language Model
COLM 2025Poster
4
Reinforcement Learning for Reasoning in Large Language Models with One Training Example
NeurIPS 2025Poster
4
Decoder-Hybrid-Decoder Architecture for Efficient Reasoning with Long Generation
NeurIPS 2025Poster
202412 篇
3
Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective
ICLR 2024Poster
4
Sparse Backpropagation for MoE Training
ICLR 2024Rejected
4
Deep Reinforcement Learning from Weak Hierarchical Preference Feedback
ICLR 2024Rejected
3
HART: Efficient Adaptation via Regularized Autoregressive Parameter Generation
ICLR 2024Rejected
4
Make Your LLM Fully Utilize the Context
NeurIPS 2024Poster
3
LoftQ: LoRA-Fine-Tuning-aware Quantization for Large Language Models
ICLR 2024Oral
4
Exploring the Mystery of Influential Data for Mathematical Reasoning
COLM 2024Poster
4
CRITIC: Large Language Models Can Self-Correct with Tool-Interactive Critiquing
ICLR 2024Poster
4
ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem Solving
ICLR 2024Poster
4
WizardArena: Post-training Large Language Models via Simulated Offline Chatbot Arena
NeurIPS 2024Poster
3
Supervised Knowledge Makes Large Language Models Better In-context Learners
ICLR 2024Poster
3
Not All Tokens Are What You Need for Pretraining
NeurIPS 2024Oral