Jiantao Jiao
~Jiantao_Jiao1
19
论文总数
9.5
年均投稿
平均评分
接收情况12/19
会议分布
ICLR
9
NeurIPS
6
ICML
2
COLM
2
发表论文 (19 篇)
202510 篇
4
Thinking LLMs: General Instruction Following with Thought Generation
ICML 2025Poster
4
Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs
ICLR 2025Rejected
4
Reasoning by Superposition: A Theoretical Perspective on Chain of Continuous Thought
NeurIPS 2025Poster
3
Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning
ICML 2025Poster
4
Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers
NeurIPS 2025Poster
3
Information-Driven Design of Imaging Systems
NeurIPS 2025Poster
3
EmbedLLM: Learning Compact Representations of Large Language Models
ICLR 2025Spotlight
4
Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge
ICLR 2025Rejected
4
How to Evaluate Reward Models for RLHF
ICLR 2025Poster
4
Watermarking using Semantic-aware Speculative Sampling: from Theory to Practice
ICLR 2025Rejected
20249 篇
4
An Analysis of Tokenization: Transformers under Markov Data
NeurIPS 2024Spotlight
-
Data Refinement: Mitigating Reward Over-Optimization in Reinforcement Learning with Human Feedback
ICLR 2024withdrawn
5
Toxicity Detection for Free
NeurIPS 2024Spotlight
4
Towards a Theoretical Understanding of the 'Reversal Curse' via Training Dynamics
NeurIPS 2024Poster
4
Pairwise Proximal Policy Optimization: Language Model Alignment with Comparative RL
COLM 2024Poster
4
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for LLM Alignment
ICLR 2024Rejected
3
End-to-end Story Plot Generator
ICLR 2024Rejected
4
Fine-Tuning Language Models with Advantage-Induced Policy Alignment
ICLR 2024Rejected
4
Starling-7B: Improving Helpfulness and Harmlessness with RLAIF
COLM 2024Poster