paper-review

an archive of posts with this tag

Aug 19, 2025 Spurious Rewards: Rethinking Training Signals in RLVR
Aug 19, 2025 ON THE GENERALIZATION OF SFT: A REINFORCEMENT LEARNING PERSPECTIVE WITH REWARD RECTIFICATION
Aug 12, 2025 What Makes a Reward Model a Good Teacher? An Optimization Perspective / The Accuracy Paradox in RLHF: When Better Reward Models Don’t Yield Better Language Models
Aug 12, 2025 The Accuracy Paradox in RLHF: When Better Reward Models Don’t Yield Better Language Models / What Makes a Reward Model a Good Teacher? An Optimization Perspective
Aug 12, 2025 ON THE EXPRESSIVENESS OF SOFTMAX ATTENTION: A RECURRENT NEURAL NETWORK PERSPECTIVE
Aug 05, 2025 Impact of Fine-Tuning Methods on Memorization in Large Language Models
Aug 05, 2025 BLOCK DIFFUSION: INTERPOLATING BETWEEN AUTOREGRESSIVE AND DIFFUSION LANGUAGE MODELS
Jul 15, 2025 Search-R1: Training LLMs to Reason and Leverage Search Engines with Reinforcement Learning
Jul 15, 2025 Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models
Jul 15, 2025 Reasoning Model is Stubborn: Diagnosing Instruction Overriding in Reasoning Models
Jul 01, 2025 Reasoning Models Can Be Effective Without Thinking
Jul 01, 2025 Between Underthinking and Overthinking: An Empirical Study of Reasoning Length and correctness in LLMs
Jun 24, 2025 See What You Are Told: Visual Attention Sink in Large Multimodal Models
Jun 17, 2025 Diffusion of Thought: Chain-of-Thought Reasoning in Diffusion Language Models
Jun 10, 2025 Towards a Generalizable Bimanual Foundation Policy via Flow-based Video Prediction
Jun 10, 2025 DRA-GRPO: Exploring Diversity-Aware Reward Adjustment for R1-Zero-Like Training of Large Language Models
Jun 03, 2025 Textgrad: Automatic “Differentiation” via Text
Jun 03, 2025 Reinforcement Learning Finetunes Small Subnetworks in Large Language Models
Apr 22, 2025 Fine-tuning Vision-Language-Action Models: Optimizing Speed and Success
Apr 15, 2025 Universal and Transferable Adversarial Attacks on Aligned Language Models
Apr 15, 2025 Model Context Protocol (MCP) - provided by Antrophic
Apr 08, 2025 Reasoning Models Don’t Always Say What They Think
Apr 08, 2025 On the Biology of a Large Language Model
Mar 25, 2025 ReFT: Reasoning with Reinforced Fine-Tuning
Mar 11, 2025 WHEN IS TASK VECTOR Provably EFFECTIVE FOR MODEL EDITING? A GENERALIZATION ANALYSIS OF NONLINEAR TRANSFORMERS
Mar 11, 2025 Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs
Mar 04, 2025 SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution
Mar 04, 2025 Logic-RL: Unleashing LLM Reasoning with Rule-Based Reinforcement Learning
Mar 04, 2025 Contextual Document Embeddings
Feb 18, 2025 DeepSeek v3
Feb 04, 2025 Titans: Learning to Memorize at Test Time
Feb 04, 2025 SSM → HIPPO → LSSL → S4 → Mamba → Mamba2
Jan 21, 2025 Agent Laboratory: Using LLM Agents as Research Assistants
Jan 14, 2025 OpenVLA: An Open-Source Vision-Language-Action Model
Jan 02, 2025 TraceVLA: Visual Trace Prompting Enhances Spatial-Temporal Awareness for Generalist Robotic Policies
Jan 02, 2025 Inferring from Logits: Exploring Best Practices for Decoding-Free Generative Candidate Selection
Jan 02, 2025 Diffusion Language Model-Mathematical foundations & inference optimization
Jan 02, 2025 DeepSeek R1
Jan 02, 2025 d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning
Oct 17, 2024 Rule Based Rewards for Language Model Safety
Oct 17, 2024 KNOWLEDGE ENTROPY DECAY DURING LANGUAGE MODEL PRETRAINING HINDERS NEW KNOWLEDGE ACQUISITION
Oct 10, 2024 FAITHEVAL: CAN YOUR LANGUAGE MODEL STAY FAITHFUL TO CONTEXT, EVEN IF “THE MOON IS MADE OF MARSHMALLOWS”
Oct 03, 2024 QCRD: Quality-guided Contrastive Rationale Distillation for Large Lanauge Models
Sep 23, 2024 Training Language Models to Self-Correct via Reinforcement Learning
Sep 23, 2024 SUPER: Evaluating Agents on Setting Up and Executing Tasks from Research Repositories
Sep 09, 2024 Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling
Sep 09, 2024 Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models
Sep 02, 2024 Many-shot jailbreaking
Sep 02, 2024 LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders
Aug 20, 2024 Knowledge-Augmented Reasoning distillation for Small Language Models in Knowledge-Intensive Tasks (KARD)
Aug 13, 2024 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process
Aug 13, 2024 Knowledge conflict survey
Jul 30, 2024 In-Context Retrieval-Augmented Language Models
Jul 23, 2024 Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Jul 23, 2024 Step-DPO : Step-wise preference optimization for long-chain reasoning of LLMs
Jul 23, 2024 Pyspark - How to preprocess Large Scale Data with Python
Jul 22, 2024 LLAVA - Visual Instruction Tuning
Jul 02, 2024 RL-JACK: Reinforcement Learning-powered Black-box Jailbreaking Attack against LLMs
Jul 02, 2024 Llama3 Tokenizer
Jun 11, 2024 Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet
Jun 11, 2024 Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Jun 11, 2024 Contextual Position Encoding: Learning to Count What’s Important
Jun 04, 2024 Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
May 28, 2024 SimPO: Simple Preference Optimization with a Reference-Free Reward
May 27, 2024 Understanding the performance gap between online and offline alignment algorithms
May 21, 2024 LLAMA PRO: Progressive LLaMA with Block Expansion
May 07, 2024 How to Train LLM? - From Data Parallel To Fully Sharded Data Parallel
May 07, 2024 How to Inference Big LLM? - Using Accelerate Library
Apr 30, 2024 Training diffusion modelse with reinforcement learning
Apr 30, 2024 Many-Shot In-Context Learning
Apr 23, 2024 ORPO: Monolithic Preference Optimization without Reference Model
Apr 23, 2024 Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?
Apr 16, 2024 Understanding Emergent Abilities of Language Models from the Loss Perspective
Apr 13, 2024 Scaling Laws for Data Filtering— Data Curation cannot be Compute Agnostic
Apr 02, 2024 Preference-free Alignment Learning with Regularized Relevance Reward
Mar 26, 2024 Search-in-the-Chain: Interactively Enhancing Large Language Models with Search for Knowledge-intensive Tasks
Mar 19, 2024 Unveiling the Generalization Power of Fine-Tuned Large Language Models
Mar 12, 2024 A Simple and Effective Pruning Approach for Large Language Models
Mar 11, 2024 BitNet: Scaling 1-bit Transformers for Large Language Models
Mar 05, 2024 Beyond Memorization: Violating Privacy Via Inferencing With LLMs
Feb 27, 2024 SELF-RAG: LEARNING TO RETRIEVE, GENERATE, AND CRITIQUE THROUGH SELF-REFLECTION
Feb 20, 2024 WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia
Feb 20, 2024 KNOWLEDGE CARD: FILLING LLMS’ KNOWLEDGE GAPS WITH PLUG-IN SPECIALIZED LANGUAGE MODELS
Feb 13, 2024 LLM AUGMENTED LLMS: EXPANDING CAPABILITIES THROUGH COMPOSITION
Feb 13, 2024 CAN SENSITIVE INFORMATION BE DELETED FROM LLMS? OBJECTIVES FOR DEFENDING AGAINST EXTRACTION ATTACKS
Feb 06, 2024 Self-Rewarding Language Models
Jan 30, 2024 Lion: Adversarial Distillation of Proprietary Large Language Models
Jan 23, 2024 OVERTHINKING THE TRUTH: UNDERSTANDING HOW LANGUAGE MODELS PROCESS FALSE DEMONSTRATIONS
Jan 23, 2024 IN-CONTEXT PRETRAINING: LANGUAGE MODELING BEYOND DOCUMENT BOUNDARIES
Jan 16, 2024 Mistral 7B & Mixtral (Mixtral of Experts)
Jan 16, 2024 BENCHMARKING COGNITIVE BIASES IN LARGE LANGUAGE MODELS AS EVALUATORS
Jan 09, 2024 Making Large Language Models A Better Foundation For Dense Retrieval
Jan 03, 2024 vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention
Jan 02, 2024 DETECTING PRETRAINING DATA FROM LARGE LANGUAGE MODELS
Dec 26, 2023 Are Emergent Abilities of Large Language Models a Mirage?
Dec 19, 2023 Learning to Tokenize for Generative Retrieval
Dec 19, 2023 Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
Dec 12, 2023 Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
Oct 31, 2023 In-Context Learning Learns Label Relationships but Is Not Conventional Learning
Oct 31, 2023 EFFICIENT STREAMING LANGUAGE MODELS WITH ATTENTION SINKS
Oct 31, 2023 A Survey on Large Language Model based Autonomous Agents
Oct 17, 2023 Resolving Interference When Merging Models
Oct 10, 2023 LongLoRA: Efficient Fine-Tuning of Long-Context Large Language Models
Oct 03, 2023 DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
Sep 19, 2023 The CRINGE Loss: Learning what language not to model
Sep 19, 2023 LARGE LANGUAGE MODELS AS OPTIMIZERS
Sep 12, 2023 SILO LANGUAGE MODELS: ISOLATING LEGAL RISK IN A NONPARAMETRIC DATASTORE
Sep 12, 2023 A Systematic Study of Knowledge Distillation for Natural Language Generation with Pseudo-Target Training
Aug 29, 2023 Code Llama: Open Foundation Models for Code
Jun 29, 2023 QLoRA: Eficient Finetuning of Quantized LLMs
Jun 22, 2023 The False Promise of Imitating Proprietary LLMs
Jun 15, 2023 Do Prompt-Based Models Really Understand the Meaning of Their Prompts?
May 25, 2023 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
May 11, 2023 Measuring Association Between Labels and Free-Text Rationales
Apr 27, 2023 Automatic chain of thought prompting in large language models
Apr 20, 2023 FALSESUM : Generating Document-level NLI Examples for Recognizing Factual Inconsistency in Summarization
Apr 13, 2023 P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
Apr 13, 2023 AdapterDrop: On the Efficiency of Adapters in Transformers
Mar 30, 2023 GPT Understands, Too
Mar 16, 2023 Calibrating Factual Knowledge in Pretrained Language Models
Feb 09, 2023 AdapterHub: A Framework for Adapting Transformers, Parameter-Efficient Transfer Learning for NLP
Feb 02, 2023 Measuring and Improving Semantic Diversity of Dialogue Generation
Jan 26, 2023 Task-aware Retrieval with Instructions
Jan 19, 2023 KALA: Knowledge-Augmented Language Model Adaptation
Jan 12, 2023 A Survey for In-context Learning