Aug 19, 2025 | Spurious Rewards: Rethinking Training Signals in RLVR |
Aug 19, 2025 | ON THE GENERALIZATION OF SFT: A REINFORCEMENT LEARNING PERSPECTIVE WITH REWARD RECTIFICATION |
Aug 05, 2025 | Impact of Fine-Tuning Methods on Memorization in Large Language Models |
Aug 05, 2025 | BLOCK DIFFUSION: INTERPOLATING BETWEEN AUTOREGRESSIVE AND DIFFUSION LANGUAGE MODELS |
Jul 15, 2025 | Search-R1: Training LLMs to Reason and Leverage Search Engines with Reinforcement Learning |
Jul 15, 2025 | Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models |
Jul 15, 2025 | Reasoning Model is Stubborn: Diagnosing Instruction Overriding in Reasoning Models |
Jul 01, 2025 | Reasoning Models Can Be Effective Without Thinking |
Jul 01, 2025 | Between Underthinking and Overthinking: An Empirical Study of Reasoning Length and correctness in LLMs |
Jun 24, 2025 | See What You Are Told: Visual Attention Sink in Large Multimodal Models |
Jun 17, 2025 | Diffusion of Thought: Chain-of-Thought Reasoning in Diffusion Language Models |
Jun 10, 2025 | DRA-GRPO: Exploring Diversity-Aware Reward Adjustment for R1-Zero-Like Training of Large Language Models |
Jun 03, 2025 | Textgrad: Automatic “Differentiation” via Text |
Jun 03, 2025 | Reinforcement Learning Finetunes Small Subnetworks in Large Language Models |
Apr 22, 2025 | Fine-tuning Vision-Language-Action Models: Optimizing Speed and Success |
Apr 15, 2025 | Universal and Transferable Adversarial Attacks on Aligned Language Models |
Apr 15, 2025 | Model Context Protocol (MCP) - provided by Antrophic |
Apr 08, 2025 | On the Biology of a Large Language Model |
Mar 25, 2025 | ReFT: Reasoning with Reinforced Fine-Tuning |
Mar 11, 2025 | WHEN IS TASK VECTOR Provably EFFECTIVE FOR MODEL EDITING? A GENERALIZATION ANALYSIS OF NONLINEAR TRANSFORMERS |
Mar 11, 2025 | Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs |
Mar 04, 2025 | SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution |
Mar 04, 2025 | Logic-RL: Unleashing LLM Reasoning with Rule-Based Reinforcement Learning |
Feb 18, 2025 | DeepSeek v3 |
Feb 04, 2025 | SSM → HIPPO → LSSL → S4 → Mamba → Mamba2 |
Jan 21, 2025 | Agent Laboratory: Using LLM Agents as Research Assistants |
Jan 14, 2025 | OpenVLA: An Open-Source Vision-Language-Action Model |
Jan 02, 2025 | TraceVLA: Visual Trace Prompting Enhances Spatial-Temporal Awareness for Generalist Robotic Policies |
Jan 02, 2025 | Inferring from Logits: Exploring Best Practices for Decoding-Free Generative Candidate Selection |
Jan 02, 2025 | DeepSeek R1 |
Jan 02, 2025 | d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning |
Oct 17, 2024 | Rule Based Rewards for Language Model Safety |
Oct 10, 2024 | FAITHEVAL: CAN YOUR LANGUAGE MODEL STAY FAITHFUL TO CONTEXT, EVEN IF “THE MOON IS MADE OF MARSHMALLOWS” |
Oct 03, 2024 | QCRD: Quality-guided Contrastive Rationale Distillation for Large Lanauge Models |
Sep 23, 2024 | Training Language Models to Self-Correct via Reinforcement Learning |
Sep 23, 2024 | SUPER: Evaluating Agents on Setting Up and Executing Tasks from Research Repositories |
Sep 09, 2024 | Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling |
Sep 09, 2024 | Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models |
Sep 02, 2024 | Many-shot jailbreaking |
Sep 02, 2024 | LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders |
Aug 20, 2024 | Knowledge-Augmented Reasoning distillation for Small Language Models in Knowledge-Intensive Tasks (KARD) |
Aug 13, 2024 | Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process |
Aug 13, 2024 | Knowledge conflict survey |
Jul 30, 2024 | In-Context Retrieval-Augmented Language Models |
Jul 23, 2024 | Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning |
Jul 23, 2024 | Step-DPO : Step-wise preference optimization for long-chain reasoning of LLMs |
Jul 23, 2024 | Pyspark - How to preprocess Large Scale Data with Python |
Jul 02, 2024 | RL-JACK: Reinforcement Learning-powered Black-box Jailbreaking Attack against LLMs |
Jun 11, 2024 | Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet |
Jun 11, 2024 | Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? |
Jun 11, 2024 | Contextual Position Encoding: Learning to Count What’s Important |
Jun 04, 2024 | Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training |
May 27, 2024 | Understanding the performance gap between online and offline alignment algorithms |
May 21, 2024 | LLAMA PRO: Progressive LLaMA with Block Expansion |
May 07, 2024 | How to Train LLM? - From Data Parallel To Fully Sharded Data Parallel |
May 07, 2024 | How to Inference Big LLM? - Using Accelerate Library |
Apr 30, 2024 | Many-Shot In-Context Learning |
Apr 23, 2024 | ORPO: Monolithic Preference Optimization without Reference Model |
Apr 23, 2024 | Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers? |
Apr 16, 2024 | Understanding Emergent Abilities of Language Models from the Loss Perspective |
Apr 13, 2024 | Scaling Laws for Data Filtering— Data Curation cannot be Compute Agnostic |
Apr 02, 2024 | Preference-free Alignment Learning with Regularized Relevance Reward |
Mar 26, 2024 | Search-in-the-Chain: Interactively Enhancing Large Language Models with Search for Knowledge-intensive Tasks |
Mar 19, 2024 | Unveiling the Generalization Power of Fine-Tuned Large Language Models |
Mar 12, 2024 | A Simple and Effective Pruning Approach for Large Language Models |
Mar 11, 2024 | BitNet: Scaling 1-bit Transformers for Large Language Models |
Mar 05, 2024 | Beyond Memorization: Violating Privacy Via Inferencing With LLMs |
Feb 27, 2024 | SELF-RAG: LEARNING TO RETRIEVE, GENERATE, AND CRITIQUE THROUGH SELF-REFLECTION |
Feb 20, 2024 | WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia |
Feb 20, 2024 | KNOWLEDGE CARD: FILLING LLMS’ KNOWLEDGE GAPS WITH PLUG-IN SPECIALIZED LANGUAGE MODELS |
Feb 13, 2024 | LLM AUGMENTED LLMS: EXPANDING CAPABILITIES THROUGH COMPOSITION |
Feb 13, 2024 | CAN SENSITIVE INFORMATION BE DELETED FROM LLMS? OBJECTIVES FOR DEFENDING AGAINST EXTRACTION ATTACKS |
Feb 06, 2024 | Self-Rewarding Language Models |
Jan 30, 2024 | Lion: Adversarial Distillation of Proprietary Large Language Models |
Jan 23, 2024 | IN-CONTEXT PRETRAINING: LANGUAGE MODELING BEYOND DOCUMENT BOUNDARIES |
Jan 16, 2024 | Mistral 7B & Mixtral (Mixtral of Experts) |
Jan 16, 2024 | BENCHMARKING COGNITIVE BIASES IN LARGE LANGUAGE MODELS AS EVALUATORS |
Jan 09, 2024 | Making Large Language Models A Better Foundation For Dense Retrieval |
Jan 03, 2024 | vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention |
Jan 02, 2024 | DETECTING PRETRAINING DATA FROM LARGE LANGUAGE MODELS |
Dec 26, 2023 | Are Emergent Abilities of Large Language Models a Mirage? |
Dec 19, 2023 | Break the Sequential Dependency of LLM Inference Using Lookahead Decoding |
Dec 12, 2023 | Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning |
Oct 31, 2023 | In-Context Learning Learns Label Relationships but Is Not Conventional Learning |
Oct 31, 2023 | EFFICIENT STREAMING LANGUAGE MODELS WITH ATTENTION SINKS |
Oct 31, 2023 | A Survey on Large Language Model based Autonomous Agents |
Oct 10, 2023 | LongLoRA: Efficient Fine-Tuning of Long-Context Large Language Models |
Oct 03, 2023 | DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models |
Sep 19, 2023 | LARGE LANGUAGE MODELS AS OPTIMIZERS |
Sep 12, 2023 | SILO LANGUAGE MODELS: ISOLATING LEGAL RISK IN A NONPARAMETRIC DATASTORE |
Sep 12, 2023 | A Systematic Study of Knowledge Distillation for Natural Language Generation with Pseudo-Target Training |
Aug 29, 2023 | Code Llama: Open Foundation Models for Code |
Jun 29, 2023 | QLoRA: Eficient Finetuning of Quantized LLMs |
Jun 22, 2023 | The False Promise of Imitating Proprietary LLMs |
May 25, 2023 | Rethinking the Role of Demonstrations: What Makes In-Context Learning Work? |
Jan 26, 2023 | Task-aware Retrieval with Instructions |
Jan 19, 2023 | KALA: Knowledge-Augmented Language Model Adaptation |
Jan 12, 2023 | A Survey for In-context Learning |