llm

an archive of posts with this tag

Aug 19, 2025 Spurious Rewards: Rethinking Training Signals in RLVR
Aug 19, 2025 ON THE GENERALIZATION OF SFT: A REINFORCEMENT LEARNING PERSPECTIVE WITH REWARD RECTIFICATION
Aug 05, 2025 Impact of Fine-Tuning Methods on Memorization in Large Language Models
Aug 05, 2025 BLOCK DIFFUSION: INTERPOLATING BETWEEN AUTOREGRESSIVE AND DIFFUSION LANGUAGE MODELS
Jul 15, 2025 Search-R1: Training LLMs to Reason and Leverage Search Engines with Reinforcement Learning
Jul 15, 2025 Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models
Jul 15, 2025 Reasoning Model is Stubborn: Diagnosing Instruction Overriding in Reasoning Models
Jul 01, 2025 Reasoning Models Can Be Effective Without Thinking
Jul 01, 2025 Between Underthinking and Overthinking: An Empirical Study of Reasoning Length and correctness in LLMs
Jun 24, 2025 See What You Are Told: Visual Attention Sink in Large Multimodal Models
Jun 17, 2025 Diffusion of Thought: Chain-of-Thought Reasoning in Diffusion Language Models
Jun 10, 2025 DRA-GRPO: Exploring Diversity-Aware Reward Adjustment for R1-Zero-Like Training of Large Language Models
Jun 03, 2025 Textgrad: Automatic “Differentiation” via Text
Jun 03, 2025 Reinforcement Learning Finetunes Small Subnetworks in Large Language Models
Apr 22, 2025 Fine-tuning Vision-Language-Action Models: Optimizing Speed and Success
Apr 15, 2025 Universal and Transferable Adversarial Attacks on Aligned Language Models
Apr 15, 2025 Model Context Protocol (MCP) - provided by Antrophic
Apr 08, 2025 On the Biology of a Large Language Model
Mar 25, 2025 ReFT: Reasoning with Reinforced Fine-Tuning
Mar 11, 2025 WHEN IS TASK VECTOR Provably EFFECTIVE FOR MODEL EDITING? A GENERALIZATION ANALYSIS OF NONLINEAR TRANSFORMERS
Mar 11, 2025 Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs
Mar 04, 2025 SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution
Mar 04, 2025 Logic-RL: Unleashing LLM Reasoning with Rule-Based Reinforcement Learning
Feb 18, 2025 DeepSeek v3
Feb 04, 2025 SSM → HIPPO → LSSL → S4 → Mamba → Mamba2
Jan 21, 2025 Agent Laboratory: Using LLM Agents as Research Assistants
Jan 14, 2025 OpenVLA: An Open-Source Vision-Language-Action Model
Jan 02, 2025 TraceVLA: Visual Trace Prompting Enhances Spatial-Temporal Awareness for Generalist Robotic Policies
Jan 02, 2025 Inferring from Logits: Exploring Best Practices for Decoding-Free Generative Candidate Selection
Jan 02, 2025 DeepSeek R1
Jan 02, 2025 d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning
Oct 17, 2024 Rule Based Rewards for Language Model Safety
Oct 10, 2024 FAITHEVAL: CAN YOUR LANGUAGE MODEL STAY FAITHFUL TO CONTEXT, EVEN IF “THE MOON IS MADE OF MARSHMALLOWS”
Oct 03, 2024 QCRD: Quality-guided Contrastive Rationale Distillation for Large Lanauge Models
Sep 23, 2024 Training Language Models to Self-Correct via Reinforcement Learning
Sep 23, 2024 SUPER: Evaluating Agents on Setting Up and Executing Tasks from Research Repositories
Sep 09, 2024 Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling
Sep 09, 2024 Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models
Sep 02, 2024 Many-shot jailbreaking
Sep 02, 2024 LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders
Aug 20, 2024 Knowledge-Augmented Reasoning distillation for Small Language Models in Knowledge-Intensive Tasks (KARD)
Aug 13, 2024 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process
Aug 13, 2024 Knowledge conflict survey
Jul 30, 2024 In-Context Retrieval-Augmented Language Models
Jul 23, 2024 Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning
Jul 23, 2024 Step-DPO : Step-wise preference optimization for long-chain reasoning of LLMs
Jul 23, 2024 Pyspark - How to preprocess Large Scale Data with Python
Jul 02, 2024 RL-JACK: Reinforcement Learning-powered Black-box Jailbreaking Attack against LLMs
Jun 11, 2024 Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet
Jun 11, 2024 Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Jun 11, 2024 Contextual Position Encoding: Learning to Count What’s Important
Jun 04, 2024 Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
May 27, 2024 Understanding the performance gap between online and offline alignment algorithms
May 21, 2024 LLAMA PRO: Progressive LLaMA with Block Expansion
May 07, 2024 How to Train LLM? - From Data Parallel To Fully Sharded Data Parallel
May 07, 2024 How to Inference Big LLM? - Using Accelerate Library
Apr 30, 2024 Many-Shot In-Context Learning
Apr 23, 2024 ORPO: Monolithic Preference Optimization without Reference Model
Apr 23, 2024 Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?
Apr 16, 2024 Understanding Emergent Abilities of Language Models from the Loss Perspective
Apr 13, 2024 Scaling Laws for Data Filtering— Data Curation cannot be Compute Agnostic
Apr 02, 2024 Preference-free Alignment Learning with Regularized Relevance Reward
Mar 26, 2024 Search-in-the-Chain: Interactively Enhancing Large Language Models with Search for Knowledge-intensive Tasks
Mar 19, 2024 Unveiling the Generalization Power of Fine-Tuned Large Language Models
Mar 12, 2024 A Simple and Effective Pruning Approach for Large Language Models
Mar 11, 2024 BitNet: Scaling 1-bit Transformers for Large Language Models
Mar 05, 2024 Beyond Memorization: Violating Privacy Via Inferencing With LLMs
Feb 27, 2024 SELF-RAG: LEARNING TO RETRIEVE, GENERATE, AND CRITIQUE THROUGH SELF-REFLECTION
Feb 20, 2024 WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia
Feb 20, 2024 KNOWLEDGE CARD: FILLING LLMS’ KNOWLEDGE GAPS WITH PLUG-IN SPECIALIZED LANGUAGE MODELS
Feb 13, 2024 LLM AUGMENTED LLMS: EXPANDING CAPABILITIES THROUGH COMPOSITION
Feb 13, 2024 CAN SENSITIVE INFORMATION BE DELETED FROM LLMS? OBJECTIVES FOR DEFENDING AGAINST EXTRACTION ATTACKS
Feb 06, 2024 Self-Rewarding Language Models
Jan 30, 2024 Lion: Adversarial Distillation of Proprietary Large Language Models
Jan 23, 2024 IN-CONTEXT PRETRAINING: LANGUAGE MODELING BEYOND DOCUMENT BOUNDARIES
Jan 16, 2024 Mistral 7B & Mixtral (Mixtral of Experts)
Jan 16, 2024 BENCHMARKING COGNITIVE BIASES IN LARGE LANGUAGE MODELS AS EVALUATORS
Jan 09, 2024 Making Large Language Models A Better Foundation For Dense Retrieval
Jan 03, 2024 vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention
Jan 02, 2024 DETECTING PRETRAINING DATA FROM LARGE LANGUAGE MODELS
Dec 26, 2023 Are Emergent Abilities of Large Language Models a Mirage?
Dec 19, 2023 Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
Dec 12, 2023 Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
Oct 31, 2023 In-Context Learning Learns Label Relationships but Is Not Conventional Learning
Oct 31, 2023 EFFICIENT STREAMING LANGUAGE MODELS WITH ATTENTION SINKS
Oct 31, 2023 A Survey on Large Language Model based Autonomous Agents
Oct 10, 2023 LongLoRA: Efficient Fine-Tuning of Long-Context Large Language Models
Oct 03, 2023 DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
Sep 19, 2023 LARGE LANGUAGE MODELS AS OPTIMIZERS
Sep 12, 2023 SILO LANGUAGE MODELS: ISOLATING LEGAL RISK IN A NONPARAMETRIC DATASTORE
Sep 12, 2023 A Systematic Study of Knowledge Distillation for Natural Language Generation with Pseudo-Target Training
Aug 29, 2023 Code Llama: Open Foundation Models for Code
Jun 29, 2023 QLoRA: Eficient Finetuning of Quantized LLMs
Jun 22, 2023 The False Promise of Imitating Proprietary LLMs
May 25, 2023 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
Jan 26, 2023 Task-aware Retrieval with Instructions
Jan 19, 2023 KALA: Knowledge-Augmented Language Model Adaptation
Jan 12, 2023 A Survey for In-context Learning