Master-Zangetsu 's Collections ToRead
updated
Rethinking Mixture-of-Agents: Is Mixing Different Large Language Models
Beneficial?
Paper
• 2502.00674
• Published
• 13
Demystifying Long Chain-of-Thought Reasoning in LLMs
Paper
• 2502.03373
• Published
• 58
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model
Paper
• 2502.02737
• Published
• 256
DeepRAG: Thinking to Retrieval Step by Step for Large Language Models
Paper
• 2502.01142
• Published
• 25
Scaling Embedding Layers in Language Models
Paper
• 2502.01637
• Published
• 24
ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning
Paper
• 2502.01100
• Published
• 21
ScoreFlow: Mastering LLM Agent Workflows via Score-based Preference
Optimization
Paper
• 2502.04306
• Published
• 20
Analyze Feature Flow to Enhance Interpretation and Steering in Language
Models
Paper
• 2502.03032
• Published
• 60
QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Paper
• 2502.05003
• Published
• 44
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM
Guardrails
Paper
• 2502.05163
• Published
• 22
CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference
Paper
• 2502.04416
• Published
• 12
ARR: Question Answering with Large Language Models via Analyzing,
Retrieving, and Reasoning
Paper
• 2502.04689
• Published
• 8