-
Self-Reflection in LLM Agents: Effects on Problem-Solving Performance
Paper β’ 2405.06682 β’ Published β’ 3 -
Self-Refine: Iterative Refinement with Self-Feedback
Paper β’ 2303.17651 β’ Published β’ 2 -
Rethinking Chain-of-Thought from the Perspective of Self-Training
Paper β’ 2412.10827 β’ Published -
Reflexion: Language Agents with Verbal Reinforcement Learning
Paper β’ 2303.11366 β’ Published β’ 5
Melisa Russak
melisa
AI & ML interests
I love definitions
Recent Activity
updated
a model
6 days ago
Writer/colab
updated
a model
8 days ago
melisa/results_bert
published
a model
8 days ago
melisa/results_bert
Organizations
Self-improving LLMs
-
Self-Taught Self-Correction for Small Language Models
Paper β’ 2503.08681 β’ Published β’ 15 -
Self-Improving Robust Preference Optimization
Paper β’ 2406.01660 β’ Published β’ 20 -
LADDER: Self-Improving LLMs Through Recursive Problem Decomposition
Paper β’ 2503.00735 β’ Published β’ 22 -
Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge
Paper β’ 2407.19594 β’ Published β’ 21
Model Merging
Daily Papers 2025
Daily Papers
-
Simple linear attention language models balance the recall-throughput tradeoff
Paper β’ 2402.18668 β’ Published β’ 21 -
Linear Transformers with Learnable Kernel Functions are Better In-Context Models
Paper β’ 2402.10644 β’ Published β’ 82 -
Repeat After Me: Transformers are Better than State Space Models at Copying
Paper β’ 2402.01032 β’ Published β’ 25 -
Zoology: Measuring and Improving Recall in Efficient Language Models
Paper β’ 2312.04927 β’ Published β’ 2
lshort-transformers
Papers useful when writing the paper: "The Not So Short Transfromers"
-
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
Paper β’ 2403.03853 β’ Published β’ 66 -
SliceGPT: Compress Large Language Models by Deleting Rows and Columns
Paper β’ 2401.15024 β’ Published β’ 74 -
Your Transformer is Secretly Linear
Paper β’ 2405.12250 β’ Published β’ 159 -
Yi: Open Foundation Models by 01.AI
Paper β’ 2403.04652 β’ Published β’ 66
self-reflection papers
-
Self-Reflection in LLM Agents: Effects on Problem-Solving Performance
Paper β’ 2405.06682 β’ Published β’ 3 -
Self-Refine: Iterative Refinement with Self-Feedback
Paper β’ 2303.17651 β’ Published β’ 2 -
Rethinking Chain-of-Thought from the Perspective of Self-Training
Paper β’ 2412.10827 β’ Published -
Reflexion: Language Agents with Verbal Reinforcement Learning
Paper β’ 2303.11366 β’ Published β’ 5
Daily Papers 2025
Self-improving LLMs
-
Self-Taught Self-Correction for Small Language Models
Paper β’ 2503.08681 β’ Published β’ 15 -
Self-Improving Robust Preference Optimization
Paper β’ 2406.01660 β’ Published β’ 20 -
LADDER: Self-Improving LLMs Through Recursive Problem Decomposition
Paper β’ 2503.00735 β’ Published β’ 22 -
Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge
Paper β’ 2407.19594 β’ Published β’ 21
Daily Papers
-
Simple linear attention language models balance the recall-throughput tradeoff
Paper β’ 2402.18668 β’ Published β’ 21 -
Linear Transformers with Learnable Kernel Functions are Better In-Context Models
Paper β’ 2402.10644 β’ Published β’ 82 -
Repeat After Me: Transformers are Better than State Space Models at Copying
Paper β’ 2402.01032 β’ Published β’ 25 -
Zoology: Measuring and Improving Recall in Efficient Language Models
Paper β’ 2312.04927 β’ Published β’ 2
Model Merging
lshort-transformers
Papers useful when writing the paper: "The Not So Short Transfromers"
-
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
Paper β’ 2403.03853 β’ Published β’ 66 -
SliceGPT: Compress Large Language Models by Deleting Rows and Columns
Paper β’ 2401.15024 β’ Published β’ 74 -
Your Transformer is Secretly Linear
Paper β’ 2405.12250 β’ Published β’ 159 -
Yi: Open Foundation Models by 01.AI
Paper β’ 2403.04652 β’ Published β’ 66