LLM agents PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models Paper • 2402.01118 • Published Feb 2, 2024 • 32
PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models Paper • 2402.01118 • Published Feb 2, 2024 • 32
Prompting In-Context Principle Learning from Mistakes Paper • 2402.05403 • Published Feb 8, 2024 • 17 Chain-of-Thought Reasoning Without Prompting Paper • 2402.10200 • Published Feb 15, 2024 • 109
LLM training A Tale of Tails: Model Collapse as a Change of Scaling Laws Paper • 2402.07043 • Published Feb 10, 2024 • 15 How to Train Data-Efficient LLMs Paper • 2402.09668 • Published Feb 15, 2024 • 42
A Tale of Tails: Model Collapse as a Change of Scaling Laws Paper • 2402.07043 • Published Feb 10, 2024 • 15
LLM pruning Shortened LLaMA: A Simple Depth Pruning for Large Language Models Paper • 2402.02834 • Published Feb 5, 2024 • 17
Shortened LLaMA: A Simple Depth Pruning for Large Language Models Paper • 2402.02834 • Published Feb 5, 2024 • 17
Math LLM InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning Paper • 2402.06332 • Published Feb 9, 2024 • 19 AutoMathText: Autonomous Data Selection with Language Models for Mathematical Texts Paper • 2402.07625 • Published Feb 12, 2024 • 16 Reformatted Alignment Paper • 2402.12219 • Published Feb 19, 2024 • 17
InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning Paper • 2402.06332 • Published Feb 9, 2024 • 19
AutoMathText: Autonomous Data Selection with Language Models for Mathematical Texts Paper • 2402.07625 • Published Feb 12, 2024 • 16
Long llm Transformers Can Achieve Length Generalization But Not Robustly Paper • 2402.09371 • Published Feb 14, 2024 • 14 In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs Miss Paper • 2402.10790 • Published Feb 16, 2024 • 42
Transformers Can Achieve Length Generalization But Not Robustly Paper • 2402.09371 • Published Feb 14, 2024 • 14
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs Miss Paper • 2402.10790 • Published Feb 16, 2024 • 42
LLM agents PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models Paper • 2402.01118 • Published Feb 2, 2024 • 32
PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models Paper • 2402.01118 • Published Feb 2, 2024 • 32
LLM pruning Shortened LLaMA: A Simple Depth Pruning for Large Language Models Paper • 2402.02834 • Published Feb 5, 2024 • 17
Shortened LLaMA: A Simple Depth Pruning for Large Language Models Paper • 2402.02834 • Published Feb 5, 2024 • 17
Prompting In-Context Principle Learning from Mistakes Paper • 2402.05403 • Published Feb 8, 2024 • 17 Chain-of-Thought Reasoning Without Prompting Paper • 2402.10200 • Published Feb 15, 2024 • 109
Math LLM InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning Paper • 2402.06332 • Published Feb 9, 2024 • 19 AutoMathText: Autonomous Data Selection with Language Models for Mathematical Texts Paper • 2402.07625 • Published Feb 12, 2024 • 16 Reformatted Alignment Paper • 2402.12219 • Published Feb 19, 2024 • 17
InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning Paper • 2402.06332 • Published Feb 9, 2024 • 19
AutoMathText: Autonomous Data Selection with Language Models for Mathematical Texts Paper • 2402.07625 • Published Feb 12, 2024 • 16
LLM training A Tale of Tails: Model Collapse as a Change of Scaling Laws Paper • 2402.07043 • Published Feb 10, 2024 • 15 How to Train Data-Efficient LLMs Paper • 2402.09668 • Published Feb 15, 2024 • 42
A Tale of Tails: Model Collapse as a Change of Scaling Laws Paper • 2402.07043 • Published Feb 10, 2024 • 15
Long llm Transformers Can Achieve Length Generalization But Not Robustly Paper • 2402.09371 • Published Feb 14, 2024 • 14 In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs Miss Paper • 2402.10790 • Published Feb 16, 2024 • 42
Transformers Can Achieve Length Generalization But Not Robustly Paper • 2402.09371 • Published Feb 14, 2024 • 14
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs Miss Paper • 2402.10790 • Published Feb 16, 2024 • 42