-
Moral Foundations of Large Language Models
Paper • 2310.15337 • Published • 1 -
Specific versus General Principles for Constitutional AI
Paper • 2310.13798 • Published • 3 -
Contrastive Prefence Learning: Learning from Human Feedback without RL
Paper • 2310.13639 • Published • 25 -
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51
Collections
Discover the best community collections!
Collections including paper arxiv:2401.01335
-
InGram: Inductive Knowledge Graph Embedding via Relation Graphs
Paper • 2305.19987 • Published • 2 -
Curating Grounded Synthetic Data with Global Perspectives for Equitable A
Paper • 2406.10258 • Published • 1 -
Peregrine: A Pattern-Aware Graph Mining System
Paper • 2004.02369 • Published • 1 -
OFFER: A Motif Dimensional Framework for Network Representation Learning
Paper • 2008.12010 • Published • 1
-
Unleashing the Power of Pre-trained Language Models for Offline Reinforcement Learning
Paper • 2310.20587 • Published • 18 -
SELF: Language-Driven Self-Evolution for Large Language Model
Paper • 2310.00533 • Published • 2 -
QLoRA: Efficient Finetuning of Quantized LLMs
Paper • 2305.14314 • Published • 54 -
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 44
-
A Critical Evaluation of AI Feedback for Aligning Large Language Models
Paper • 2402.12366 • Published • 3 -
Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation
Paper • 2401.08417 • Published • 37 -
Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks
Paper • 2404.14723 • Published • 10 -
Self-Play Preference Optimization for Language Model Alignment
Paper • 2405.00675 • Published • 28
-
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Paper • 2406.08464 • Published • 70 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 102 -
argilla/magpie-ultra-v1.0
Viewer • Updated • 3.22M • 1.33k • 47 -
simplescaling/s1K-1.1
Viewer • Updated • 1k • 8.06k • 126
-
Textbooks Are All You Need
Paper • 2306.11644 • Published • 145 -
Textbooks Are All You Need II: phi-1.5 technical report
Paper • 2309.05463 • Published • 87 -
TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
Paper • 2305.07759 • Published • 36 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 102
-
Chain-of-Thought Reasoning Without Prompting
Paper • 2402.10200 • Published • 110 -
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 43 -
BitDelta: Your Fine-Tune May Only Be Worth One Bit
Paper • 2402.10193 • Published • 23 -
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts
Paper • 2402.09727 • Published • 39
-
Moral Foundations of Large Language Models
Paper • 2310.15337 • Published • 1 -
Specific versus General Principles for Constitutional AI
Paper • 2310.13798 • Published • 3 -
Contrastive Prefence Learning: Learning from Human Feedback without RL
Paper • 2310.13639 • Published • 25 -
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51
-
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Paper • 2406.08464 • Published • 70 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 102 -
argilla/magpie-ultra-v1.0
Viewer • Updated • 3.22M • 1.33k • 47 -
simplescaling/s1K-1.1
Viewer • Updated • 1k • 8.06k • 126
-
InGram: Inductive Knowledge Graph Embedding via Relation Graphs
Paper • 2305.19987 • Published • 2 -
Curating Grounded Synthetic Data with Global Perspectives for Equitable A
Paper • 2406.10258 • Published • 1 -
Peregrine: A Pattern-Aware Graph Mining System
Paper • 2004.02369 • Published • 1 -
OFFER: A Motif Dimensional Framework for Network Representation Learning
Paper • 2008.12010 • Published • 1
-
Textbooks Are All You Need
Paper • 2306.11644 • Published • 145 -
Textbooks Are All You Need II: phi-1.5 technical report
Paper • 2309.05463 • Published • 87 -
TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
Paper • 2305.07759 • Published • 36 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 102
-
Unleashing the Power of Pre-trained Language Models for Offline Reinforcement Learning
Paper • 2310.20587 • Published • 18 -
SELF: Language-Driven Self-Evolution for Large Language Model
Paper • 2310.00533 • Published • 2 -
QLoRA: Efficient Finetuning of Quantized LLMs
Paper • 2305.14314 • Published • 54 -
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 44
-
A Critical Evaluation of AI Feedback for Aligning Large Language Models
Paper • 2402.12366 • Published • 3 -
Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation
Paper • 2401.08417 • Published • 37 -
Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks
Paper • 2404.14723 • Published • 10 -
Self-Play Preference Optimization for Language Model Alignment
Paper • 2405.00675 • Published • 28
-
Chain-of-Thought Reasoning Without Prompting
Paper • 2402.10200 • Published • 110 -
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 43 -
BitDelta: Your Fine-Tune May Only Be Worth One Bit
Paper • 2402.10193 • Published • 23 -
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts
Paper • 2402.09727 • Published • 39