-
Memory Augmented Language Models through Mixture of Word Experts
Paper • 2311.10768 • Published • 19 -
System 2 Attention (is something you might need too)
Paper • 2311.11829 • Published • 43 -
Fine-tuning Language Models for Factuality
Paper • 2311.08401 • Published • 30 -
Orca 2: Teaching Small Language Models How to Reason
Paper • 2311.11045 • Published • 77
Collections
Discover the best community collections!
Collections including paper arxiv:2312.01552
-
#InsTag: Instruction Tagging for Analyzing Supervised Fine-tuning of Large Language Models
Paper • 2308.07074 • Published -
Evoke: Evoking Critical Thinking Abilities in LLMs via Reviewer-Author Prompt Editing
Paper • 2310.13855 • Published • 1 -
LIMIT: Less Is More for Instruction Tuning Across Evaluation Paradigms
Paper • 2311.13133 • Published -
Group Preference Optimization: Few-Shot Alignment of Large Language Models
Paper • 2310.11523 • Published
-
Eureka: Human-Level Reward Design via Coding Large Language Models
Paper • 2310.12931 • Published • 26 -
GENOME: GenerativE Neuro-symbOlic visual reasoning by growing and reusing ModulEs
Paper • 2311.04901 • Published • 9 -
Hiformer: Heterogeneous Feature Interactions Learning with Transformers for Recommender Systems
Paper • 2311.05884 • Published • 9 -
PolyMaX: General Dense Prediction with Mask Transformer
Paper • 2311.05770 • Published • 8
-
LongLoRA: Efficient Fine-tuning of Long-Context Large Language Models
Paper • 2309.12307 • Published • 89 -
NEFTune: Noisy Embeddings Improve Instruction Finetuning
Paper • 2310.05914 • Published • 14 -
SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling
Paper • 2312.15166 • Published • 60 -
Soaring from 4K to 400K: Extending LLM's Context with Activation Beacon
Paper • 2401.03462 • Published • 27
-
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator
Paper • 2312.04474 • Published • 33 -
Training Chain-of-Thought via Latent-Variable Inference
Paper • 2312.02179 • Published • 10 -
The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Paper • 2312.01552 • Published • 32 -
AppAgent: Multimodal Agents as Smartphone Users
Paper • 2312.13771 • Published • 54
-
The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMs
Paper • 2210.14986 • Published • 5 -
Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
Paper • 2311.10702 • Published • 19 -
Large Language Models as Optimizers
Paper • 2309.03409 • Published • 78 -
From Sparse to Dense: GPT-4 Summarization with Chain of Density Prompting
Paper • 2309.04269 • Published • 33
-
Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions
Paper • 2309.10150 • Published • 25 -
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper • 2310.10638 • Published • 30 -
Farzi Data: Autoregressive Data Distillation
Paper • 2310.09983 • Published • 10 -
LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Paper • 2311.05437 • Published • 51
-
Memory Augmented Language Models through Mixture of Word Experts
Paper • 2311.10768 • Published • 19 -
System 2 Attention (is something you might need too)
Paper • 2311.11829 • Published • 43 -
Fine-tuning Language Models for Factuality
Paper • 2311.08401 • Published • 30 -
Orca 2: Teaching Small Language Models How to Reason
Paper • 2311.11045 • Published • 77
-
LongLoRA: Efficient Fine-tuning of Long-Context Large Language Models
Paper • 2309.12307 • Published • 89 -
NEFTune: Noisy Embeddings Improve Instruction Finetuning
Paper • 2310.05914 • Published • 14 -
SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling
Paper • 2312.15166 • Published • 60 -
Soaring from 4K to 400K: Extending LLM's Context with Activation Beacon
Paper • 2401.03462 • Published • 27
-
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator
Paper • 2312.04474 • Published • 33 -
Training Chain-of-Thought via Latent-Variable Inference
Paper • 2312.02179 • Published • 10 -
The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Paper • 2312.01552 • Published • 32 -
AppAgent: Multimodal Agents as Smartphone Users
Paper • 2312.13771 • Published • 54
-
#InsTag: Instruction Tagging for Analyzing Supervised Fine-tuning of Large Language Models
Paper • 2308.07074 • Published -
Evoke: Evoking Critical Thinking Abilities in LLMs via Reviewer-Author Prompt Editing
Paper • 2310.13855 • Published • 1 -
LIMIT: Less Is More for Instruction Tuning Across Evaluation Paradigms
Paper • 2311.13133 • Published -
Group Preference Optimization: Few-Shot Alignment of Large Language Models
Paper • 2310.11523 • Published
-
The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMs
Paper • 2210.14986 • Published • 5 -
Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
Paper • 2311.10702 • Published • 19 -
Large Language Models as Optimizers
Paper • 2309.03409 • Published • 78 -
From Sparse to Dense: GPT-4 Summarization with Chain of Density Prompting
Paper • 2309.04269 • Published • 33
-
Eureka: Human-Level Reward Design via Coding Large Language Models
Paper • 2310.12931 • Published • 26 -
GENOME: GenerativE Neuro-symbOlic visual reasoning by growing and reusing ModulEs
Paper • 2311.04901 • Published • 9 -
Hiformer: Heterogeneous Feature Interactions Learning with Transformers for Recommender Systems
Paper • 2311.05884 • Published • 9 -
PolyMaX: General Dense Prediction with Mask Transformer
Paper • 2311.05770 • Published • 8
-
Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions
Paper • 2309.10150 • Published • 25 -
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper • 2310.10638 • Published • 30 -
Farzi Data: Autoregressive Data Distillation
Paper • 2310.09983 • Published • 10 -
LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Paper • 2311.05437 • Published • 51