-
Attention Is All You Need
Paper • 1706.03762 • Published • 61 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 39 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 58 -
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 40
Collections
Discover the best community collections!
Collections including paper arxiv:2106.09685
-
Agents: An Open-source Framework for Autonomous Language Agents
Paper • 2309.07870 • Published • 42 -
Language Agents with Reinforcement Learning for Strategic Play in the Werewolf Game
Paper • 2310.18940 • Published -
Advances and Challenges in Foundation Agents: From Brain-Inspired Intelligence to Evolutionary, Collaborative, and Safe Systems
Paper • 2504.01990 • Published • 275 -
AutoWebGLM: Bootstrap And Reinforce A Large Language Model-based Web Navigating Agent
Paper • 2404.03648 • Published • 29
-
Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning
Paper • 2211.04325 • Published -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 18 -
On the Opportunities and Risks of Foundation Models
Paper • 2108.07258 • Published -
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Paper • 2204.07705 • Published • 1
-
Adapters: A Unified Library for Parameter-Efficient and Modular Transfer Learning
Paper • 2311.11077 • Published • 28 -
Tensor Product Attention Is All You Need
Paper • 2501.06425 • Published • 88 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 39 -
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
Paper • 2403.03853 • Published • 65
-
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Paper • 1910.10683 • Published • 13 -
AutoTrain: No-code training for state-of-the-art models
Paper • 2410.15735 • Published • 60 -
LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Paper • 2405.00732 • Published • 122 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 39