-
Nemotron-4 15B Technical Report
Paper ⢠2402.16819 ⢠Published ⢠47 -
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper ⢠2402.19427 ⢠Published ⢠57 -
RWKV: Reinventing RNNs for the Transformer Era
Paper ⢠2305.13048 ⢠Published ⢠19 -
Reformer: The Efficient Transformer
Paper ⢠2001.04451 ⢠Published
Collections
Discover the best community collections!
Collections including paper arxiv:2311.16867
-
Attention Is All You Need
Paper ⢠1706.03762 ⢠Published ⢠61 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper ⢠1810.04805 ⢠Published ⢠18 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper ⢠1907.11692 ⢠Published ⢠7 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper ⢠1910.01108 ⢠Published ⢠14
-
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Paper ⢠2312.00752 ⢠Published ⢠143 -
Schrodinger Bridges Beat Diffusion Models on Text-to-Speech Synthesis
Paper ⢠2312.03491 ⢠Published ⢠35 -
Order Matters in the Presence of Dataset Imbalance for Multilingual Learning
Paper ⢠2312.06134 ⢠Published ⢠3 -
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper ⢠2312.11514 ⢠Published ⢠257
-
Cognitive Architectures for Language Agents
Paper ⢠2309.02427 ⢠Published ⢠8 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper ⢠2305.18290 ⢠Published ⢠58 -
Orca 2: Teaching Small Language Models How to Reason
Paper ⢠2311.11045 ⢠Published ⢠75 -
Pretraining Data Mixtures Enable Narrow Model Selection Capabilities in Transformer Models
Paper ⢠2311.00871 ⢠Published ⢠2
-
Attention Is All You Need
Paper ⢠1706.03762 ⢠Published ⢠61 -
FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
Paper ⢠2307.08691 ⢠Published ⢠8 -
Mixtral of Experts
Paper ⢠2401.04088 ⢠Published ⢠159 -
Mistral 7B
Paper ⢠2310.06825 ⢠Published ⢠48