Attention System 2 Attention (is something you might need too) Paper • 2311.11829 • Published Nov 20, 2023 • 44 Transformers are Multi-State RNNs Paper • 2401.06104 • Published Jan 11, 2024 • 39 The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
System 2 Attention (is something you might need too) Paper • 2311.11829 • Published Nov 20, 2023 • 44
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
Transformers Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs Paper • 2403.20041 • Published Mar 29, 2024 • 34
Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs Paper • 2403.20041 • Published Mar 29, 2024 • 34
Mamba+Transformers Jamba: A Hybrid Transformer-Mamba Language Model Paper • 2403.19887 • Published Mar 28, 2024 • 111
Attention System 2 Attention (is something you might need too) Paper • 2311.11829 • Published Nov 20, 2023 • 44 Transformers are Multi-State RNNs Paper • 2401.06104 • Published Jan 11, 2024 • 39 The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
System 2 Attention (is something you might need too) Paper • 2311.11829 • Published Nov 20, 2023 • 44
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
Mamba+Transformers Jamba: A Hybrid Transformer-Mamba Language Model Paper • 2403.19887 • Published Mar 28, 2024 • 111
Transformers Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs Paper • 2403.20041 • Published Mar 29, 2024 • 34
Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs Paper • 2403.20041 • Published Mar 29, 2024 • 34