2-D Transformer: Extending Large Language Models to Long-Context With Few Memory.
2-D Transformer:擴展大型語言模型以應對長上下文與少量記憶。
IEEE Trans Neural Netw Learn Syst 2025-03-21
Enabling Energy-Efficient Deployment of Large Language Models on Memristor Crossbar: A Synergy of Large and Small.
在 Memristor Crossbar 上實現大型語言模型的能源效率部署:大與小的協同作用。
IEEE Trans Pattern Anal Mach Intell 2024-10-18
Efficient scaling of large language models with mixture of experts and 3D analog in-memory computing.
大型語言模型的高效擴展:專家混合與3D類比內存計算。
Nat Comput Sci 2025-01-08