2-D Transformer: Extending Large Language Models to Long-Context With Few Memory.
2-D Transformer:擴展大型語言模型以應對長上下文與少量記憶。
IEEE Trans Neural Netw Learn Syst 2025-03-21
Efficient scaling of large language models with mixture of experts and 3D analog in-memory computing.
大型語言模型的高效擴展:專家混合與3D類比內存計算。
Nat Comput Sci 2025-01-08
Optimizing Large Language Models in Radiology and Mitigating Pitfalls: Prompt Engineering and Fine-tuning.
優化放射學中的大型語言模型及減少陷阱:提示工程與微調。
Radiographics 2025-03-06