Apple、限られたメモリで効率的な大規模言語モデル推論を行う方法を開発
2023.12.20 22:35
Macotakara.jp
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
AppleのMachine Learning Researchは、arXivに「LLM in a flash: Efficient Large Language Model Inference with Limited Memory」…