Meta AI researchers propose scalable memory layers to improve the factual knowledge and reduce hallucinations in large language models (LLMs) by enhancing their learning capacity without additional compute resources. These layers use sparse activations and key-value lookup mechanisms, making them more memory-intensive but
Sort: