According to Meta, memory layers may be the the answer to LLM hallucinations as they don't require huge compute resources at inference time.