MIT researchers developed Attention Matching, a KV cache compaction technique that compresses LLM memory by 50x in seconds — ...
In a computer, the entire memory can be separated into different levels based on access time and capacity. Figure 1 shows different levels in the memory hierarchy. Smaller and faster memories are kept ...