Skip to main content

LLM

Infini-Attention Paper Review
·438 words·3 mins
Infini-Attention introduces a novel approach to scaling Transformer models for infinitely long inputs while managing memory and computation.