1 link tagged with all of: language-models + long-context + ropemethods
Click any tag below to further narrow down your results
Links
Long-context large language models (LLMs) have made significant progress due to methods such as Rotary Position Embedding (RoPE). This paper analyzes various attention mechanisms, revealing performance limitations of RoPE and proposing a new hybrid attention architecture that effectively combines global and local attention spans, resulting in improved performance and efficiency for long-context tasks.