3 links
tagged with all of: language-models + long-context
Click any tag below to further narrow down your results
Links
HELMET (How to Evaluate Long-Context Models Effectively and Thoroughly) is introduced as a comprehensive benchmark for evaluating long-context language models (LCLMs), addressing limitations in existing evaluation methods. The blog outlines HELMET's design, key findings from evaluations of 59 recent LCLMs, and offers a quickstart guide for practitioners to utilize HELMET in their research and applications.
Long-context large language models (LLMs) have made significant progress due to methods such as Rotary Position Embedding (RoPE). This paper analyzes various attention mechanisms, revealing performance limitations of RoPE and proposing a new hybrid attention architecture that effectively combines global and local attention spans, resulting in improved performance and efficiency for long-context tasks.
The article presents EntropyLong, a novel method for training long-context language models by utilizing predictive uncertainty to verify the quality of long-range dependencies. This approach constructs training samples by combining original documents with semantically relevant contexts, leading to significant improvements in tasks requiring distant information according to the RULER benchmarks and LongBenchv2. The study emphasizes the effectiveness of entropy-based verification in enhancing long-context understanding in machine learning models.