2 links tagged with all of: machine-learning + language-models + reasoning
Click any tag below to further narrow down your results
Links
This article introduces Dynamic Large Concept Models (DLCM), a new framework that enhances language processing by shifting focus from individual tokens to broader concepts. It learns semantic boundaries and reallocates computational resources for better reasoning, achieving improvements in language model performance on various benchmarks.
Recent advancements in large language models (LLMs) have prompted discussions about their reasoning capabilities. This study introduces a representation engineering approach that leverages model activations to create control vectors, enhancing reasoning performance on various tasks without additional training. The results indicate that modulating model activations can effectively improve LLMs' reasoning abilities.