News Score: Score the News, Sort the News, Rewrite the Headlines

Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens

Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens (via) Very significant new release from Alibaba's Qwen team. Their openly licensed (sometimes Apache 2, sometimes Qwen license, I've had trouble keeping up) Qwen 2.5 LLM previously had an input token limit of 128,000 tokens. This new model increases that to 1 million, using a new technique called Dual Chunk Attention, first described in this paper from February 2024. They've released two models on Hugging Face: Qwen2.5-7B-Instr...

Read more at simonwillison.net

© News Score  score the news, sort the news, rewrite the headlines