The provided text is an excerpt from the pre-print service arXiv, promoting its support for Open Access Week while presenting information about a new paper submission. The paper, titled "Glyph: Scaling Context Windows via Visual-Text Compression," proposes a novel framework called Glyph that addresses the computational challenges of large language models (LLMs) with extensive context windows by rendering long texts into images for processing by vision-language models (VLMs). The authors state that this visual approach achieves significant token compression (3-4x faster prefilling and decoding) while maintaining accuracy, potentially allowing 1M-token-level text tasks to be handled by smaller 128K-context VLMs. The entry includes bibliographic details, submission history, links to access the paper(PDF/HTML), and various citation and code-related tools, all within the context of Computer Vision and Pattern Recognition.
資訊
- 節目
- 頻率每週更新
- 發佈時間2025年11月2日 下午12:18 [UTC]
- 長度16 分鐘
- 年齡分級兒少適宜
