
The provided text is an excerpt from the pre-print service arXiv, promoting its support for Open Access Week while presenting information about a new paper submission. The paper, titled "Glyph: Scaling Context Windows via Visual-Text Compression," proposes a novel framework called Glyph that addresses the computational challenges of large language models (LLMs) with extensive context windows by rendering long texts into images for processing by vision-language models (VLMs). The authors state that this visual approach achieves significant token compression (3-4x faster prefilling and decoding) while maintaining accuracy, potentially allowing 1M-token-level text tasks to be handled by smaller 128K-context VLMs. The entry includes bibliographic details, submission history, links to access the paper(PDF/HTML), and various citation and code-related tools, all within the context of Computer Vision and Pattern Recognition.