Skip to main content

🏒 University of Washington

Byte Latent Transformer: Patches Scale Better Than Tokens
·3839 words·19 mins· loading · loading
AI Generated πŸ€— Daily Papers Natural Language Processing Large Language Models 🏒 University of Washington
BLT: λ°”μ΄νŠΈ 기반 LLM, 토큰보닀 패치 μš°μ„ .