Skip to main content

Natural Language Processing

Fietje: An open, efficient LLM for Dutch
·2556 words·12 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข KU Leuven
Fietje: ์˜คํ”ˆ์†Œ์Šค ์†Œํ˜• ๋„ค๋œ๋ž€๋“œ์–ด LLM ๊ณต๊ฐœ!
AceMath: Advancing Frontier Math Reasoning with Post-Training and Reward Modeling
·2682 words·13 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข NVIDIA Research
AceMath๋Š” ์‚ฌ์ „ ํ›ˆ๋ จ ๋ฐ ๋ณด์ƒ ๋ชจ๋ธ๋ง์„ ํ†ตํ•ด ์ตœ์ฒจ๋‹จ ์ˆ˜ํ•™ ์ถ”๋ก  ๋Šฅ๋ ฅ์„ ๋‹ฌ์„ฑํ•œ ํ”„๋Ÿฐํ‹ฐ์–ด๊ธ‰ ๋ชจ๋ธ ์‹œ๋ฆฌ์ฆˆ์ž…๋‹ˆ๋‹ค.
TheAgentCompany: Benchmarking LLM Agents on Consequential Real World Tasks
·2422 words·12 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Carnegie Mellon University
TheAgentCompany ๋ฒค์น˜๋งˆํฌ๋Š” ์‹ค์ œ ์†Œํ”„ํŠธ์›จ์–ด ํšŒ์‚ฌ ํ™˜๊ฒฝ์„ ๋ชจ๋ฐฉํ•˜์—ฌ LLM ์—์ด์ „ํŠธ์˜ ์‹ค์ œ ์—…๋ฌด ์ˆ˜ํ–‰ ๋Šฅ๋ ฅ์„ ํ‰๊ฐ€ํ•˜๋ฉฐ, AI ์—์ด์ „ํŠธ์˜ ํ˜„์‹ค ์„ธ๊ณ„ ์ ์šฉ ๊ฐ€๋Šฅ์„ฑ๊ณผ ํ•œ๊ณ„๋ฅผ ๋ณด์—ฌ์ค๋‹ˆ๋‹ค.
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for Fast, Memory Efficient, and Long Context Finetuning and Inference
·2449 words·12 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Answer.AI
ModernBERT: ๋น ๋ฅด๊ณ  ๋ฉ”๋ชจ๋ฆฌ ํšจ์œจ์ ์ธ ์žฅ๋ฌธ ์ปจํ…์ŠคํŠธ ๋ฏธ์„ธ ์กฐ์ • ๋ฐ ์ถ”๋ก ์„ ์œ„ํ•œ ์ตœ์ฒจ๋‹จ ์–‘๋ฐฉํ–ฅ ์ธ์ฝ”๋”!
RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment
·2978 words·14 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข University of Chinese Academy of Sciences
RAG-RewardBench: RAG ํ™˜๊ฒฝ์—์„œ ๋ณด์ƒ ๋ชจ๋ธ ํ‰๊ฐ€๋ฅผ ์œ„ํ•œ ์ตœ์ดˆ์˜ ๋ฒค์น˜๋งˆํฌ ์ œ์‹œ!
AntiLeak-Bench: Preventing Data Contamination by Automatically Constructing Benchmarks with Updated Real-World Knowledge
·3149 words·15 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Nanyang Technological University
AntiLeak-Bench: ์ž๋™ํ™”๋œ ๋ฒค์น˜๋งˆํ‚น์œผ๋กœ LLM ๋ฐ์ดํ„ฐ ์˜ค์—ผ ๋ฐฉ์ง€
DateLogicQA: Benchmarking Temporal Biases in Large Language Models
·2927 words·14 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข University of Aberdeen
DateLogicQA: LLM์˜ ์‹œ๊ฐ„์  ์ถ”๋ก  ํŽธํ–ฅ ๋ฒค์น˜๋งˆํฌ ์ œ์‹œ! ํ† ํฐํ™”, ํ‘œ์ƒ ๋ฐ ๋…ผ๋ฆฌ ์ˆ˜์ค€ ํŽธํ–ฅ ๋ถ„์„์œผ๋กœ ์‹œ๊ฐ„์  ๋ฐ์ดํ„ฐ ์ฒ˜๋ฆฌ ๊ฐœ์„  ๋ฐฉ์•ˆ ์ œ์‹œ!
Whisper-GPT: A Hybrid Representation Audio Large Language Model
·1322 words·7 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Stanford University
Whisper-GPT: ํ•˜์ด๋ธŒ๋ฆฌ๋“œ ์Œ์„ฑ ๋ฐ ์Œ์•… LLM์œผ๋กœ, ์—ฐ์† ์˜ค๋””์˜ค์™€ ์ด์‚ฐ ํ† ํฐ์„ ๊ฒฐํ•ฉํ•˜์—ฌ ํ–ฅ์ƒ๋œ ์„ฑ๋Šฅ์„ ์ œ๊ณตํ•ฉ๋‹ˆ๋‹ค.
The Open Source Advantage in Large Language Models (LLMs)
·248 words·2 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Rollins College
์˜คํ”ˆ์†Œ์Šค LLM, ํ์‡„ํ˜• LLM ๋Œ€๋น„ ํˆฌ๋ช…์„ฑ๊ณผ ์ ‘๊ทผ์„ฑ์€ ๋†’์ง€๋งŒ, ์„ฑ๋Šฅ์€ ๋‚ฎ์Œ. ํ•˜์ด๋ธŒ๋ฆฌ๋“œ ์ „๋žต์ด ๋ฏธ๋ž˜.
SPaR: Self-Play with Tree-Search Refinement to Improve Instruction-Following in Large Language Models
·3260 words·16 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Tsinghua University
Self-play with refinement boosts instruction-following in LLMs.
SepLLM: Accelerate Large Language Models by Compressing One Segment into One Separator
·2998 words·15 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Huawei Noah's Ark Lab
SepLLM์€ ํŠน์ˆ˜ ํ† ํฐ์˜ ์ค‘์š”์„ฑ์„ ํ™œ์šฉํ•˜์—ฌ LLM ์ถ”๋ก ์„ ๊ฐ€์†ํ™”ํ•˜๊ณ  ๊ธด ์‹œํ€€์Šค๋ฅผ ํšจ์œจ์ ์œผ๋กœ ์ฒ˜๋ฆฌํ•ฉ๋‹ˆ๋‹ค.
RetroLLM: Empowering Large Language Models to Retrieve Fine-grained Evidence within Generation
·3747 words·18 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Question Answering ๐Ÿข Renmin University of China
RetroLLM: ๊ฒ€์ƒ‰๊ณผ ์ƒ์„ฑ์„ ํ†ตํ•ฉํ•œ RAG ์‹œ์Šคํ…œ
Smaller Language Models Are Better Instruction Evolvers
·4310 words·21 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Beijing University of Posts and Telecommunications
์†Œํ˜• ์–ธ์–ด ๋ชจ๋ธ์ด ๋” ๋‚˜์€ ๋ช…๋ น ์ƒ์„ฑ์ž!
Reliable, Reproducible, and Really Fast Leaderboards with Evalica
·1243 words·6 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข JetBrains
Evalica: ๋ฒค์น˜๋งˆํ‚น์„ ์‰ฝ๊ณ  ๋น ๋ฅด๊ณ  ์‹ ๋ขฐํ•  ์ˆ˜ ์žˆ๊ฒŒ ๋งŒ๋“œ๋Š” ํˆดํ‚ท
SCBench: A KV Cache-Centric Analysis of Long-Context Methods
·4642 words·22 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Microsoft Corporation
SCBench๋Š” ๋ฉ€ํ‹ฐํ„ด ๋ฐ ๋ฉ€ํ‹ฐ๋ฆฌํ€˜์ŠคํŠธ ์‹œ๋‚˜๋ฆฌ์˜ค์—์„œ ์žฅ๋ฌธ ๋งฅ๋ฝ ๋ฉ”์„œ๋“œ๋ฅผ ํ‰๊ฐ€ํ•˜๋Š” ์ƒˆ๋กœ์šด ๋ฒค์น˜๋งˆํฌ์ž…๋‹ˆ๋‹ค.
Large Action Models: From Inception to Implementation
·2067 words·10 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Microsoft
LLM์—์„œ LAM์œผ๋กœ: ์‹ค์ œ ์ž‘์—…์„ ์ˆ˜ํ–‰ํ•˜๋Š” AI ์—์ด์ „ํŠธ ๊ตฌ์ถ•.
Byte Latent Transformer: Patches Scale Better Than Tokens
·3839 words·19 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข University of Washington
BLT: ๋ฐ”์ดํŠธ ๊ธฐ๋ฐ˜ LLM, ํ† ํฐ๋ณด๋‹ค ํŒจ์น˜ ์šฐ์„ .
Phi-4 Technical Report
·2236 words·11 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Microsoft Research
Phi-4: 140์–ต ๋งค๊ฐœ๋ณ€์ˆ˜ ์–ธ์–ด ๋ชจ๋ธ์€ ๋ฐ์ดํ„ฐ ํ’ˆ์งˆ์— ์ค‘์ ์„ ๋‘” ํ›ˆ๋ จ ๋ ˆ์‹œํ”ผ๋กœ ๊ฐœ๋ฐœ๋˜์–ด ์ถ”๋ก  ๋Šฅ๋ ฅ์„ ๋Œ€ํญ ํ–ฅ์ƒ์‹œ์ผฐ์Šต๋‹ˆ๋‹ค.
GReaTer: Gradients over Reasoning Makes Smaller Language Models Strong Prompt Optimizers
·7101 words·34 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Pennsylvania State University
GREATER๋Š” ์ถ”๋ก ์— ๋Œ€ํ•œ ๊ทธ๋ ˆ์ด๋””์–ธํŠธ๋ฅผ ํ™œ์šฉํ•˜์—ฌ ์†Œ๊ทœ๋ชจ ์–ธ์–ด ๋ชจ๋ธ์˜ ํ”„๋กฌํ”„ํŠธ๋ฅผ ์ตœ์ ํ™”ํ•˜์—ฌ ๋Œ€๊ทœ๋ชจ LLM ์—†์ด๋„ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚ต๋‹ˆ๋‹ค.
SmolTulu: Higher Learning Rate to Batch Size Ratios Can Lead to Better Reasoning in SLMs
·2378 words·12 mins· loading · loading
AI Generated ๐Ÿค— Daily Papers Natural Language Processing Large Language Models ๐Ÿข Saudi Data & Artificial Intelligence Authority
Smaller language models reason better with fine-tuned training recipes.