2025-01-29s
2025
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training
·3663 words·18 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 UC Berkeley
Reinforcement learning (RL) surpasses supervised fine-tuning (SFT) in fostering generalization in foundation models, while SFT aids RL’s stability; a comparative study across text and visual domains r…
Over-Tokenized Transformer: Vocabulary is Generally Worth Scaling
·3794 words·18 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Seed-Foundation-Model Team, Bytedance
Boosting Large Language Model (LLM) performance, researchers introduce Over-Tokenized Transformers, decoupling input/output vocabularies to improve language modeling. Scaling input vocabularies improv…
Optimizing Large Language Model Training Using FP4 Quantization
·1562 words·8 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Microsoft Research
First-ever FP4 training framework for LLMs achieves accuracy comparable to BF16 and FP8, enabling efficient ultra-low precision training.
Histoires Morales: A French Dataset for Assessing Moral Alignment
·8270 words·39 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Laboratoire Hubert Curien
HISTOIRESMORALES: a new French dataset tackles the crucial issue of aligning language models with human moral values, providing valuable resources for ethical AI research in a previously underserved l…
DiffSplat: Repurposing Image Diffusion Models for Scalable Gaussian Splat Generation
·3227 words·16 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Computer Vision
Image Generation
🏢 Peking University
DIFFSPLAT repurposes 2D image diffusion models to natively generate high-quality 3D Gaussian splats, overcoming limitations in existing 3D generation methods.
IndicMMLU-Pro: Benchmarking Indic Large Language Models on Multi-Task Language Understanding
·2564 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Artificial Intelligence Institute, University of South Carolina
IndicMMLU-Pro: a new benchmark rigorously evaluates large language models’ multi-task language understanding capabilities across nine major Indian languages, pushing Indic language AI research forward…
Low-Rank Adapters Meet Neural Architecture Search for LLM Compression
·2154 words·11 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Intel Labs
Low-rank adapters combined with neural architecture search revolutionize LLM compression, enabling efficient fine-tuning and significantly reduced memory footprint.