Paper Reviews by AI
2025
Is Safety Standard Same for Everyone? User-Specific Safety Evaluation of Large Language Models
·5119 words·25 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Safety
🏢 KAIST
LLMs fail to act safely when considering user-specific safety standards, which were made to be solved via new benchmark.
InterFeedback: Unveiling Interactive Intelligence of Large Multimodal Models via Human Feedback
·3063 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Multimodal Learning
Human-AI Interaction
🏢 National University of Singapore
InterFeedback: LMMs need better human feedback to enhance AI assistants!
How Much Knowledge Can You Pack into a LoRA Adapter without Harming LLM?
·2645 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 AIRI
Packing new knowledge into LoRA adapters can harm LLMs! A delicate balance is needed to prevent performance decline.
FR-Spec: Accelerating Large-Vocabulary Language Models via Frequency-Ranked Speculative Sampling
·3445 words·17 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Text Generation
🏢 Tsinghua University
FR-Spec: Frequency-Ranked Speculative Sampling accelerates LLMs by optimizing vocabulary space compression, achieving 1.12x speedup over EAGLE-2.
Dynamic Concepts Personalization from Single Videos
·2668 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Computer Vision
Video Understanding
🏢 SNAP RESEARCH
Personalizing video models for dynamic concepts is now achievable with Set-and-Sequence: enabling high-fidelity generation, editing, and composition!
Does Time Have Its Place? Temporal Heads: Where Language Models Recall Time-specific Information
·4876 words·23 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Korea University
LLMs have ‘Temporal Heads’ that process time-specific facts!
Discovering highly efficient low-weight quantum error-correcting codes with reinforcement learning
·6573 words·31 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Optimization
🏢 University of Texas at Austin
RL optimizes quantum error-correcting codes, slashing physical qubit overhead for fault-tolerant quantum computing.
CrossOver: 3D Scene Cross-Modal Alignment
·5760 words·28 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Computer Vision
Scene Understanding
🏢 Stanford University
CrossOver: Flexible scene-level cross-modal alignment via modality-agnostic embeddings, unlocking robust 3D scene understanding.
AlphaMaze: Enhancing Large Language Models' Spatial Intelligence via GRPO
·402 words·2 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Multimodal Learning
Vision-Language Models
🏢 Menlo Research
AlphaMaze enhances LLMs’ spatial intelligence via GRPO, achieving 93% accuracy in maze navigation and showing emergent reasoning.
Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region
·2482 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
AI Theory
Safety
🏢 Hong Kong Polytechnic University
Aligned LLMs’ safety often anchors in the template region, creating vulnerabilities. Detaching safety mechanisms shows promise in mitigation.
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
·3075 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Zhejiang University
LORAM: Train small, infer large LLMs by memory-efficient LoRA training. Enables 70B parameter model training on a 20G HBM GPU, replacing A100-80G. Reduces parameter storage cost by 15.81x.
Slamming: Training a Speech Language Model on One GPU in a Day
·2787 words·14 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Speech and Audio
Speech Synthesis
🏢 Hebrew University of Jerusalem
Slam: Train SLMs on one GPU in a day!
SIFT: Grounding LLM Reasoning in Contexts via Stickers
·3144 words·15 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Shanghai Jiao Tong University
SIFT: Grounds LLM reasoning with ‘Stickers’ to highlight context and improve accuracy without extra training.
REFIND: Retrieval-Augmented Factuality Hallucination Detection in Large Language Models
·582 words·3 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Pohang University of Science and Technology
REFIND: Detects LLM hallucinations by directly leveraging retrieved documents, using a novel Context Sensitivity Ratio.
Noise May Contain Transferable Knowledge: Understanding Semi-supervised Heterogeneous Domain Adaptation from an Empirical Perspective
·6916 words·33 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Machine Learning
Transfer Learning
🏢 Beijing Teleinfo Technology Company Ltd., China Academy of Information and Communications Technology
Unveiling the surprising potential of noise: transferable knowledge in semi-supervised heterogeneous domain adaptation (SHDA).
MoM: Linear Sequence Modeling with Mixture-of-Memories
·2764 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 Shanghai AI Laboratory
MoM: Enhancing linear sequence modeling via mixture-of-memories for improved recall and reduced memory interference.
LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization
·2370 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 National University of Singapore
LongPO: Self-evolve LLMs to excel in long contexts via short-to-long preference optimization, boosting performance without sacrificing short-context skills.
JL1-CD: A New Benchmark for Remote Sensing Change Detection and a Robust Multi-Teacher Knowledge Distillation Framework
·3675 words·18 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Computer Vision
Image Segmentation
🏢 Tsinghua University
JL1-CD: New all-inclusive dataset & multi-teacher knowledge distillation framework for robust remote sensing change detection, achieving state-of-the-art results!
Is That Your Final Answer? Test-Time Scaling Improves Selective Question Answering
·2478 words·12 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Question Answering
🏢 Johns Hopkins University
Test-time scaling + confidence = better QA!
Geolocation with Real Human Gameplay Data: A Large-Scale Dataset and Human-Like Reasoning Framework
·2585 words·13 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Computer Vision
Scene Understanding
🏢 MBZUAI
New geolocation dataset & reasoning framework enhance accuracy and interpretability by leveraging human gameplay data.