Posters
2024
ALPINE: Unveiling The Planning Capability of Autoregressive Learning in Language Models
·2122 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
ALPINE reveals how Transformer-based LLMs learn planning by embedding graph information into their weights, but also highlights their inability to handle transitive relationships.
AlphaTablets: A Generic Plane Representation for 3D Planar Reconstruction from Monocular Videos
·2409 words·12 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 Tsinghua University
AlphaTablets revolutionizes 3D planar reconstruction from monocular videos with its novel rectangle-based representation featuring continuous surfaces and precise boundaries, achieving state-of-the-ar…
AlphaPruning: Using Heavy-Tailed Self Regularization Theory for Improved Layer-wise Pruning of Large Language Models
·3772 words·18 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Nankai University
AlphaPruning leverages Heavy-Tailed Self-Regularization theory to allocate optimal layer-wise sparsity ratios in LLMs, achieving 80% sparsity in LLaMA-7B with reasonable perplexity.
AlphaMath Almost Zero: Process Supervision without Process
·2731 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Tongyi Lab
AlphaMath: LLMs excel at math reasoning without human-annotated process supervision, using Monte Carlo Tree Search.
Almost-Linear RNNs Yield Highly Interpretable Symbolic Codes in Dynamical Systems Reconstruction
·3184 words·15 mins·
loading
·
loading
AI Theory
Interpretability
🏢 Dept. of Theoretical Neuroscience, Central Institute of Mental Health, Medical Faculty, Heidelberg University, Germany
Almost-linear RNNs (AL-RNNs) offer highly interpretable symbolic codes for dynamical systems reconstruction, simplifying the analysis of complex systems.
Almost Surely Asymptotically Constant Graph Neural Networks
·1976 words·10 mins·
loading
·
loading
AI Theory
Generalization
🏢 University of Oxford
Many graph neural networks (GNNs) surprisingly converge to constant outputs with increasing graph size, limiting their expressiveness.
Almost Minimax Optimal Best Arm Identification in Piecewise Stationary Linear Bandits
·1649 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 National University of Singapore
PSɛBAI+ is a near-optimal algorithm for best arm identification in piecewise stationary linear bandits, efficiently detecting changepoints and aligning contexts for improved accuracy and minimal sampl…
Almost Free: Self-concordance in Natural Exponential Families and an Application to Bandits
·359 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 University of Alberta
Generalized linear bandits with subexponential reward distributions are self-concordant, enabling second-order regret bounds free of exponential dependence on problem parameters.
Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced Optimization
·1718 words·9 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Shenzhen Institute for Advanced Study
New Hallucination-Induced Optimization (HIO) significantly reduces hallucinations in Large Vision-Language Models (LVLMs) by amplifying contrast between correct and incorrect tokens, outperforming exi…
Alleviating Distortion in Image Generation via Multi-Resolution Diffusion Models and Time-Dependent Layer Normalization
·2692 words·13 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 ByteDance
DiMR boosts image generation fidelity by cleverly combining multi-resolution networks with time-dependent layer normalization in diffusion models, achieving state-of-the-art results on ImageNet.
Alleviate Anchor-Shift: Explore Blind Spots with Cross-View Reconstruction for Incomplete Multi-View Clustering
·1873 words·9 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 National University of Defense Technology
AIMC-CVR: A novel approach that alleviates anchor-shift in incomplete multi-view clustering via cross-view reconstruction, improving accuracy and scalability.
All-in-One Image Coding for Joint Human-Machine Vision with Multi-Path Aggregation
·5531 words·26 mins·
loading
·
loading
AI Generated
Computer Vision
Image Coding
🏢 Nanjing University
Multi-Path Aggregation (MPA) achieves comparable performance to state-of-the-art methods in multi-task image coding, by unifying feature representations with a novel all-in-one architecture and a two-…
Alignment for Honesty
·3666 words·18 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
This paper introduces a novel framework for aligning LLMs with honesty, proposing new metrics and training techniques to make LLMs more truthful and less prone to confidently incorrect responses.
Alignment at Pre-training! Towards Native Alignment for Arabic LLMs
·2342 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 King Abdullah University of Science and Technology
This study introduces ’native alignment’ for Arabic LLMs, achieving state-of-the-art results by aligning models during pre-training, rather than post-training.
Aligning Vision Models with Human Aesthetics in Retrieval: Benchmarks and Algorithms
·4509 words·22 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Southeast University
This paper presents a novel method to align vision models with human aesthetics in image retrieval, using large language models (LLMs) for query rephrasing and preference-based reinforcement learning …
Aligning to Thousands of Preferences via System Message Generalization
·3279 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 KAIST AI
JANUS, a 7B LLM, achieves high alignment to thousands of user preferences by generalizing from diverse system messages, outperforming existing LLMs on various benchmarks.
Aligning Target-Aware Molecule Diffusion Models with Exact Energy Optimization
·2379 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Stanford University
ALIDIFF aligns target-aware molecule diffusion models with exact energy optimization, generating molecules with state-of-the-art binding energies and improved properties.
Aligning Model Properties via Conformal Risk Control
·1981 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Safety
🏢 Stanford University
Post-processing pre-trained models for alignment using conformal risk control and property testing guarantees better alignment, even when training data is biased.
Aligning LLM Agents by Learning Latent Preference from User Edits
·2688 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Microsoft Research
PRELUDE, a novel framework, leverages user edits of LLM outputs to learn latent preferences, improving agent alignment and minimizing edit costs. CIPHER, its efficient algorithm, infers preferences f…
Aligning Large Language Models with Representation Editing: A Control Perspective
·2249 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Cornell University
RE-Control: Aligning LLMs via dynamic representation editing using optimal control theory, achieving superior alignment with significantly fewer resources than fine-tuning.