Skip to main content

Posters

2024

How to Continually Adapt Text-to-Image Diffusion Models for Flexible Customization?
·2761 words·13 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 Mohamed Bin Zayed University of Artificial Intelligence
Concept-Incremental Flexible Customization (CIFC) model tackles catastrophic forgetting and concept neglect in continually adapting text-to-image diffusion models, enabling flexible personalization.
How to Boost Any Loss Function
·3432 words·17 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Google Research
Boosting, traditionally limited by assumptions about loss functions, is proven in this paper to efficiently optimize any loss function regardless of differentiability or convexity.
How Sparse Can We Prune A Deep Network: A Fundamental Limit Perspective
·2596 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 Huazhong University of Science and Technology
Deep network pruning’s fundamental limits are characterized, revealing how weight magnitude and network sharpness determine the maximum achievable sparsity.
How Molecules Impact Cells: Unlocking Contrastive PhenoMolecular Retrieval
·3595 words·17 mins· loading · loading
Multimodal Learning Cross-Modal Retrieval 🏢 University of Toronto
MolPhenix, a novel multi-modal model, drastically improves zero-shot molecular retrieval by leveraging a pre-trained phenomics model and a novel similarity-aware loss, achieving an 8.1x improvement ov…
How many classifiers do we need?
·1821 words·9 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Berkeley
Boost ensemble accuracy by predicting performance with fewer classifiers using a novel polarization law and refined error bounds.
How Far Can Transformers Reason? The Globality Barrier and Inductive Scratchpad
·3573 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Apple
Transformers struggle with complex reasoning tasks. This paper introduces ‘globality degree’ to measure task difficulty and shows that high globality hinders efficient learning. However, using ‘induc…
How Does Variance Shape the Regret in Contextual Bandits?
·334 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 MIT
Low reward variance drastically improves contextual bandit regret, defying minimax assumptions and highlighting the crucial role of eluder dimension.
How does PDE order affect the convergence of PINNs?
·2124 words·10 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of California, Los Angeles
Higher-order PDEs hinder Physics-Informed Neural Network (PINN) convergence; this paper provides theoretical explanation and proposes variable splitting for improved accuracy.
How Does Message Passing Improve Collaborative Filtering?
·1963 words·10 mins· loading · loading
Machine Learning Representation Learning 🏢 University of California, Riverside
TAG-CF boosts collaborative filtering accuracy by up to 39.2% on cold users, using only a single message-passing step at test time, avoiding costly training-time computations.
How does Inverse RL Scale to Large State Spaces? A Provably Efficient Approach
·1501 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Politecnico Di Milano
CATY-IRL: A novel, provably efficient algorithm solves Inverse Reinforcement Learning’s scalability issues for large state spaces, improving upon state-of-the-art methods.
How does Gradient Descent Learn Features --- A Local Analysis for Regularized Two-Layer Neural Networks
·1326 words·7 mins· loading · loading
AI Theory Optimization 🏢 University of Washington
Neural networks learn features effectively through gradient descent, not just at the beginning, but also at the end of training, even with carefully regularized objectives.
How Does Black-Box Impact the Learning Guarantee of Stochastic Compositional Optimization?
·1242 words·6 mins· loading · loading
AI Theory Optimization 🏢 Huazhong Agricultural University
This study reveals how black-box settings affect the learning guarantee of stochastic compositional optimization, offering sharper generalization bounds and novel learning guarantees for derivative-fr…
How does Architecture Influence the Base Capabilities of Pre-trained Language Models? A Case Study Based on FFN-Wider and MoE Transformers
·3480 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Research Center for Social Computing and Information Retrieval, Harbin Institute of Technology
Pre-trained language models’ base capabilities are significantly influenced by architecture, not just scale; a novel Combination Enhanced Architecture (CEA) improves performance by addressing FFN-Wide…
How do Large Language Models Handle Multilingualism?
·2895 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 DAMO Academy, Alibaba Group, Singapore
LLMs surprisingly process multilingual queries via an English-centric intermediate stage before generating responses in the original language, a phenomenon explained by the proposed MWork framework an…
How Do Large Language Models Acquire Factual Knowledge During Pretraining?
·4632 words·22 mins· loading · loading
Natural Language Processing Large Language Models 🏢 KAIST
LLMs’ factual knowledge acquisition during pretraining is surprisingly non-linear: more data doesn’t guarantee better knowledge retention, and forgetting follows a power law.
How Diffusion Models Learn to Factorize and Compose
·3926 words·19 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 MIT
Diffusion models surprisingly learn factorized representations, enabling compositional generalization, but struggle with interpolation; training with independent factors drastically improves data effi…
How Control Information Influences Multilingual Text Image Generation and Editing?
·2075 words·10 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 University of Science and Technology of China
TextGen enhances multilingual visual text generation and editing by optimizing control information using Fourier analysis and a two-stage framework, achieving state-of-the-art results.
HORSE: Hierarchical Representation for Large-Scale Neural Subset Selection
·1821 words·9 mins· loading · loading
Machine Learning Deep Learning 🏢 Chinese University of Hong Kong
HORSE: A novel attention-based neural network significantly improves large-scale neural subset selection by up to 20%, addressing limitations in existing methods.
HOPE: Shape Matching Via Aligning Different K-hop Neighbourhoods
·1940 words·10 mins· loading · loading
Computer Vision 3D Vision 🏢 Hong Kong University of Science and Technology
HOPE: a novel shape matching method achieving both accuracy and smoothness by aligning different k-hop neighborhoods and refining maps via local map distortion.
HonestLLM: Toward an Honest and Helpful Large Language Model
·3514 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Peking University
HonestLLM boosts LLM honesty & helpfulness by 65.3% (Llama3-8b) and 124.7% (Mistral-7b) using training-free and fine-tuning methods, establishing principles and a new dataset (HONESET) for honesty eva…