Skip to main content

Posters

2024

A two-scale Complexity Measure for Deep Learning Models
·1709 words·9 mins· loading · loading
Machine Learning Deep Learning 🏢 IBM Research
New 2sED measure effectively bounds deep learning model complexity, correlating well with training error and offering efficient computation, particularly for deep models via a layerwise approach.
A Tractable Inference Perspective of Offline RL
·2824 words·14 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Peking University
Trifle: Tractable inference for Offline RL achieves state-of-the-art results by using tractable generative models to overcome the inference-time suboptimality of existing sequence modeling approaches.
A Topology-aware Graph Coarsening Framework for Continual Graph Learning
·2623 words·13 mins· loading · loading
Machine Learning Deep Learning 🏢 Stevens Institute of Technology
TACO, a novel topology-aware graph coarsening framework, tackles catastrophic forgetting in continual graph learning by efficiently preserving topological information during experience replay, signifi…
A Theory of Optimistically Universal Online Learnability for General Concept Classes
·411 words·2 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Purdue University
This paper fully characterizes concept classes optimistically universally learnable online, introducing novel algorithms and revealing equivalences between agnostic and realizable settings.
A Theoretical Understanding of Self-Correction through In-context Alignment
·1997 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 MIT CSAIL
LLMs improve through self-correction, but the mechanisms are unclear. This paper provides a theoretical framework and empirical evidence demonstrating that self-correction arises from in-context align…
A Theoretical Perspective for Speculative Decoding Algorithm
·1873 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Princeton University
This paper theoretically analyzes speculative decoding, revealing its optimality and providing formulas for expected rejections, paving the way for more efficient large language model inference.
A theoretical design of concept sets: improving the predictability of concept bottleneck models
·1648 words·8 mins· loading · loading
AI Theory Interpretability 🏢 University of Cambridge
Boosting concept bottleneck model predictability, this paper introduces a theoretical framework linking concept set properties to model performance, proposing a method for effective concept identifica…
A theoretical case-study of Scalable Oversight in Hierarchical Reinforcement Learning
·414 words·2 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Carnegie Mellon University
Bounded human feedback hinders large AI model training. This paper introduces hierarchical reinforcement learning to enable scalable oversight, efficiently acquiring feedback and learning optimal poli…
A teacher-teacher framework for clinical language representation learning
·1643 words·8 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Harvard University
A lightweight knowledge alignment module enables two pre-trained LLMs to mutually learn and improve clinical language representation, exceeding individual model performance on various downstream tasks…
A Swiss Army Knife for Heterogeneous Federated Learning: Flexible Coupling via Trace Norm
·2684 words·13 mins· loading · loading
Machine Learning Federated Learning 🏢 Sun Yat-Sen University
FedSAK, a novel federated multi-task learning framework, flexibly handles data, model, and task heterogeneity using tensor trace norm to learn correlations among client models, achieving superior perf…
A Surprisingly Simple Approach to Generalized Few-Shot Semantic Segmentation
·2178 words·11 mins· loading · loading
AI Generated Computer Vision Image Segmentation 🏢 IBM Research
Simple rule-based base-class mining (BCM) significantly boosts generalized few-shot semantic segmentation (GFSS) performance, surpassing complex existing methods.
A Structure-Aware Framework for Learning Device Placements on Computation Graphs
·1503 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Intel Labs
Learn optimal device placement for neural networks with HSDAG, a novel framework boosting inference speed by up to 58.2%!
A Sober Look at the Robustness of CLIPs to Spurious Features
·4936 words·24 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Hong Kong Baptist University
CounterAnimal: a new dataset exposes CLIP’s reliance on spurious correlations, challenging its perceived robustness and highlighting the need for more comprehensive evaluation benchmarks in vision-lan…
A Single-Step, Sharpness-Aware Minimization is All You Need to Achieve Efficient and Accurate Sparse Training
·2195 words·11 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Clemson University
Single-step Sharpness-Aware Minimization (S2-SAM) achieves efficient and accurate sparse training by approximating sharpness perturbation via prior gradient information, incurring zero extra cost and …
A Simple yet Universal Framework for Depth Completion
·2167 words·11 mins· loading · loading
Computer Vision 3D Vision 🏢 AI Graduate School GIST
UniDC framework achieves universal depth completion across various sensors and scenes using minimal labeled data, leveraging a foundation model and hyperbolic embedding for enhanced generalization.
A Simple yet Scalable Granger Causal Structural Learning Approach for Topological Event Sequences
·1497 words·8 mins· loading · loading
AI Theory Causality 🏢 East China Normal University
S²GCSL: A novel scalable Granger causal structural learning approach efficiently identifies root causes of telecommunication network alarms by leveraging a linear kernel and incorporating expert knowl…
A Simple Remedy for Dataset Bias via Self-Influence: A Mislabeled Sample Perspective
·3350 words·16 mins· loading · loading
AI Theory Fairness 🏢 Korea Advanced Institute of Science and Technology
This paper introduces Bias-Conditioned Self-Influence (BCSI) for precise bias-conflicting sample detection and model rectification, enhancing fairness in machine learning.
A Simple Framework for Generalization in Visual RL under Dynamic Scene Perturbations
·6610 words·32 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Ewha Womans University
SimGRL: A novel framework boosts visual reinforcement learning’s generalization by mitigating imbalanced saliency and observational overfitting through a feature-level frame stack and shifted random o…
A Simple and Optimal Approach for Universal Online Learning with Gradient Variations
·244 words·2 mins· loading · loading
AI Theory Optimization 🏢 Nanjing University
A novel universal online learning algorithm achieves optimal gradient-variation regret across diverse function curvatures, boasting efficiency with only one gradient query per round.
A Simple and Adaptive Learning Rate for FTRL in Online Learning with Minimax Regret of Θ(T^{2/3}) and its Application to Best-of-Both-Worlds
·334 words·2 mins· loading · loading
AI Theory Optimization 🏢 University of Tokyo
A new adaptive learning rate for FTRL achieves minimax regret of O(T²/³) in online learning, improving existing best-of-both-worlds algorithms for various hard problems.