Skip to main content

Posters

2024

Dissecting the Interplay of Attention Paths in a Statistical Mechanics Theory of Transformers
·2710 words·13 mins· loading · loading
AI Generated AI Theory Interpretability 🏢 Harvard University
Researchers dissected attention paths in Transformers using statistical mechanics, revealing a task-relevant kernel combination mechanism boosting generalization performance.
Dissecting the Failure of Invariant Learning on Graphs
·4452 words·21 mins· loading · loading
AI Generated AI Theory Generalization 🏢 Peking University
Cross-environment Intra-class Alignment (CIA) and its label-free variant, CIA-LRA, significantly improve node-level OOD generalization on graphs by aligning representations and eliminating spurious fe…
Dissect Black Box: Interpreting for Rule-Based Explanations in Unsupervised Anomaly Detection
·1770 words·9 mins· loading · loading
Machine Learning Unsupervised Learning 🏢 Tsinghua University
SCD-Tree & GBD: Unlocking interpretable rules for unsupervised anomaly detection!
DISP-LLM: Dimension-Independent Structural Pruning for Large Language Models
·3179 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Samsung Research
DISP-LLM: A novel dimension-independent structural pruning method for LLMs achieves accuracy similar to semi-structural pruning while improving flexibility and efficiency, outperforming state-of-the-a…
Disentangling Linear Quadratic Control with Untrusted ML Predictions
·1894 words·9 mins· loading · loading
AI Applications Robotics 🏢 Chinese University of Hong Kong, Shenzhen
DISC, a novel control policy, disentangles untrusted ML predictions to achieve near-optimal performance when accurate, while guaranteeing competitive ratio bounds even with significant prediction erro…
Disentangling Interpretable Factors with Supervised Independent Subspace Principal Component Analysis
·3550 words·17 mins· loading · loading
AI Generated Machine Learning Representation Learning 🏢 Columbia University
Supervised Independent Subspace PCA (sisPCA) disentangles interpretable factors in high-dimensional data by leveraging supervision to maximize subspace dependence on target variables while minimizing …
Disentangling and mitigating the impact of task similarity for continual learning
·2158 words·11 mins· loading · loading
Machine Learning Transfer Learning 🏢 Washington University in St Louis
This study reveals that high input similarity paired with low output similarity is detrimental to continual learning, whereas the opposite scenario is relatively benign; offering insights into mitigat…
Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning
·1850 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Texas at Austin
DUSDi: A novel method for learning disentangled skills in unsupervised reinforcement learning, enabling efficient reuse for diverse downstream tasks.
Disentangled Style Domain for Implicit $z$-Watermark Towards Copyright Protection
·1999 words·10 mins· loading · loading
Computer Vision Image Generation 🏢 Fudan University
This paper introduces a novel implicit Zero-Watermarking scheme using disentangled style domains to detect unauthorized dataset usage in text-to-image models, offering robust copyright protection via …
Disentangled Representation Learning in Non-Markovian Causal Systems
·2882 words·14 mins· loading · loading
AI Theory Causality 🏢 Columbia University
This paper introduces graphical criteria and an algorithm for disentangling causal factors from heterogeneous data in non-Markovian settings, advancing causal representation learning.
DisenGCD: A Meta Multigraph-assisted Disentangled Graph Learning Framework for Cognitive Diagnosis
·2314 words·11 mins· loading · loading
AI Applications Education 🏢 Anhui University
DisenGCD, a meta multigraph framework, disentangles graph learning for cognitive diagnosis, achieving robust student knowledge assessment.
Discretely beyond $1/e$: Guided Combinatorial Algortihms for Submodular Maximization
·3091 words·15 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Texas A&M University
Researchers surpass the 1/e barrier in submodular maximization with novel combinatorial algorithms!
Discrete-state Continuous-time Diffusion for Graph Generation
·2084 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Illinois Urbana-Champaign
DISCO: a novel discrete-state continuous-time diffusion model for flexible and efficient graph generation, outperforming state-of-the-art methods.
Discrete Modeling via Boundary Conditional Diffusion Processes
·2908 words·14 mins· loading · loading
AI Generated Natural Language Processing Text Generation 🏢 Harbin Institute of Technology
Bridging the gap between continuous diffusion models and discrete data, this work introduces a novel boundary-conditional approach achieving superior performance in language modeling and image generat…
Discrete Dictionary-based Decomposition Layer for Structured Representation Learning
·4466 words·21 mins· loading · loading
AI Generated Machine Learning Representation Learning 🏢 Kyungpook National University
Boosting structured representation learning, a novel Discrete Dictionary-based Decomposition (D3) layer significantly improves systematic generalization in TPR-based models by efficiently decomposing …
Discovery of the Hidden World with Large Language Models
·6303 words·30 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Hong Kong Baptist University
COAT leverages LLMs to identify high-level causal factors from unstructured data, enabling causal discovery in real-world scenarios where well-defined variables are lacking.
Discovering Sparsity Allocation for Layer-wise Pruning of Large Language Models
·1939 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Hong Kong University of Science and Technology
DSA, a novel automated framework, discovers optimal sparsity allocation for layer-wise LLM pruning, achieving significant performance gains across various models and tasks.
Discovering Preference Optimization Algorithms with and for Large Language Models
·4948 words·24 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Sakana AI
LLMs discover novel offline preference optimization algorithms, achieving state-of-the-art performance on various tasks.
Discovering plasticity rules that organize and maintain neural circuits
·1657 words·8 mins· loading · loading
Machine Learning Meta Learning 🏢 University of Washington
AI discovers robust, biologically-plausible plasticity rules that self-organize and maintain neural circuits’ sequential activity, even with synaptic turnover.
Discovering Creative Behaviors through DUPLEX: Diverse Universal Features for Policy Exploration
·1669 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Texas at Austin
DUPLEX: a novel RL method trains diverse, near-optimal policies in complex, dynamic environments by explicitly maximizing policy diversity using successor features. It outperforms existing methods in…