Posters
2024
Dissecting the Interplay of Attention Paths in a Statistical Mechanics Theory of Transformers
·2710 words·13 mins·
loading
·
loading
AI Generated
AI Theory
Interpretability
🏢 Harvard University
Researchers dissected attention paths in Transformers using statistical mechanics, revealing a task-relevant kernel combination mechanism boosting generalization performance.
Dissecting the Failure of Invariant Learning on Graphs
·4452 words·21 mins·
loading
·
loading
AI Generated
AI Theory
Generalization
🏢 Peking University
Cross-environment Intra-class Alignment (CIA) and its label-free variant, CIA-LRA, significantly improve node-level OOD generalization on graphs by aligning representations and eliminating spurious fe…
Dissect Black Box: Interpreting for Rule-Based Explanations in Unsupervised Anomaly Detection
·1770 words·9 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Tsinghua University
SCD-Tree & GBD: Unlocking interpretable rules for unsupervised anomaly detection!
DISP-LLM: Dimension-Independent Structural Pruning for Large Language Models
·3179 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Samsung Research
DISP-LLM: A novel dimension-independent structural pruning method for LLMs achieves accuracy similar to semi-structural pruning while improving flexibility and efficiency, outperforming state-of-the-a…
Disentangling Linear Quadratic Control with Untrusted ML Predictions
·1894 words·9 mins·
loading
·
loading
AI Applications
Robotics
🏢 Chinese University of Hong Kong, Shenzhen
DISC, a novel control policy, disentangles untrusted ML predictions to achieve near-optimal performance when accurate, while guaranteeing competitive ratio bounds even with significant prediction erro…
Disentangling Interpretable Factors with Supervised Independent Subspace Principal Component Analysis
·3550 words·17 mins·
loading
·
loading
AI Generated
Machine Learning
Representation Learning
🏢 Columbia University
Supervised Independent Subspace PCA (sisPCA) disentangles interpretable factors in high-dimensional data by leveraging supervision to maximize subspace dependence on target variables while minimizing …
Disentangling and mitigating the impact of task similarity for continual learning
·2158 words·11 mins·
loading
·
loading
Machine Learning
Transfer Learning
🏢 Washington University in St Louis
This study reveals that high input similarity paired with low output similarity is detrimental to continual learning, whereas the opposite scenario is relatively benign; offering insights into mitigat…
Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning
·1850 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Texas at Austin
DUSDi: A novel method for learning disentangled skills in unsupervised reinforcement learning, enabling efficient reuse for diverse downstream tasks.
Disentangled Style Domain for Implicit $z$-Watermark Towards Copyright Protection
·1999 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Fudan University
This paper introduces a novel implicit Zero-Watermarking scheme using disentangled style domains to detect unauthorized dataset usage in text-to-image models, offering robust copyright protection via …
Disentangled Representation Learning in Non-Markovian Causal Systems
·2882 words·14 mins·
loading
·
loading
AI Theory
Causality
🏢 Columbia University
This paper introduces graphical criteria and an algorithm for disentangling causal factors from heterogeneous data in non-Markovian settings, advancing causal representation learning.
DisenGCD: A Meta Multigraph-assisted Disentangled Graph Learning Framework for Cognitive Diagnosis
·2314 words·11 mins·
loading
·
loading
AI Applications
Education
🏢 Anhui University
DisenGCD, a meta multigraph framework, disentangles graph learning for cognitive diagnosis, achieving robust student knowledge assessment.
Discretely beyond $1/e$: Guided Combinatorial Algortihms for Submodular Maximization
·3091 words·15 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Texas A&M University
Researchers surpass the 1/e barrier in submodular maximization with novel combinatorial algorithms!
Discrete-state Continuous-time Diffusion for Graph Generation
·2084 words·10 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Illinois Urbana-Champaign
DISCO: a novel discrete-state continuous-time diffusion model for flexible and efficient graph generation, outperforming state-of-the-art methods.
Discrete Modeling via Boundary Conditional Diffusion Processes
·2908 words·14 mins·
loading
·
loading
AI Generated
Natural Language Processing
Text Generation
🏢 Harbin Institute of Technology
Bridging the gap between continuous diffusion models and discrete data, this work introduces a novel boundary-conditional approach achieving superior performance in language modeling and image generat…
Discrete Dictionary-based Decomposition Layer for Structured Representation Learning
·4466 words·21 mins·
loading
·
loading
AI Generated
Machine Learning
Representation Learning
🏢 Kyungpook National University
Boosting structured representation learning, a novel Discrete Dictionary-based Decomposition (D3) layer significantly improves systematic generalization in TPR-based models by efficiently decomposing …
Discovery of the Hidden World with Large Language Models
·6303 words·30 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Hong Kong Baptist University
COAT leverages LLMs to identify high-level causal factors from unstructured data, enabling causal discovery in real-world scenarios where well-defined variables are lacking.
Discovering Sparsity Allocation for Layer-wise Pruning of Large Language Models
·1939 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Hong Kong University of Science and Technology
DSA, a novel automated framework, discovers optimal sparsity allocation for layer-wise LLM pruning, achieving significant performance gains across various models and tasks.
Discovering Preference Optimization Algorithms with and for Large Language Models
·4948 words·24 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Sakana AI
LLMs discover novel offline preference optimization algorithms, achieving state-of-the-art performance on various tasks.
Discovering plasticity rules that organize and maintain neural circuits
·1657 words·8 mins·
loading
·
loading
Machine Learning
Meta Learning
🏢 University of Washington
AI discovers robust, biologically-plausible plasticity rules that self-organize and maintain neural circuits’ sequential activity, even with synaptic turnover.
Discovering Creative Behaviors through DUPLEX: Diverse Universal Features for Policy Exploration
·1669 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Texas at Austin
DUPLEX: a novel RL method trains diverse, near-optimal policies in complex, dynamic environments by explicitly maximizing policy diversity using successor features. It outperforms existing methods in…