Skip to main content

Posters

2024

Exact, Tractable Gauss-Newton Optimization in Deep Reversible Architectures Reveal Poor Generalization
·2136 words·11 mins· loading · loading
AI Theory Optimization 🏢 MediaTek Research
Exact Gauss-Newton optimization in deep reversible networks surprisingly reveals poor generalization, despite faster training, challenging existing deep learning optimization theories.
Exact Gradients for Stochastic Spiking Neural Networks Driven by Rough Signals
·528 words·3 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of Copenhagen
New framework uses rough path theory to enable gradient-based training of SSNNs driven by rough signals, allowing for noise in spike timing and network dynamics.
Ex Uno Pluria: Insights on Ensembling in Low Precision Number Systems
·2322 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 Kim Jaechul Graduate School of AI, KAIST
Low Precision Ensembling (LPE) boosts large model accuracy using training-free ensemble creation via stochastic rounding in low-precision number systems.
EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models
·2805 words·14 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Show Lab, National University of Singapore
EvolveDirector trains competitive text-to-image models using publicly available data by cleverly leveraging large vision-language models to curate and refine training datasets, dramatically reducing d…
Evidential Stochastic Differential Equations for Time-Aware Sequential Recommendation
·2228 words·11 mins· loading · loading
AI Applications Recommendation Systems 🏢 Rochester Institute of Technology
E-NSDE, a novel time-aware sequential recommendation model, integrates neural stochastic differential equations and evidential learning to improve recommendation accuracy by effectively handling varia…
Evidential Mixture Machines: Deciphering Multi-Label Correlations for Active Learning Sensitivity
·2512 words·12 mins· loading · loading
Machine Learning Active Learning 🏢 Rochester Institute of Technology
Evidential Mixture Machines (EMM) enhances multi-label active learning by deciphering label correlations for improved accuracy and uncertainty quantification in large, sparse label spaces.
Evidence of Learned Look-Ahead in a Chess-Playing Neural Network
·4142 words·20 mins· loading · loading
AI Generated AI Theory Interpretability 🏢 UC Berkeley
Chess AI Leela Zero surprisingly uses learned look-ahead, internally representing future optimal moves, significantly improving its strategic decision-making.
Everyday Object Meets Vision-and-Language Navigation Agent via Backdoor
·2050 words·10 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Tsinghua University
Researchers introduce object-aware backdoors in Vision-and-Language Navigation, enabling malicious behavior upon encountering specific objects, demonstrating the vulnerability of real-world AI agents.
Event-3DGS: Event-based 3D Reconstruction Using 3D Gaussian Splatting
·2242 words·11 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Tsinghua University
Event-3DGS: First event-based 3D reconstruction using 3D Gaussian splatting, enabling high-quality, efficient, and robust 3D scene reconstruction in challenging real-world conditions.
Even Sparser Graph Transformers
·2059 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 University of British Columbia
Spexphormer achieves significant memory reduction in graph Transformers by leveraging a two-stage training process that leverages attention score consistency across network widths to effectively spars…
Evaluation of Text-to-Video Generation Models: A Dynamics Perspective
·3278 words·16 mins· loading · loading
Natural Language Processing Vision-Language Models 🏢 University of Chinese Academy of Sciences
DEVIL: a novel text-to-video evaluation protocol focusing on video dynamics, resulting in more realistic video generation.
Evaluating the design space of diffusion-based generative models
·378 words·2 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Berkeley
This paper provides the first complete error analysis for diffusion models, theoretically justifying optimal training and sampling strategies and design choices for enhanced generative capabilities.
Evaluating alignment between humans and neural network representations in image-based learning tasks
·3856 words·19 mins· loading · loading
AI Generated AI Theory Representation Learning 🏢 Helmholtz Computational Health Center
Pretrained neural networks surprisingly capture fundamental aspects of human cognition, enabling generalization in image-based learning tasks, as demonstrated by aligning neural network representation…
Evaluate then Cooperate: Shapley-based View Cooperation Enhancement for Multi-view Clustering
·1847 words·9 mins· loading · loading
Machine Learning Unsupervised Learning 🏢 National University of Defence Technology
Shapley-based Cooperation Enhancing Multi-view Clustering (SCE-MVC) improves deep multi-view clustering by using game theory to fairly evaluate and enhance individual view contributions.
Euclidean distance compression via deep random features
·1451 words·7 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 UC Davis
Deep random features enable efficient Euclidean distance compression, offering improved bit storage compared to linear methods for specific parameter ranges, thus significantly advancing high-dimensio…
ET-Flow: Equivariant Flow-Matching for Molecular Conformer Generation
·2269 words·11 mins· loading · loading
AI Applications Healthcare 🏢 University of British Columbia
ET-Flow, a novel equivariant flow-matching model, generates highly accurate and physically realistic molecular conformers significantly faster than existing methods.
Estimating the Hallucination Rate of Generative AI
·3412 words·17 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Department of Statistics, Columbia University
New method estimates hallucination rates in generative AI’s in-context learning, improving model reliability.
Estimating Heterogeneous Treatment Effects by Combining Weak Instruments and Observational Data
·1848 words·9 mins· loading · loading
AI Theory Causality 🏢 Cornell University
This study develops a novel two-stage framework for accurately predicting conditional average treatment effects using both observational data and weak instrumental variables, overcoming limitations of…
Estimating Generalization Performance Along the Trajectory of Proximal SGD in Robust Regression
·1899 words·9 mins· loading · loading
AI Theory Optimization 🏢 Rutgers University
New consistent estimators precisely track generalization error during robust regression’s iterative model training, enabling optimal stopping iteration for minimized error.
Estimating Epistemic and Aleatoric Uncertainty with a Single Model
·2171 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 University of Maryland
HyperDM accurately estimates both epistemic and aleatoric uncertainty using a single model, overcoming the computational limitations of existing ensemble methods.