🏢 ETH Zurich
Policy Mirror Descent with Lookahead
·1918 words·10 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 ETH Zurich
Boosting reinforcement learning, this paper introduces h-PMD, a novel algorithm enhancing policy mirror descent with lookahead for faster convergence and improved sample complexity.
On Affine Homotopy between Language Encoders
·2070 words·10 mins·
loading
·
loading
AI Generated
Natural Language Processing
Representation Learning
🏢 ETH Zurich
This paper introduces a novel notion of intrinsic similarity between language encoders, based on affine homotopy, and demonstrates its strong correlation with extrinsic similarity (downstream task per…
NeoRL: Efficient Exploration for Nonepisodic RL
·1407 words·7 mins·
loading
·
loading
Reinforcement Learning
🏢 ETH Zurich
NEORL: Novel nonepisodic RL algorithm guarantees optimal average cost with sublinear regret for nonlinear systems!
MultiOOD: Scaling Out-of-Distribution Detection for Multiple Modalities
·3642 words·18 mins·
loading
·
loading
Multimodal Learning
Multimodal Understanding
🏢 ETH Zurich
MultiOOD benchmark and novel A2D & NP-Mix algorithms drastically improve multimodal out-of-distribution detection.
LiteVAE: Lightweight and Efficient Variational Autoencoders for Latent Diffusion Models
·3223 words·16 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 ETH Zurich
LiteVAE: A new autoencoder design for latent diffusion models boosts efficiency sixfold without sacrificing image quality, achieving faster training and lower memory needs via the 2D discrete wavelet …
Learning diffusion at lightspeed
·1990 words·10 mins·
loading
·
loading
AI Theory
Optimization
🏢 ETH Zurich
JKOnet* learns diffusion processes at unprecedented speed and accuracy by directly minimizing a simple quadratic loss function, bypassing complex bilevel optimization problems.
Learning Bregman Divergences with Application to Robustness
·2210 words·11 mins·
loading
·
loading
Computer Vision
Image Classification
🏢 ETH Zurich
Learned Bregman divergences significantly improve image corruption robustness in adversarial training.
Implicit Regularization of Sharpness-Aware Minimization for Scale-Invariant Problems
·2239 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 ETH Zurich
Boosting deep learning generalization, this work unveils SAM’s implicit regularization using ‘balancedness’, a new metric. A resource-efficient variant, BAR, achieves 95% computational savings with i…
FUSE: Fast Unified Simulation and Estimation for PDEs
·7308 words·35 mins·
loading
·
loading
AI Generated
AI Applications
Healthcare
🏢 ETH Zurich
FUSE, a novel framework, efficiently predicts continuous fields & estimates discrete parameters in PDEs, significantly improving accuracy and robustness.
Fairness in Social Influence Maximization via Optimal Transport
·2682 words·13 mins·
loading
·
loading
AI Theory
Fairness
🏢 ETH Zurich
Fairness in social influence maximization is achieved via optimal transport, optimizing both outreach and a new ‘mutual fairness’ metric that considers variability in outreach scenarios.
Exploiting LLM Quantization
·1836 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 ETH Zurich
LLM quantization, while improving efficiency, creates a security risk: attackers can craft seemingly benign models that exhibit malicious behavior only when quantized.
Dynamic 3D Gaussian Fields for Urban Areas
·2544 words·12 mins·
loading
·
loading
3D Vision
🏢 ETH Zurich
4DGF, a novel neural scene representation, achieves interactive-speed novel view synthesis for large-scale dynamic urban areas by efficiently combining 3D Gaussians and neural fields.
DeltaDEQ: Exploiting Heterogeneous Convergence for Accelerating Deep Equilibrium Iterations
·2888 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
Video Understanding
🏢 ETH Zurich
DeltaDEQ accelerates deep equilibrium model inference by 73-84% via a novel ‘heterogeneous convergence’ exploitation technique, maintaining accuracy.
Cooperate or Collapse: Emergence of Sustainable Cooperation in a Society of LLM Agents
·6245 words·30 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 ETH Zurich
LLMs struggle to cooperate sustainably; GOVSIM reveals this, showing communication and ‘universalization’ reasoning improve outcomes.
Contextual Bilevel Reinforcement Learning for Incentive Alignment
·3140 words·15 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 ETH Zurich
Contextual Bilevel Reinforcement Learning (CB-RL) tackles real-world strategic decision-making where optimal policies depend on environmental configurations and exogenous events, proposing a stochasti…
ConStat: Performance-Based Contamination Detection in Large Language Models
·4433 words·21 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 ETH Zurich
ConStat: Exposing hidden LLM contamination!
Confidence Regulation Neurons in Language Models
·3393 words·16 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 ETH Zurich
LLMs regulate uncertainty via specialized ’entropy’ and ’token frequency’ neurons, impacting prediction confidence without directly altering logits.
Compositional PAC-Bayes: Generalization of GNNs with persistence and beyond
·2208 words·11 mins·
loading
·
loading
AI Theory
Generalization
🏢 ETH Zurich
Novel compositional PAC-Bayes framework delivers data-dependent generalization bounds for persistence-enhanced Graph Neural Networks, improving model design and performance.
Causal Effect Identification in a Sub-Population with Latent Variables
·1896 words·9 mins·
loading
·
loading
AI Theory
Causality
🏢 ETH Zurich
This paper introduces a novel algorithm to accurately compute causal effects within specific sub-populations, even when hidden factors influence the data, advancing causal inference significantly.
Can an AI Agent Safely Run a Government? Existence of Probably Approximately Aligned Policies
·506 words·3 mins·
loading
·
loading
AI Theory
Safety
🏢 ETH Zurich
This paper introduces a novel quantitative definition of AI alignment for social decision-making, proposing probably approximately aligned policies and a method to safeguard any autonomous agent’s act…