Posters
2024
Sketchy Moment Matching: Toward Fast and Provable Data Selection for Finetuning
·1518 words·8 mins·
loading
·
loading
AI Theory
Generalization
🏢 Courant Institute
Sketchy Moment Matching (SkMM) is a fast and theoretically sound data selection method for deep learning finetuning. By controlling variance-bias tradeoffs in high dimensions, SkMM drastically reduces…
Sketching for Distributed Deep Learning: A Sharper Analysis
·3663 words·18 mins·
loading
·
loading
AI Generated
Machine Learning
Federated Learning
🏢 University of Illinois Urbana-Champaign
This work presents a sharper analysis of sketching for distributed deep learning, eliminating the problematic dependence on ambient dimension in convergence analysis and proving ambient dimension-inde…
Sketched Lanczos uncertainty score: a low-memory summary of the Fisher information
·2226 words·11 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Technical University of Denmark
SLU: a novel, low-memory uncertainty score for neural networks, achieves logarithmic memory scaling with model parameters, providing well-calibrated uncertainties and outperforming existing methods.
SIRIUS : Contexual Sparisty with Correction for Efficient LLMs
·5392 words·26 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Carnegie Mellon University
SIRIUS: A novel correction mechanism boosts the efficiency of contextually sparse LLMs for complex reasoning tasks, achieving significant latency reduction.
Single-Loop Stochastic Algorithms for Difference of Max-Structured Weakly Convex Functions
·1750 words·9 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
🏢 Texas A&M University
SMAG, a novel single-loop stochastic algorithm, achieves state-of-the-art convergence for solving non-smooth non-convex optimization problems involving differences of max-structured weakly convex func…
Single Image Unlearning: Efficient Machine Unlearning in Multimodal Large Language Models
·3521 words·17 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 School of Cyber Science and Engineering, Southeast University, Nanjing, China
Single Image Unlearning (SIU) efficiently removes visual data from Multimodal Large Language Models (MLLMs) using only one image per concept, outperforming existing methods and defending against attac…
Single Image Reflection Separation via Dual-Stream Interactive Transformers
·2158 words·11 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 College of Intelligence and Computing, Tianjin University
Dual-Stream Interactive Transformers (DSIT) revolutionizes single image reflection separation by using a novel dual-attention mechanism that captures inter- and intra-layer correlations, significantly…
SimVG: A Simple Framework for Visual Grounding with Decoupled Multi-modal Fusion
·2883 words·14 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Southeast University
SimVG: A simpler, faster visual grounding framework with decoupled multi-modal fusion, achieving state-of-the-art performance.
Simulation-Free Training of Neural ODEs on Paired Data
·3545 words·17 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 KAIST
Train Neural ODEs without simulations, achieving high performance on regression and classification by using flow matching in the embedding space of data pairs.
SimPO: Simple Preference Optimization with a Reference-Free Reward
·3091 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Princeton University
SimPO: a simpler, reference-free reward algorithm significantly outperforming existing offline preference optimization methods, achieving higher accuracy and efficiency in aligning LLMs with human pre…
Simplifying Latent Dynamics with Softly State-Invariant World Models
·2423 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Max Planck Institute for Biological Cybernetics
This paper introduces the Parsimonious Latent Space Model (PLSM), a novel world model that regularizes latent dynamics to improve action predictability, enhancing RL performance.
Simplifying Constraint Inference with Inverse Reinforcement Learning
·1653 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of Toronto
This paper simplifies constraint inference in reinforcement learning, demonstrating that standard inverse RL methods can effectively infer constraints from expert data, surpassing complex, previously …
Simplified and Generalized Masked Diffusion for Discrete Data
·2082 words·10 mins·
loading
·
loading
Natural Language Processing
Text Generation
🏢 Google DeepMind
Simplified and generalized masked diffusion models achieve state-of-the-art results in discrete data generation, surpassing previous methods in text and image modeling.
Simple and Fast Distillation of Diffusion Models
·3151 words·15 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Zhejiang University
Simple and Fast Distillation (SFD) drastically accelerates diffusion model training by 1000x, achieving state-of-the-art results in few-step image generation with minimal fine-tuning.
Simple and Effective Masked Diffusion Language Models
·2145 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Cornell Tech
Simple masked discrete diffusion models achieve state-of-the-art language modeling results, closing the performance gap with autoregressive methods by using a novel training recipe and a Rao-Blackwell…
Similarity-Navigated Conformal Prediction for Graph Neural Networks
·2658 words·13 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
🏢 State Key Laboratory of Novel Software Technology, Nanjing University
SNAPS: a novel algorithm boosts graph neural network accuracy by efficiently aggregating non-conformity scores, improving prediction sets without sacrificing validity.
SimGen: Simulator-conditioned Driving Scene Generation
·2751 words·13 mins·
loading
·
loading
AI Applications
Autonomous Vehicles
🏢 UC Los Angeles
SimGen: Simulator-conditioned driving scene generation, uses a novel cascade diffusion pipeline to generate diverse driving scenes by mixing real-world and simulator data, addressing Sim2Real gaps.
SILENCE: Protecting privacy in offloaded speech understanding on resource-constrained devices
·2275 words·11 mins·
loading
·
loading
Natural Language Processing
Speech Recognition
🏢 Peking University
SILENCE, a novel lightweight system, protects user privacy in offloaded speech understanding on resource-constrained devices by selectively masking short-term audio details without impacting long-term…
Sigmoid Gating is More Sample Efficient than Softmax Gating in Mixture of Experts
·1350 words·7 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Texas at Austin
Sigmoid gating significantly boosts sample efficiency in Mixture of Experts models compared to softmax gating, offering faster convergence rates for various expert functions.
Shuffling Gradient-Based Methods for Nonconvex-Concave Minimax Optimization
·337 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 IBM Research
New shuffling gradient methods achieve state-of-the-art oracle complexity for nonconvex-concave minimax optimization problems, offering improved performance and efficiency.