Posters
2024
Stealth edits to large language models
·3221 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 King's College London
Researchers unveil stealth edits for large language models, offering a new metric to assess editability and reveal vulnerability to malicious attacks.
Statistical-Computational Trade-offs for Density Estimation
·433 words·3 mins·
loading
·
loading
AI Theory
Optimization
🏢 MIT
Density estimation algorithms face inherent trade-offs: reducing sample needs often increases query time. This paper proves these trade-offs are fundamental, showing limits to how much improvement is…
Statistical and Geometrical properties of the Kernel Kullback-Leibler divergence
·1547 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 CREST, ENSAE, IP Paris
Regularized Kernel Kullback-Leibler divergence solves the original KKL’s disjoint support limitation, enabling comparison of any probability distributions with a closed-form solution and efficient gra…
State-free Reinforcement Learning
·357 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Boston University
State-free Reinforcement Learning (SFRL) framework eliminates the need for state-space information in RL algorithms, achieving regret bounds independent of the state space size and adaptive to the rea…
State Space Models on Temporal Graphs: A First-Principles Study
·1794 words·9 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 Sun Yat-Sen University
GRAPHSSM: a novel graph state space model efficiently captures temporal graph dynamics, overcoming limitations of existing sequence models.
State Chrono Representation for Enhancing Generalization in Reinforcement Learning
·2535 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 University of California, Santa Barbara
State Chrono Representation (SCR) enhances reinforcement learning generalization by incorporating extensive temporal information and cumulative rewards into state representations, improving performanc…
START: A Generalized State Space Model with Saliency-Driven Token-Aware Transformation
·2428 words·12 mins·
loading
·
loading
Computer Vision
Domain Generalization
🏢 Nanjing University
START, a novel SSM-based architecture with saliency-driven token-aware transformation, achieves state-of-the-art domain generalization performance with efficient linear complexity.
Star-Agents: Automatic Data Optimization with LLM Agents for Instruction Tuning
·1847 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Huawei Noah's Ark Lab
Star-Agents automates data optimization for instruction-tuned LLMs via multi-agent collaboration, achieving a 12% average performance boost.
Stable-Pose: Leveraging Transformers for Pose-Guided Text-to-Image Generation
·3451 words·17 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Munich Center for Machine Learning
Stable-Pose: Precise human pose guidance for text-to-image synthesis.
Stabilizing Zero-Shot Prediction: A Novel Antidote to Forgetting in Continual Vision-Language Tasks
·2243 words·11 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Tsinghua University
ZAF: a novel replay-free continual learning method for vision-language models, significantly reduces forgetting by stabilizing zero-shot predictions.
Stabilizing Linear Passive-Aggressive Online Learning with Weighted Reservoir Sampling
·7304 words·35 mins·
loading
·
loading
AI Generated
AI Applications
Security
🏢 Harvard University
Weighted reservoir sampling stabilizes online learning algorithms by creating a robust ensemble of intermediate solutions, significantly improving accuracy and mitigating sensitivity to outliers.
Stabilize the Latent Space for Image Autoregressive Modeling: A Unified Perspective
·2011 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 School of Data Science, University of Science and Technology of China
DiGIT stabilizes image autoregressive models’ latent space using a novel discrete tokenizer from self-supervised learning, achieving state-of-the-art image generation.
Stability and Generalization of Asynchronous SGD: Sharper Bounds Beyond Lipschitz and Smoothness
·1414 words·7 mins·
loading
·
loading
AI Theory
Generalization
🏢 National University of Defense Technology
Sharper ASGD generalization bounds achieved by leveraging on-average model stability, even without Lipschitz and smoothness assumptions; validated with diverse machine learning models.
Stability and Generalization of Adversarial Training for Shallow Neural Networks with Smooth Activation
·201 words·1 min·
loading
·
loading
AI Generated
AI Theory
Robustness
🏢 Johns Hopkins University
This paper provides novel theoretical guarantees for adversarial training of shallow neural networks, improving generalization bounds via early stopping and Moreau’s envelope smoothing.
Stability and Generalizability in SDE Diffusion Models with Measure-Preserving Dynamics
·2499 words·12 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Oxford
D³GM, a novel score-based diffusion model, enhances stability & generalizability in solving inverse problems by leveraging measure-preserving dynamics, enabling robust image reconstruction across dive…
ST$_k$: A Scalable Module for Solving Top-k Problems
·2467 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 School of Mathematical Sciences
STk: a novel, differentiable module solves Top-k problems in neural networks without extra time/GPU memory, boosting performance in long-tailed learning.
SSDM: Scalable Speech Dysfluency Modeling
·2807 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Berkeley
SSDM: Scalable Speech Dysfluency Modeling tackles challenges in speech dysfluency analysis by using articulatory gestures for scalable alignment, a connectionist subsequence aligner for efficient dysf…
SSDiff: Spatial-spectral Integrated Diffusion Model for Remote Sensing Pansharpening
·2088 words·10 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 University of Electronic Science and Technology of China
SSDiff: A novel spatial-spectral integrated diffusion model for superior remote sensing pansharpening.
SSA-Seg: Semantic and Spatial Adaptive Pixel-level Classifier for Semantic Segmentation
·2332 words·11 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 Huawei Noah's Ark Lab Zhejiang University
SSA-Seg improves semantic segmentation by adapting pixel-level classifiers to the test image’s semantic and spatial features, achieving state-of-the-art performance with minimal extra computational co…
SS1: Accelerating Inference with Fast and Expressive Sketch Structured Transform
·2142 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Rice University
SS1: A novel GPU-friendly operator accelerates deep learning inference by leveraging structured parameter sharing, achieving superior quality-efficiency tradeoffs compared to existing methods.