Posters
2024
SAMPa: Sharpness-aware Minimization Parallelized
·2453 words·12 mins·
loading
·
loading
Machine Learning
Optimization
π’ EPFL
SAMPa: Parallelizing gradient computations in Sharpness-Aware Minimization (SAM) achieves a 2x speedup and superior generalization.
Samba: Severity-aware Recurrent Modeling for Cross-domain Medical Image Grading
·2230 words·11 mins·
loading
·
loading
Computer Vision
Image Classification
π’ Westlake University
Samba: a novel severity-aware recurrent model, tackles cross-domain medical image grading by sequentially encoding image patches and recalibrating states using EM, significantly improving accuracy.
SAM-Guided Masked Token Prediction for 3D Scene Understanding
·1740 words·9 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ Clemson University
This paper introduces SAM-guided masked token prediction, a novel framework for 3D scene understanding that leverages foundation models to significantly improve 3D object detection and semantic segmen…
SafeWorld: Geo-Diverse Safety Alignment
·3977 words·19 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ UC Los Angeles
SAFEWORLD: a new benchmark reveals and fixes LLMs’ struggle with diverse safety standards.
Safety through feedback in Constrained RL
·2526 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Singapore Management University
Reinforcement Learning from Safety Feedback (RLSF) efficiently infers cost functions from trajectory-level feedback, enabling safe policy learning in complex environments.
SAFE: Slow and Fast Parameter-Efο¬cient Tuning for Continual Learning with Pre-Trained Models
·2317 words·11 mins·
loading
·
loading
Machine Learning
Continual Learning
π’ Tencent AI Lab
SAFE, a novel parameter-efficient tuning framework, boosts pre-trained model performance in continual learning by balancing model stability and plasticity through slow and fast learning stages, signif…
Safe Time-Varying Optimization based on Gaussian Processes with Spatio-Temporal Kernel
·1841 words·9 mins·
loading
·
loading
AI Applications
Robotics
π’ ETH Zurich
TVSAFEOPT: Safe time-varying optimization using spatio-temporal kernels ensures safety while tracking time-varying reward and safety functions, providing optimality guarantees in stationary settings.
Safe Exploitative Play with Untrusted Type Beliefs
·1930 words·10 mins·
loading
·
loading
AI Theory
Optimization
π’ School of Data Science, the Chinese University of Hong Kong, Shenzhen
This paper characterizes the fundamental tradeoff between trusting and distrusting learned type beliefs in games, establishing upper and lower bounds for optimal strategies in both normal-form and sto…
Safe and Sparse Newton Method for Entropic-Regularized Optimal Transport
·2040 words·10 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Shanghai University of Finance and Economics
A novel safe & sparse Newton method (SSNS) for entropic-regularized optimal transport boasts strict error control, avoids singularity, needs no hyperparameter tuning, and offers rigorous convergence a…
Safe and Efficient: A Primal-Dual Method for Offline Convex CMDPs under Partial Data Coverage
·1556 words·8 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ ShanghaiTech University
A novel primal-dual method boosts offline safe reinforcement learning efficiency for convex CMDPs by using uncertainty parameters and achieving a sample complexity of O(1/(1-Ξ³)βn).
S2HPruner: Soft-to-Hard Distillation Bridges the Discretization Gap in Pruning
·2415 words·12 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Fudan University
S2HPruner bridges the discretization gap in neural network pruning via a novel soft-to-hard distillation framework, achieving superior performance across various benchmarks without fine-tuning.
S$^{2}$FT: Efficient, Scalable and Generalizable LLM Fine-tuning by Structured Sparsity
·1908 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Carnegie Mellon University
S2FT: Structured Sparse Fine-Tuning achieves state-of-the-art LLM fine-tuning performance, training efficiency, and inference scalability by selecting sparsely and computing densely.
S-STE: Continuous Pruning Function for Efficient 2:4 Sparse Pre-training
·2718 words·13 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ Tsinghua University
S-STE achieves efficient 2:4 sparse pre-training by introducing a novel continuous pruning function, overcoming the limitations of previous methods and leading to improved accuracy and speed.
S-SOS: Stochastic Sum-Of-Squares for Parametric Polynomial Optimization
·1441 words·7 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Chicago
S-SOS: A new algorithm solves complex, parameterized polynomial problems with provable convergence, enabling efficient solutions for high-dimensional applications like sensor network localization.
S-MolSearch: 3D Semi-supervised Contrastive Learning for Bioactive Molecule Search
·1909 words·9 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
π’ Renmin University of China
S-MolSearch: a novel semi-supervised framework using 3D molecular data and contrastive learning achieves state-of-the-art in bioactive molecule search, outperforming existing methods.
Rule Based Rewards for Language Model Safety
·3342 words·16 mins·
loading
·
loading
AI Theory
Safety
π’ OpenAI
Rule-Based Rewards (RBRs) enhance LLM safety by using AI feedback and a few-shot prompt-based approach, achieving higher safety-behavior accuracy with less human annotation than existing methods.
RTify: Aligning Deep Neural Networks with Human Behavioral Decisions
·1884 words·9 mins·
loading
·
loading
Computer Vision
Image Classification
π’ Brown University
RTify: A novel framework aligns deep neural networks’ dynamics with human reaction times for improved visual decision-making models.
RSA: Resolving Scale Ambiguities in Monocular Depth Estimators through Language Descriptions
·2064 words·10 mins·
loading
·
loading
Natural Language Processing
Vision-Language Models
π’ Yale University
RSA: Language unlocks metric depth from single images!
RouterDC: Query-Based Router by Dual Contrastive Learning for Assembling Large Language Models
·3132 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ Hong Kong University of Science and Technology
RouterDC: A query-based router trained via dual contrastive learning assembles multiple LLMs, significantly outperforming individual LLMs and existing routing methods on both in- and out-of-distributi…
Rough Transformers: Lightweight Continuous-Time Sequence Modelling with Path Signatures
·4237 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ University of Oxford
Rough Transformers: A lightweight continuous-time sequence modeling approach using path signatures to significantly reduce computational costs, improving efficiency and accuracy, particularly for long…