Skip to main content

Posters

2024

SMART: Towards Pre-trained Missing-Aware Model for Patient Health Status Prediction
·2281 words·11 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 Peking University
SMART: a novel self-supervised model tackles missing EHR data, improving patient health status prediction via missing-aware attention and robust pre-training.
SMART: Scalable Multi-agent Real-time Motion Generation via Next-token Prediction
·1973 words·10 mins· loading · loading
AI Applications Autonomous Vehicles 🏢 Tsinghua University
SMART: a scalable, real-time multi-agent driving simulator using next-token prediction, achieves state-of-the-art results and zero-shot generalization.
SmallToLarge (S2L): Scalable Data Selection for Fine-tuning Large Language Models by Summarizing Training Trajectories of Small Models
·2599 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 UC Los Angeles
SMALLTOLARGE (S2L) revolutionizes large language model (LLM) fine-tuning by using a small model to summarize training loss trajectories, enabling efficient data selection for larger models.
Small steps no more: Global convergence of stochastic gradient bandits for arbitrary learning rates
·447 words·3 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Google DeepMind
Stochastic gradient bandit algorithms now guaranteed to globally converge, using ANY constant learning rate!
Sm: enhanced localization in Multiple Instance Learning for medical imaging classification
·2471 words·12 mins· loading · loading
AI Applications Healthcare 🏢 University of Granada
SmMIL enhances medical image classification by introducing a novel smooth operator to improve instance-level localization in multiple instance learning, achieving state-of-the-art results.
SLTrain: a sparse plus low rank approach for parameter and memory efficient pretraining
·4422 words·21 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 RIKEN AIP
SLTrain: Sparsity+low-rank pretraining boosts LLM efficiency by up to 73% memory reduction without performance loss!
SlowFocus: Enhancing Fine-grained Temporal Understanding in Video LLM
·4826 words·23 mins· loading · loading
AI Generated Natural Language Processing Vision-Language Models 🏢 School of Data Science, Fudan University
SlowFocus significantly improves fine-grained temporal understanding in video LLMs by using mixed-frequency sampling and a novel multi-frequency attention mechanism.
SLowcalSGD : Slow Query Points Improve Local-SGD for Stochastic Convex Optimization
·362 words·2 mins· loading · loading
Machine Learning Federated Learning 🏢 Technion
SLowcal-SGD, a new local update method for distributed learning, provably outperforms Minibatch-SGD and Local-SGD in heterogeneous settings by using a slow querying technique, mitigating bias from loc…
Slot-VLM: Object-Event Slots for Video-Language Modeling
·4378 words·21 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Microsoft Research
Slot-VLM generates semantically decomposed video tokens using an Object-Event Slots module, improving video-language model performance.
Slot State Space Models
·2613 words·13 mins· loading · loading
Computer Vision Video Understanding 🏢 Rutgers University
SlotSSMs: a novel framework for modular sequence modeling, achieving significant performance gains by incorporating independent mechanisms and sparse interactions into State Space Models.
SlimSAM: 0.1% Data Makes Segment Anything Slim
·2447 words·12 mins· loading · loading
Computer Vision Image Segmentation 🏢 National University of Singapore
SlimSAM achieves near original SAM performance using 0.1% of its training data by employing a novel alternate slimming framework and disturbed Taylor pruning, significantly advancing data-efficient mo…
SlimGPT: Layer-wise Structured Pruning for Large Language Models
·2966 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Alibaba Group
SlimGPT: Achieve near-optimal LLM structured pruning via Batched Greedy Pruning and Incremental Pruning Ratio, improving efficiency without sacrificing accuracy.
SLIM: Style-Linguistics Mismatch Model for Generalized Audio Deepfake Detection
·1964 words·10 mins· loading · loading
Speech and Audio Speaker Recognition 🏢 Reality Defender Inc.
SLIM: A novel audio deepfake detection model leverages style-linguistics mismatch for superior generalization and explainability.
Slicing Vision Transformer for Flexibile Inference
·2922 words·14 mins· loading · loading
Computer Vision Image Classification 🏢 Snap Inc.
Scala: One-shot training enables flexible ViT inference!
SleeperNets: Universal Backdoor Poisoning Attacks Against Reinforcement Learning Agents
·2849 words·14 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Khoury College of Computer Sciences, Northeastern University
SleeperNets: A universal backdoor attack against RL agents, achieving 100% success rate across diverse environments while preserving benign performance.
SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Models
·2353 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Google Research
Self Logits Evolution Decoding (SLED) boosts LLM factuality by up to 20% without extra data or fine-tuning!
Slack-Free Spiking Neural Network Formulation for Hypergraph Minimum Vertex Cover
·1466 words·7 mins· loading · loading
AI Theory Optimization 🏢 Intel Labs
A novel slack-free spiking neural network efficiently solves the Hypergraph Minimum Vertex Cover problem on neuromorphic hardware, outperforming CPU-based methods in both speed and energy consumption.
SkipPredict: When to Invest in Predictions for Scheduling
·2285 words·11 mins· loading · loading
AI Theory Optimization 🏢 Harvard University
SkipPredict optimizes scheduling by prioritizing cheap predictions and using expensive ones only when necessary, achieving cost-effective performance.
Skill-aware Mutual Information Optimisation for Zero-shot Generalisation in Reinforcement Learning
·5509 words·26 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 University of Edinburgh
Skill-aware Mutual Information optimization enhances RL agent generalization across diverse tasks by distinguishing context embeddings based on skills, leading to improved zero-shot performance and ro…
SkiLD: Unsupervised Skill Discovery Guided by Factor Interactions
·2028 words·10 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Texas at Austin
SkiLD, a novel unsupervised skill discovery method, uses state factorization and a new objective function to learn skills inducing diverse interactions between state factors, outperforming existing me…