Skip to main content

Posters

2024

Symmetry-Informed Governing Equation Discovery
·2189 words·11 mins· loading · loading
Machine Learning Deep Learning 🏒 UC San Diego
Leveraging symmetry in automated equation discovery improves accuracy and simplicity of learned governing equations, enhancing robustness against noise and achieving higher success rates across divers…
Symmetric Linear Bandits with Hidden Symmetry
·1466 words·7 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Warwick
Researchers unveil a novel algorithm for high-dimensional symmetric linear bandits, achieving a regret bound of O(d^(2/3)T^(2/3)log(d)), surpassing limitations of existing approaches that assume expli…
SymILO: A Symmetry-Aware Learning Framework for Integer Linear Optimization
·1858 words·9 mins· loading · loading
AI Theory Optimization 🏒 Shenzhen Research Institute of Big Data
SymILO: A novel symmetry-aware learning framework dramatically improves integer linear program (ILP) solutions by addressing data variability caused by ILP symmetry.
Symbolic Regression with a Learned Concept Library
·2112 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Texas at Austin
LASR, a novel symbolic regression method, uses zero-shot LLM queries to discover and evolve abstract concepts, substantially outperforming state-of-the-art approaches and discovering a new LLM scaling…
SWT-Bench: Testing and Validating Real-World Bug-Fixes with Code Agents
·3127 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 ETH Zurich
SWT-Bench, a new benchmark, reveals that LLMs excel at generating tests for real-world bug fixes, surpassing dedicated test generation systems and significantly improving code-fix precision.
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
·3239 words·16 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 Stanford University
SwitchHead: A novel MoE attention mechanism accelerates Transformers by significantly reducing computation and memory, matching baseline performance.
Swift Sampler: Efficient Learning of Sampler by 10 Parameters
·2624 words·13 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Washington
Swift Sampler (SS) automates the learning of efficient data samplers for deep learning, achieving significant performance gains (e.g., 1.5% on ImageNet) with minimal computational cost using only 10 p…
SWE-agent: Agent-Computer Interfaces Enable Automated Software Engineering
·10845 words·51 mins· loading · loading
AI Applications Security 🏒 Princeton University
SWE-agent achieves state-of-the-art performance on software engineering benchmarks by creating a custom agent-computer interface that enhances LM agents’ ability to use computers.
SVFT: Parameter-Efficient Fine-Tuning with Singular Vectors
·2772 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Texas at Austin
SVFT: a novel parameter-efficient fine-tuning method achieves near full fine-tuning accuracy using only 0.006% to 0.25% of parameters, significantly outperforming existing techniques.
Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling
·2206 words·11 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 Tencent Hunyuan
Deep learning’s Adam-style optimizers exhibit a surprising surge phenomenon: optimal learning rates initially increase, then decrease, before converging to a non-zero value as batch size grows.
SureMap: Simultaneous mean estimation for single-task and multi-task disaggregated evaluation
·2443 words·12 mins· loading · loading
AI Theory Fairness 🏒 Princeton University
SureMap, a new method, significantly boosts accuracy in single and multi-task disaggregated evaluations of AI models using limited data by transforming the problem into Gaussian mean estimation and cl…
Supra-Laplacian Encoding for Transformer on Dynamic Graphs
·2790 words·14 mins· loading · loading
Machine Learning Deep Learning 🏒 Conservatoire National Des Arts Et Métiers
SLATE: Supra-Laplacian encoding for spatio-temporal Transformers achieves state-of-the-art dynamic link prediction by innovatively using a multi-layer graph representation and a unique cross-attention…
Suppress Content Shift: Better Diffusion Features via Off-the-Shelf Generation Techniques
·3213 words·16 mins· loading · loading
Computer Vision Image Generation 🏒 Institute of Information Engineering, CAS
Boosting diffusion model features: This paper introduces GATE, a novel method to suppress ‘content shift’ in diffusion features, improving their quality via off-the-shelf generation techniques.
SuperVLAD: Compact and Robust Image Descriptors for Visual Place Recognition
·3456 words·17 mins· loading · loading
AI Generated Computer Vision Visual Place Recognition 🏒 Tsinghua University
SuperVLAD: A new visual place recognition method boasts superior robustness and compactness, outperforming state-of-the-art techniques by significantly reducing parameters and dimensions.
Supervised Kernel Thinning
·1753 words·9 mins· loading · loading
AI Generated Machine Learning Supervised Learning 🏒 Cornell University
Supervised Kernel Thinning accelerates kernel regression by cleverly compressing data, achieving quadratic speedups in training and inference with minimal accuracy loss.
Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
·2903 words·14 mins· loading · loading
Natural Language Processing Text Generation 🏒 University of Washington
Generate multiple text drafts from a single language model pass with Superposed Decoding, significantly boosting efficiency!
SuperDeepFool: a new fast and accurate minimal adversarial attack
·4315 words·21 mins· loading · loading
AI Generated AI Theory Robustness 🏒 EPFL
SuperDeepFool: a fast, accurate algorithm generating minimal adversarial perturbations, significantly improving deep learning model robustness evaluation and adversarial training.
Super Consistency of Neural Network Landscapes and Learning Rate Transfer
·3859 words·19 mins· loading · loading
Machine Learning Deep Learning 🏒 ETH Zurich
Neural network hyperparameter transferability across vastly different model sizes is achieved via a newly discovered property called ‘Super Consistency’ of loss landscapes.
Suitable is the Best: Task-Oriented Knowledge Fusion in Vulnerability Detection
·3138 words·15 mins· loading · loading
AI Generated AI Applications Security 🏒 Institute of Systems Engineering, Academy of Military Sciences, PLA
KF-GVD: a novel knowledge fusion-based method boosts vulnerability detection accuracy by integrating task-specific knowledge into graph neural networks, achieving significant performance gains and dis…
Subwords as Skills: Tokenization for Sparse-Reward Reinforcement Learning
·2209 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 TTI-Chicago
This paper introduces Subwords as Skills (SaS), a fast and efficient skill extraction method for sparse-reward reinforcement learning that uses tokenization. SaS enables 1000x faster skill extraction…