Posters
2024
SequentialAttention++ for Block Sparsification: Differentiable Pruning Meets Combinatorial Optimization
·1692 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ Google Research
SequentialAttention++ unites differentiable pruning with combinatorial optimization for efficient and accurate neural network block sparsification, achieving state-of-the-art results.
Sequential Signal Mixing Aggregation for Message Passing Graph Neural Networks
·2468 words·12 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Technion
Sequential Signal Mixing Aggregation (SSMA) boosts message-passing graph neural network performance by effectively mixing neighbor features, achieving state-of-the-art results across various benchmark…
Sequential Probability Assignment with Contexts: Minimax Regret, Contextual Shtarkov Sums, and Contextual Normalized Maximum Likelihood
·217 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ University of Toronto
This paper introduces contextual Shtarkov sums, a new complexity measure characterizing minimax regret in sequential probability assignment with contexts, and derives the minimax optimal algorithm, co…
Sequential Harmful Shift Detection Without Labels
·2657 words·13 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ J.P. Morgan AI Research
This paper introduces a novel, label-free method for detecting harmful distribution shifts in machine learning models deployed in production environments, leveraging a proxy error derived from an erro…
Sequential Decision Making with Expert Demonstrations under Unobserved Heterogeneity
·1771 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ University of Toronto
ExPerior leverages expert demonstrations to enhance online decision-making, even when experts use hidden contextual information unseen by the learner.
Sequence-Augmented SE(3)-Flow Matching For Conditional Protein Generation
·4573 words·22 mins·
loading
·
loading
AI Generated
AI Applications
Healthcare
π’ University of Oxford
Sequence-augmented SE(3)-Flow model, FOLDFLOW-2, excels at generating diverse, designable protein structures, surpassing existing methods in unconditional and conditional design tasks.
Separations in the Representational Capabilities of Transformers and Recurrent Architectures
·1587 words·8 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
π’ University of Oxford
Transformers and RNNs show contrasting representational capabilities: Transformers excel at tasks requiring associative recall, while RNNs are better suited for hierarchical language processing. This …
Separation and Bias of Deep Equilibrium Models on Expressivity and Learning Dynamics
·2192 words·11 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
π’ Peking University
Deep Equilibrium Models (DEQs) outperform standard neural networks, but lack theoretical understanding. This paper provides general separation results showing DEQ’s superior expressivity and character…
Separate and Reconstruct: Asymmetric Encoder-Decoder for Speech Separation
·3807 words·18 mins·
loading
·
loading
AI Generated
Speech and Audio
Speech Recognition
π’ Sogang University
SepReformer: Asymmetric encoder-decoder model for efficient speech separation, achieving state-of-the-art performance with less computation.
Semidefinite Relaxations of the Gromov-Wasserstein Distance
·2209 words·11 mins·
loading
·
loading
AI Theory
Optimization
π’ National University of Singapore
This paper introduces a novel, tractable semidefinite program (SDP) relaxation for the Gromov-Wasserstein distance, enabling the computation of globally optimal transportation plans.
Semi-supervised Knowledge Transfer Across Multi-omic Single-cell Data
·2468 words·12 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
π’ Georgia Institute of Technology
DANCE, a novel semi-supervised framework, efficiently transfers cell types across multi-omic single-cell data even with limited labeled samples, outperforming current state-of-the-art methods.
Semi-Random Matrix Completion via Flow-Based Adaptive Reweighting
·349 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ MIT
New nearly-linear time algorithm achieves high-accuracy semi-random matrix completion, overcoming previous limitations on accuracy and noise tolerance.
Semi-Open 3D Object Retrieval via Hierarchical Equilibrium on Hypergraph
·2346 words·12 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
π’ Tsinghua University
HERT: a novel framework for semi-open 3D object retrieval using hierarchical hypergraph equilibrium, achieving state-of-the-art performance on four new benchmark datasets.
SemFlow: Binding Semantic Segmentation and Image Synthesis via Rectified Flow
·2658 words·13 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
π’ Peking University
SemFlow: A unified framework uses rectified flow to seamlessly bridge semantic segmentation and image synthesis, achieving competitive results and offering reversible image-mask transformations.
SemCoder: Training Code Language Models with Comprehensive Semantics Reasoning
·1972 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Columbia University
SEMCODER: A novel 6.7B parameter code LLM surpasses GPT-3.5-turbo’s performance on code generation and execution reasoning by employing ‘monologue reasoning’βtraining the model to verbally explain cod…
Semantics and Spatiality of Emergent Communication
·2868 words·14 mins·
loading
·
loading
AI Generated
Natural Language Processing
Emergent Communication
π’ Technion - Israel Institute of Technology
Emergent communication protocols are surprisingly inconsistent; this paper proves reconstruction-based objectives yield semantically consistent protocols, unlike discrimination-based ones, highlightin…
Semantic Routing via Autoregressive Modeling
·2894 words·14 mins·
loading
·
loading
Natural Language Processing
AI Applications
π’ Google Research
Learning-based semantic routing, a scalable approach to route planning using rich user queries, is introduced, accompanied by a large-scale public benchmark and a proof-of-concept model demonstrating …
Semantic Feature Learning for Universal Unsupervised Cross-Domain Retrieval
·1809 words·9 mins·
loading
·
loading
Computer Vision
Cross-Modal Retrieval
π’ Northwestern University
Universal Unsupervised Cross-Domain Retrieval (U2CDR) framework learns semantic features to enable accurate retrieval even when category spaces differ across domains.
SELMA: Learning and Merging Skill-Specific Text-to-Image Experts with Auto-Generated Data
·2539 words·12 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
π’ UNC Chapel Hill
SELMA boosts text-to-image fidelity by merging skill-specific models trained on automatically generated image-text datasets.
SelfCodeAlign: Self-Alignment for Code Generation
·1983 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ University of Illinois Urbana-Champaign
SelfCodeAlign is a novel self-alignment method for code generation LLMs that surpasses existing methods by avoiding reliance on expensive human annotation or proprietary LLMs. The method achieves thi…