Skip to main content

Posters

2024

SequentialAttention++ for Block Sparsification: Differentiable Pruning Meets Combinatorial Optimization
·1692 words·8 mins· loading · loading
Machine Learning Deep Learning 🏒 Google Research
SequentialAttention++ unites differentiable pruning with combinatorial optimization for efficient and accurate neural network block sparsification, achieving state-of-the-art results.
Sequential Signal Mixing Aggregation for Message Passing Graph Neural Networks
·2468 words·12 mins· loading · loading
AI Theory Representation Learning 🏒 Technion
Sequential Signal Mixing Aggregation (SSMA) boosts message-passing graph neural network performance by effectively mixing neighbor features, achieving state-of-the-art results across various benchmark…
Sequential Probability Assignment with Contexts: Minimax Regret, Contextual Shtarkov Sums, and Contextual Normalized Maximum Likelihood
·217 words·2 mins· loading · loading
AI Theory Optimization 🏒 University of Toronto
This paper introduces contextual Shtarkov sums, a new complexity measure characterizing minimax regret in sequential probability assignment with contexts, and derives the minimax optimal algorithm, co…
Sequential Harmful Shift Detection Without Labels
·2657 words·13 mins· loading · loading
Machine Learning Deep Learning 🏒 J.P. Morgan AI Research
This paper introduces a novel, label-free method for detecting harmful distribution shifts in machine learning models deployed in production environments, leveraging a proxy error derived from an erro…
Sequential Decision Making with Expert Demonstrations under Unobserved Heterogeneity
·1771 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 University of Toronto
ExPerior leverages expert demonstrations to enhance online decision-making, even when experts use hidden contextual information unseen by the learner.
Sequence-Augmented SE(3)-Flow Matching For Conditional Protein Generation
·4573 words·22 mins· loading · loading
AI Generated AI Applications Healthcare 🏒 University of Oxford
Sequence-augmented SE(3)-Flow model, FOLDFLOW-2, excels at generating diverse, designable protein structures, surpassing existing methods in unconditional and conditional design tasks.
Separations in the Representational Capabilities of Transformers and Recurrent Architectures
·1587 words·8 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 University of Oxford
Transformers and RNNs show contrasting representational capabilities: Transformers excel at tasks requiring associative recall, while RNNs are better suited for hierarchical language processing. This …
Separation and Bias of Deep Equilibrium Models on Expressivity and Learning Dynamics
·2192 words·11 mins· loading · loading
AI Generated AI Theory Optimization 🏒 Peking University
Deep Equilibrium Models (DEQs) outperform standard neural networks, but lack theoretical understanding. This paper provides general separation results showing DEQ’s superior expressivity and character…
Separate and Reconstruct: Asymmetric Encoder-Decoder for Speech Separation
·3807 words·18 mins· loading · loading
AI Generated Speech and Audio Speech Recognition 🏒 Sogang University
SepReformer: Asymmetric encoder-decoder model for efficient speech separation, achieving state-of-the-art performance with less computation.
Semidefinite Relaxations of the Gromov-Wasserstein Distance
·2209 words·11 mins· loading · loading
AI Theory Optimization 🏒 National University of Singapore
This paper introduces a novel, tractable semidefinite program (SDP) relaxation for the Gromov-Wasserstein distance, enabling the computation of globally optimal transportation plans.
Semi-supervised Knowledge Transfer Across Multi-omic Single-cell Data
·2468 words·12 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏒 Georgia Institute of Technology
DANCE, a novel semi-supervised framework, efficiently transfers cell types across multi-omic single-cell data even with limited labeled samples, outperforming current state-of-the-art methods.
Semi-Random Matrix Completion via Flow-Based Adaptive Reweighting
·349 words·2 mins· loading · loading
AI Theory Optimization 🏒 MIT
New nearly-linear time algorithm achieves high-accuracy semi-random matrix completion, overcoming previous limitations on accuracy and noise tolerance.
Semi-Open 3D Object Retrieval via Hierarchical Equilibrium on Hypergraph
·2346 words·12 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏒 Tsinghua University
HERT: a novel framework for semi-open 3D object retrieval using hierarchical hypergraph equilibrium, achieving state-of-the-art performance on four new benchmark datasets.
SemFlow: Binding Semantic Segmentation and Image Synthesis via Rectified Flow
·2658 words·13 mins· loading · loading
AI Generated Computer Vision Image Generation 🏒 Peking University
SemFlow: A unified framework uses rectified flow to seamlessly bridge semantic segmentation and image synthesis, achieving competitive results and offering reversible image-mask transformations.
SemCoder: Training Code Language Models with Comprehensive Semantics Reasoning
·1972 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Columbia University
SEMCODER: A novel 6.7B parameter code LLM surpasses GPT-3.5-turbo’s performance on code generation and execution reasoning by employing ‘monologue reasoning’β€”training the model to verbally explain cod…
Semantics and Spatiality of Emergent Communication
·2868 words·14 mins· loading · loading
AI Generated Natural Language Processing Emergent Communication 🏒 Technion - Israel Institute of Technology
Emergent communication protocols are surprisingly inconsistent; this paper proves reconstruction-based objectives yield semantically consistent protocols, unlike discrimination-based ones, highlightin…
Semantic Routing via Autoregressive Modeling
·2894 words·14 mins· loading · loading
Natural Language Processing AI Applications 🏒 Google Research
Learning-based semantic routing, a scalable approach to route planning using rich user queries, is introduced, accompanied by a large-scale public benchmark and a proof-of-concept model demonstrating …
Semantic Feature Learning for Universal Unsupervised Cross-Domain Retrieval
·1809 words·9 mins· loading · loading
Computer Vision Cross-Modal Retrieval 🏒 Northwestern University
Universal Unsupervised Cross-Domain Retrieval (U2CDR) framework learns semantic features to enable accurate retrieval even when category spaces differ across domains.
SELMA: Learning and Merging Skill-Specific Text-to-Image Experts with Auto-Generated Data
·2539 words·12 mins· loading · loading
Multimodal Learning Vision-Language Models 🏒 UNC Chapel Hill
SELMA boosts text-to-image fidelity by merging skill-specific models trained on automatically generated image-text datasets.
SelfCodeAlign: Self-Alignment for Code Generation
·1983 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏒 University of Illinois Urbana-Champaign
SelfCodeAlign is a novel self-alignment method for code generation LLMs that surpasses existing methods by avoiding reliance on expensive human annotation or proprietary LLMs. The method achieves thi…