Skip to main content

Posters

2024

ShowMaker: Creating High-Fidelity 2D Human Video via Fine-Grained Diffusion Modeling
·2221 words·11 mins· loading · loading
Computer Vision Image Generation 🏒 Tsinghua University
ShowMaker: Generating high-fidelity 2D human conversational videos using fine-grained diffusion modeling and 2D key points.
Should We Really Edit Language Models? On the Evaluation of Edited Language Models
·3638 words·18 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 Hong Kong University of Science and Technology
Language model editing’s limitations exposed: Scaling current methods leads to knowledge loss and compromised safety, urging research into more robust techniques.
SHMT: Self-supervised Hierarchical Makeup Transfer via Latent Diffusion Models
·3049 words·15 mins· loading · loading
AI Generated Computer Vision Image Generation 🏒 DAMO Academy, Alibaba Group
SHMT: Self-supervised Hierarchical Makeup Transfer uses latent diffusion models to realistically and precisely apply diverse makeup styles to faces, even without paired training data, achieving high f…
ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization
·3020 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Google DeepMind
ShiftAddLLM accelerates pretrained LLMs via post-training, multiplication-less reparameterization, achieving significant memory and energy reductions with comparable or better accuracy than existing m…
SHED: Shapley-Based Automated Dataset Refinement for Instruction Fine-Tuning
·2560 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 University of Maryland
SHED, a Shapley value-based framework, efficiently refines instruction-tuning datasets for LLMs, producing high-performing subsets, only 10% of original size, that transfer well across different model…
Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
·2661 words·13 mins· loading · loading
Machine Learning Deep Learning 🏒 UC San Diego
SharpBalance, a novel training approach, effectively improves deep ensemble performance by addressing the sharpness-diversity trade-off, leading to significant improvements in both in-distribution and…
Sharpness-Aware Minimization Activates the Interactive Teaching's Understanding and Optimization
·1829 words·9 mins· loading · loading
AI Theory Optimization 🏒 School of Artificial Intelligence, Jilin University
Sharpness Reduction Interactive Teaching (SRIT) boosts interactive teaching’s performance by integrating SAM’s generalization capabilities, leading to improved model accuracy and generalization.
Sharing Key Semantics in Transformer Makes Efficient Image Restoration
·3184 words·15 mins· loading · loading
Computer Vision Image Restoration 🏒 Peking University
SemanIR boosts image restoration efficiency by cleverly sharing key semantic information within Transformer layers, achieving state-of-the-art results across multiple tasks.
Shared Autonomy with IDA: Interventional Diffusion Assistance
·1793 words·9 mins· loading · loading
AI Applications Robotics 🏒 UC Los Angeles
IDA, a novel intervention assistance, dynamically shares control between human and AI copilots by intervening only when the AI’s action is superior across all goals, maximizing performance and preserv…
Shaping the distribution of neural responses with interneurons in a recurrent circuit model
·1538 words·8 mins· loading · loading
AI Theory Optimization 🏒 Center for Computational Neuroscience, Flatiron Institute
Researchers developed a recurrent neural circuit model that efficiently transforms sensory signals into neural representations by dynamically adjusting interneuron connectivity and activation function…
Shape analysis for time series
·2156 words·11 mins· loading · loading
Machine Learning Representation Learning 🏒 Université Paris-Saclay
TS-LDDMM: Unsupervised time-series analysis handles irregular data, offering interpretable shape-based representations & exceeding existing methods in benchmarks.
Shadowheart SGD: Distributed Asynchronous SGD with Optimal Time Complexity Under Arbitrary Computation and Communication Heterogeneity
·2146 words·11 mins· loading · loading
AI Generated Machine Learning Federated Learning 🏒 KAUST AIRI
Shadowheart SGD achieves optimal time complexity for asynchronous SGD in distributed settings with arbitrary computation and communication heterogeneity.
Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models
·3138 words·15 mins· loading · loading
Multimodal Learning Vision-Language Models 🏒 University of Maryland, College Park
Shadowcast: A new data poisoning attack manipulates vision-language models by injecting visually similar, yet deceptively misleading, image-text pairs, causing them to generate false information.
SGLang: Efficient Execution of Structured Language Model Programs
·1898 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏒 UC Berkeley
SGLang: A new system boosts LLM program execution speed by up to 6.4x, simplifying complex LLM application programming.
SGD vs GD: Rank Deficiency in Linear Networks
·381 words·2 mins· loading · loading
AI Theory Optimization 🏒 EPFL
SGD surprisingly diminishes network rank, unlike GD, due to a repulsive force between eigenvalues, offering insights into deep learning generalization.
SG-Nav: Online 3D Scene Graph Prompting for LLM-based Zero-shot Object Navigation
·2121 words·10 mins· loading · loading
Multimodal Learning Embodied AI 🏒 Tsinghua University
SG-Nav achieves state-of-the-art zero-shot object navigation by leveraging a novel 3D scene graph to provide rich context for LLM-based reasoning.
SfPUEL: Shape from Polarization under Unknown Environment Light
·2725 words·13 mins· loading · loading
Computer Vision 3D Vision 🏒 Peking University
SfPUEL: A novel end-to-end SfP method achieves robust single-shot surface normal estimation under diverse lighting, integrating PS priors and material segmentation.
SF-V: Single Forward Video Generation Model
·1607 words·8 mins· loading · loading
Computer Vision Video Understanding 🏒 Snap Inc.
Researchers developed SF-V, a single-step image-to-video generation model, achieving a 23x speedup compared to existing models without sacrificing quality, paving the way for real-time video synthesis…
SeTAR: Out-of-Distribution Detection with Selective Low-Rank Approximation
·5697 words·27 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏒 Southern University of Science and Technology
SeTAR: Training-free OOD detection via selective low-rank approximation, improving robustness and efficiency.
Set-based Neural Network Encoding Without Weight Tying
·5047 words·24 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏒 University of Oxford
Set-based Neural Network Encoder (SNE) efficiently encodes neural network weights for property prediction, eliminating the need for architecture-specific models and improving generalization across dat…