Skip to main content

Posters

2024

Schedule Your Edit: A Simple yet Effective Diffusion Noise Schedule for Image Editing
·3730 words·18 mins· loading · loading
Computer Vision Image Generation 🏢 State Grid Corporation of China
Logistic Schedule: A novel noise schedule revolutionizes image editing by improving DDIM inversion, enhancing content preservation and edit fidelity without model retraining!
SceneDiffuser: Efficient and Controllable Driving Simulation Initialization and Rollout
·3039 words·15 mins· loading · loading
AI Applications Autonomous Vehicles 🏢 Waymo LLC
SceneDiffuser: a scene-level diffusion model revolutionizes driving simulation by uniting scene initialization and rollout, enabling efficient and controllable closed-loop traffic generation.
SceneCraft: Layout-Guided 3D Scene Generation
·2040 words·10 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Shanghai Jiao Tong University
SceneCraft generates highly detailed indoor scenes from user-provided textual descriptions and spatial layouts, overcoming limitations of previous text-to-3D methods in scale and control.
Scene Graph Generation with Role-Playing Large Language Models
·2597 words·13 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Zhejiang University
SDSGG outperforms leading scene graph generation methods by using LLMs to create scene-specific descriptions, adapting to diverse visual relations.
SCaR: Refining Skill Chaining for Long-Horizon Robotic Manipulation via Dual Regularization
·3516 words·17 mins· loading · loading
AI Applications Robotics 🏢 Nanjing University
SCaR refines skill chaining for long-horizon robotic manipulation via dual regularization, achieving higher success rates and robustness.
Scanning Trojaned Models Using Out-of-Distribution Samples
·2406 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 Sharif University of Technology
TRODO: a novel trojan detection method using out-of-distribution samples, effectively identifies trojaned classifiers even against adversarial attacks and with limited data.
Scaling White-Box Transformers for Vision
·2209 words·11 mins· loading · loading
Computer Vision Image Classification 🏢 UC Santa Cruz
CRATE-a: A new white-box vision transformer architecture achieves 85.1% ImageNet accuracy by strategically scaling model size and datasets, outperforming prior white-box models and preserving interpre…
Scaling transformer neural networks for skillful and reliable medium-range weather forecasting
·2398 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 UC Los Angeles
Stormer, a simple transformer model, achieves state-of-the-art medium-range weather forecasting accuracy by using weather-specific embedding, randomized dynamics forecasting, and a pressure-weighted l…
Scaling the Codebook Size of VQ-GAN to 100,000 with a Utilization Rate of 99%
·2947 words·14 mins· loading · loading
Computer Vision Image Generation 🏢 Microsoft Research
VQGAN-LC massively scales VQGAN’s codebook to 100,000 entries while maintaining a 99% utilization rate, significantly boosting image generation and downstream task performance.
Scaling Sign Language Translation
·4741 words·23 mins· loading · loading
AI Generated Natural Language Processing Machine Translation 🏢 Google DeepMind
Researchers dramatically improved sign language translation by scaling up data, model size, and the number of languages, achieving state-of-the-art results.
Scaling Retrieval-Based Language Models with a Trillion-Token Datastore
·4019 words·19 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Washington
Massive language models improve with bigger datastores at inference time. A 1.4 trillion-token datastore, MASSIVEDS, shows that retrieval-based LMs outperform larger, solely-trained models on knowled…
Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies
·2496 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Stanford University
Boosting LLM performance: This research shows how larger language models need bigger vocabularies for optimal efficiency and performance.
Scaling Laws in Linear Regression: Compute, Parameters, and Data
·1352 words·7 mins· loading · loading
AI Theory Optimization 🏢 UC Berkeley
Deep learning’s neural scaling laws defy conventional wisdom; this paper uses infinite-dimensional linear regression to theoretically explain this phenomenon, showing that implicit regularization of S…
Scaling Laws for Reward Model Overoptimization in Direct Alignment Algorithms
·2949 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Stanford University
Direct Alignment Algorithms (DAAs) for LLM alignment suffer from over-optimization, even without explicit reward models; this paper empirically demonstrates this and proposes scaling laws to understan…
Scaling laws for learning with real and surrogate data
·4942 words·24 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Granica Computing Inc.
Boost machine learning with surrogate data! A novel weighted ERM method effectively integrates surrogate data, significantly reducing test errors even with unrelated data, guided by a predictable sca…
Scaling Law for Time Series Forecasting
·2211 words·11 mins· loading · loading
Machine Learning Deep Learning 🏢 Tsinghua University
Unlocking the potential of deep learning for time series forecasting: this study reveals a scaling law influenced by dataset size, model complexity, and the crucial look-back horizon, leading to impro…
ScaleKD: Strong Vision Transformers Could Be Excellent Teachers
·3783 words·18 mins· loading · loading
AI Generated Computer Vision Image Classification 🏢 Intel Labs
ScaleKD: Pre-trained vision transformers make excellent teachers for diverse student networks, improving efficiency and performance in knowledge distillation.
Scale-invariant Optimal Sampling for Rare-events Data and Sparse Models
·2539 words·12 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 University of Connecticut
Scale-invariant optimal subsampling tackles computational challenges in analyzing massive rare-events data with sparse models, enhancing parameter estimation and variable selection without being affec…
Scalable Optimization in the Modular Norm
·2001 words·10 mins· loading · loading
Machine Learning Deep Learning 🏢 MIT
Deep learning optimization gets a major upgrade with Modula, a new method that uses the modular norm to normalize weight updates, enabling learning rate transfer across network widths and depths, thus…
Scalable Neural Network Verification with Branch-and-bound Inferred Cutting Planes
·2551 words·12 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of Illinois Urbana-Champaign
BICCOS: Scalable neural network verification via branch-and-bound inferred cutting planes.