Posters
2024
Is One GPU Enough? Pushing Image Generation at Higher-Resolutions with Foundation Models.
·3743 words·18 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 University of Glasgow
Pixelsmith: Generate gigapixel images with a single GPU, surpassing limitations of existing methods through a cascading approach and innovative guidance mechanism.
Is O(log N) practical? Near-Equivalence Between Delay Robustness and Bounded Regret in Bandits and RL
·403 words·2 mins·
loading
·
loading
AI Theory
Robustness
🏢 University of Washington
Zero Graves-Lai constant ensures both bounded regret and delay robustness in online decision-making, particularly for linear models.
Is Multiple Object Tracking a Matter of Specialization?
·2391 words·12 mins·
loading
·
loading
Computer Vision
Object Detection
🏢 University of Modena and Reggio Emilia
PASTA: A novel modular framework boosts MOT tracker generalization by using parameter-efficient fine-tuning and avoiding negative interference through specialized modules for various scenario attribut…
Is Mamba Compatible with Trajectory Optimization in Offline Reinforcement Learning?
·3014 words·15 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 National University of Defence Technology
Decision Mamba (DeMa) outperforms Decision Transformer (DT) in offline RL trajectory optimization with 30% fewer parameters in Atari and a quarter in MuJoCo, demonstrating the efficacy of Mamba’s line…
Is Knowledge Power? On the (Im)possibility of Learning from Strategic Interactions
·385 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 UC Berkeley
In strategic settings, repeated interactions alone may not enable uninformed players to achieve optimal outcomes, highlighting the persistent impact of information asymmetry.
Is Cross-validation the Gold Standard to Estimate Out-of-sample Model Performance?
·1790 words·9 mins·
loading
·
loading
AI Theory
Optimization
🏢 Columbia University
Cross-validation isn’t always superior; simple plug-in methods often perform equally well for estimating out-of-sample model performance, especially when considering computational costs.
Is A Picture Worth A Thousand Words? Delving Into Spatial Reasoning for Vision Language Models
·3090 words·15 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Microsoft Research
SpatialEval benchmark reveals that current vision-language models struggle with spatial reasoning, highlighting the need for improved multimodal models that effectively integrate visual and textual in…
IRCAN: Mitigating Knowledge Conflicts in LLM Generation via Identifying and Reweighting Context-Aware Neurons
·2251 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 College of Intelligence and Computing, Tianjin University
IRCAN tackles LLM knowledge conflicts by identifying and reweighting context-aware neurons, significantly improving context-sensitive outputs.
IR-CM: The Fast and Universal Image Restoration Method Based on Consistency Model
·2449 words·12 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Huazhong University of Science and Technology
IR-CM: One-step image restoration using a novel consistency model for fast and universal performance.
IQA-EVAL: Automatic Evaluation of Human-Model Interactive Question Answering
·3104 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Question Answering
🏢 University of Texas at Dallas
IQA-EVAL: An automatic evaluation framework uses LLMs to simulate human-AI interactions and evaluate interactive question answering, achieving high correlation with human judgments.
IPO: Interpretable Prompt Optimization for Vision-Language Models
·3712 words·18 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 AIM Lab, University of Amsterdam
This paper introduces IPO, a novel interpretable prompt optimizer for vision-language models. IPO uses large language models (LLMs) to dynamically generate human-understandable prompts, improving acc…
IPM-LSTM: A Learning-Based Interior Point Method for Solving Nonlinear Programs
·2991 words·15 mins·
loading
·
loading
AI Generated
AI Theory
Optimization
🏢 Xi'an Jiaotong University
IPM-LSTM accelerates nonlinear program solving by up to 70% using LSTM networks to approximate linear system solutions within the interior point method.
IODA: Instance-Guided One-shot Domain Adaptation for Super-Resolution
·2808 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 Nanjing University
IODA achieves efficient one-shot domain adaptation for super-resolution using a novel instance-guided strategy and image-level domain alignment, significantly improving performance with limited target…
Invertible Consistency Distillation for Text-Guided Image Editing in Around 7 Steps
·4066 words·20 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 HSE University
Invertible Consistency Distillation (iCD) achieves high-quality image editing in ~7 steps by enabling both fast editing and strong generation using a generalized distillation framework and dynamic cla…
InversionView: A General-Purpose Method for Reading Information from Neural Activations
·10684 words·51 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Saarland University
InversionView unveils neural network inner workings by decoding information from activations. It identifies inputs producing similar activations, revealing the information content. Case studies on v…
Inversion-based Latent Bayesian Optimization
·4093 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Optimization
🏢 Korea University
InvBO: Inversion-based Latent Bayesian Optimization solves the misalignment problem in LBO, boosting optimization accuracy and efficiency.
Inverse M-Kernels for Linear Universal Approximators of Non-Negative Functions
·1416 words·7 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 NTT Corporation
Unlocking efficient non-negative function approximation: This paper introduces inverse M-kernels, enabling flexible, linear universal approximators for one-dimensional inputs.
Inverse Factorized Soft Q-Learning for Cooperative Multi-agent Imitation Learning
·3040 words·15 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Singapore Management University
New multi-agent imitation learning algorithm (MIFQ) leverages inverse soft Q-learning and factorization for stable, efficient training, achieving state-of-the-art results on challenging benchmarks.
Invariant Tokenization of Crystalline Materials for Language Model Enabled Generation
·2045 words·10 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Texas A&M University
Mat2Seq revolutionizes crystal structure generation using language models by creating unique, invariant 1D sequences from 3D crystal structures, enabling accurate and efficient crystal discovery with …
Invariant subspaces and PCA in nearly matrix multiplication time
·336 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 IBM Research
Generalized eigenvalue problems get solved in nearly matrix multiplication time, providing new, faster PCA algorithms!