Skip to main content

Posters

2024

MambaAD: Exploring State Space Models for Multi-class Unsupervised Anomaly Detection
·3318 words·16 mins· loading · loading
AI Generated Computer Vision Anomaly Detection 🏢 Zhejiang University Youtu Lab
MambaAD: Linear-complexity multi-class unsupervised anomaly detection using a novel Mamba-based decoder with Locality-Enhanced State Space modules.
MALT Powers Up Adversarial Attacks
·1855 words·9 mins· loading · loading
AI Theory Robustness 🏢 Weizmann Institute of Science
MALT: a novel adversarial attack, is 5x faster than AutoAttack, achieving higher success rates on CIFAR-100 and ImageNet by exploiting mesoscopic almost linearity in neural networks.
Making Offline RL Online: Collaborative World Models for Offline Visual Reinforcement Learning
·3140 words·15 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 MoE Key Lab of Artificial Intelligence
CoWorld: a novel model-based RL approach tackles offline visual RL challenges by using online simulators as testbeds, enabling flexible value estimation & mitigating overestimation bias for effective …
Make-it-Real: Unleashing Large Multimodal Model for Painting 3D Objects with Realistic Materials
·4579 words·22 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Stanford University
Make-it-Real uses a large multimodal language model to automatically paint realistic materials onto 3D objects, drastically improving realism and saving developers time.
Make-An-Agent: A Generalizable Policy Network Generator with Behavior-Prompted Diffusion
·3025 words·15 mins· loading · loading
AI Generated AI Applications Robotics 🏢 Tsinghua University
Make-An-Agent generates high-performing robotic control policies from single behavioral demonstrations using behavior-prompted diffusion, showcasing impressive generalization and real-world applicabil…
Make Your LLM Fully Utilize the Context
·2445 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Microsoft
FILM-7B, trained with Information-Intensive (IN2) training, significantly overcomes the ’lost-in-the-middle’ problem in long-context LLMs, enabling robust information retrieval from all context positi…
Make Continual Learning Stronger via C-Flat
·2055 words·10 mins· loading · loading
Machine Learning Continual Learning 🏢 Tsinghua University
Boost continual learning with C-Flat: a novel, one-line-code optimizer creating flatter loss landscapes for enhanced stability and generalization across various continual learning scenarios.
Maia-2: A Unified Model for Human-AI Alignment in Chess
·2577 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 University of Toronto
Maia-2: A unified model for human-AI alignment in chess, coherently captures human play across skill levels, significantly improving AI-human alignment and paving the way for AI-guided teaching.
MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
·1848 words·9 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University at Albany, SUNY
MagR: a novel preprocessing technique boosts post-training quantization of LLMs by reducing weight magnitudes without inference overhead, achieving state-of-the-art performance.
Magnet: We Never Know How Text-to-Image Diffusion Models Work, Until We Learn How Vision-Language Models Function
·3895 words·19 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Nanjing University of Aeronautics and Astronautics
Magnet: Enhancing Text-to-Image Synthesis by Disentangling Attributes in CLIP.
MAGNET: Improving the Multilingual Fairness of Language Models with Adaptive Gradient-Based Tokenization
·2236 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Washington
MAGNET, a novel adaptive gradient-based tokenization method, tackles multilingual language model bias by employing language-specific boundary predictors to achieve equitable segmentation across divers…
MAGIS: LLM-Based Multi-Agent Framework for GitHub Issue Resolution
·3263 words·16 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Hong Kong
MAGIS: A novel LLM-based multi-agent framework significantly boosts GitHub issue resolution by leveraging agent collaboration for planning and coding, achieving an eight-fold performance increase comp…
MADiff: Offline Multi-agent Learning with Diffusion Models
·2719 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Shanghai Jiao Tong University
MADIFF: Offline multi-agent learning uses attention-based diffusion models to achieve effective coordination and teammate modeling, outperforming existing methods.
MACM: Utilizing a Multi-Agent System for Condition Mining in Solving Complex Mathematical Problems
·2015 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Minnesota
Multi-Agent System for Condition Mining (MACM) dramatically boosts large language model accuracy in complex math problem-solving, exceeding existing methods by achieving higher accuracy and better gen…
MAC Advice for facility location mechanism design
·1881 words·9 mins· loading · loading
AI Theory Robustness 🏢 Tel Aviv University
Improved facility location mechanisms are designed using ‘Mostly Approximately Correct’ predictions, exceeding prior bounds despite large prediction errors.
M$^3$GPT: An Advanced Multimodal, Multitask Framework for Motion Comprehension and Generation
·4046 words·19 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Tencent AI Lab
M³GPT, a novel multimodal framework, achieves superior motion comprehension and generation by integrating text, music, and motion data into a unified LLM representation.
LuSh-NeRF: Lighting up and Sharpening NeRFs for Low-light Scenes
·2414 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 City University of Hong Kong
LuSh-NeRF: A novel model reconstructs sharp, bright NeRFs from hand-held low-light photos by sequentially modeling and removing noise and blur, outperforming existing methods.
Lumina-Next : Making Lumina-T2X Stronger and Faster with Next-DiT
·2931 words·14 mins· loading · loading
Multimodal Learning Multimodal Generation 🏢 Beijing University of Posts and Telecommunications
Lumina-Next supercharges image generation: faster, more efficient, and better resolution with new architecture and sampling techniques.
Lumen: Unleashing Versatile Vision-Centric Capabilities of Large Multimodal Models
·2500 words·12 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Fudan University
Lumen: A novel LMM architecture decouples perception learning into task-agnostic and task-specific stages, enabling versatile vision-centric capabilities and surpassing existing LMM-based approaches.
LT-Defense: Searching-free Backdoor Defense via Exploiting the Long-tailed Effect
·2148 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Beijing University of Posts and Telecommunications
LT-Defense: a searching-free backdoor defense for language models leveraging the long-tailed effect of poisoned data. It achieves 98% accuracy across 1440 models with less than 1% time cost of existin…