Skip to main content

Posters

2024

Monoculture in Matching Markets
·1724 words·9 mins· loading · loading
AI Theory Fairness 🏢 Cornell Tech
Algorithmic monoculture harms applicant selection and market efficiency; this paper introduces a model to analyze its effects in two-sided matching markets.
MonkeySee: Space-time-resolved reconstructions of natural images from macaque multi-unit activity
·2882 words·14 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 Donders Institute for Brain, Cognition and Behaviour
MonkeySee reconstructs natural images from macaque brain signals with high accuracy using a novel CNN decoder, advancing neural decoding and offering insights into visual perception.
MoMu-Diffusion: On Learning Long-Term Motion-Music Synchronization and Correspondence
·2462 words·12 mins· loading · loading
Multimodal Learning Multimodal Generation 🏢 Zhejiang University
MoMu-Diffusion: a novel framework that learns long-term motion-music synchronization, generating realistic and beat-matched sequences surpassing existing methods.
MomentumSMoE: Integrating Momentum into Sparse Mixture of Experts
·2790 words·14 mins· loading · loading
Natural Language Processing Large Language Models 🏢 National University of Singapore
MomentumSMoE boosts Sparse Mixture of Experts’ (SMoE) performance by integrating momentum, resulting in more stable training and robust models.
MoME: Mixture of Multimodal Experts for Generalist Multimodal Large Language Models
·2146 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Harbin Institute of Technology, Shenzhen
MoME, a novel Mixture of Multimodal Experts, significantly improves generalist Multimodal Large Language Models (MLLMs) by mitigating task interference through specialized vision and language experts,…
Molecule Generation with Fragment Retrieval Augmentation
·2469 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 KAIST
f-RAG: A novel fragment-based molecular generation framework boosts drug discovery by combining retrieval augmentation with a generative model, enabling exploration beyond existing fragments and signi…
MoLE: Enhancing Human-centric Text-to-image Diffusion via Mixture of Low-rank Experts
·4224 words·20 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Peking University
MoLE: Mixture of Low-rank Experts enhances human-centric text-to-image diffusion models by using low-rank modules trained on high-quality face and hand datasets to improve the realism of faces and han…
MoGU: A Framework for Enhancing Safety of LLMs While Preserving Their Usability
·2311 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Harbin Institute of Technology
MoGU: A framework dynamically balances safety and usability in LLMs by routing benign and malicious instructions to different LLM variants, leading to safer, more useful responses.
MoGenTS: Motion Generation based on Spatial-Temporal Joint Modeling
·2749 words·13 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Alibaba Group
MoGenTS revolutionizes human motion generation by quantizing individual joints into 2D tokens, enabling efficient spatial-temporal modeling and significantly outperforming existing methods.
MoEUT: Mixture-of-Experts Universal Transformers
·2486 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Stanford University
MoEUT: Mixture-of-Experts Universal Transformers significantly improves the compute efficiency of Universal Transformers, making them competitive with standard Transformers in large-scale language mod…
MoE Jetpack: From Dense Checkpoints to Adaptive Mixture of Experts for Vision Tasks
·2159 words·11 mins· loading · loading
Computer Vision Image Classification 🏢 Huazhong University of Science and Technology
MoE Jetpack efficiently transforms readily available dense checkpoints into high-performing MoE models, drastically accelerating convergence and improving accuracy.
Modeling Latent Neural Dynamics with Gaussian Process Switching Linear Dynamical Systems
·1594 words·8 mins· loading · loading
Machine Learning Deep Learning 🏢 Stanford University
gpSLDS, a novel model, balances expressiveness and interpretability in modeling complex neural dynamics by combining Gaussian processes with switching linear dynamical systems, improving accuracy and …
Model-free Low-Rank Reinforcement Learning via Leveraged Entry-wise Matrix Estimation
·1610 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 KTH
LoRa-PI: a model-free RL algorithm learns and exploits low-rank MDP structures for order-optimal sample complexity, achieving ε-optimal policies with O(poly(A)) samples.
Model-Based Transfer Learning for Contextual Reinforcement Learning
·3972 words·19 mins· loading · loading
AI Applications Smart Cities 🏢 MIT
Model-Based Transfer Learning (MBTL) boosts deep RL sample efficiency by strategically selecting training tasks, achieving up to 50x improvement over traditional methods.
Model-based Diffusion for Trajectory Optimization
·2272 words·11 mins· loading · loading
AI Applications Robotics 🏢 Carnegie Mellon University
Model-Based Diffusion (MBD) uses diffusion processes and model information for data-free trajectory optimization, outperforming existing methods on complex tasks.
Model Sensitivity Aware Continual Learning
·2062 words·10 mins· loading · loading
Machine Learning Continual Learning 🏢 University of Maryland College Park
Model Sensitivity Aware Continual Learning (MACL) tackles the CL challenge by optimizing model performance based on parameter distribution, achieving superior old knowledge retention and new task perf…
Model Reconstruction Using Counterfactual Explanations: A Perspective From Polytope Theory
·4015 words·19 mins· loading · loading
AI Generated AI Theory Interpretability 🏢 University of Maryland
Counterfactual Clamping Attack (CCA) improves model reconstruction using counterfactual explanations by leveraging decision boundary proximity, offering theoretical guarantees and enhanced fidelity.
Model LEGO: Creating Models Like Disassembling and Assembling Building Blocks
·2425 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 Zhejiang University
Model LEGO (MDA) revolutionizes deep learning by enabling the creation of new models by assembling and disassembling task-aware components from pre-trained models, eliminating the need for retraining.
Model Decides How to Tokenize: Adaptive DNA Sequence Tokenization with MxDNA
·2786 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Shanghai Artificial Intelligence Laboratory
MxDNA: Model learns optimal DNA tokenization via gradient descent, outperforming existing methods.
Model Collapse Demystified: The Case of Regression
·1683 words·8 mins· loading · loading
AI Theory Generalization 🏢 Meta
Training AI models on AI-generated data leads to performance degradation, known as model collapse. This paper offers analytical formulas that precisely quantify this effect in high-dimensional regress…