Posters
2024
MGF: Mixed Gaussian Flow for Diverse Trajectory Prediction
·2012 words·10 mins·
loading
·
loading
AI Applications
Robotics
🏢 Carnegie Mellon University
MGF: Mixed Gaussian Flow enhances trajectory prediction by using a mixed Gaussian prior, achieving state-of-the-art diversity and alignment accuracy.
MG-Net: Learn to Customize QAOA with Circuit Depth Awareness
·2515 words·12 mins·
loading
·
loading
AI Theory
Optimization
🏢 School of Computer Science, Faculty of Engineering, University of Sydney
MG-Net dynamically designs optimal mixer Hamiltonians for QAOA, overcoming the limitation of fixed-depth quantum circuits and significantly improving approximation ratios.
Metric Space Magnitude for Evaluating the Diversity of Latent Representations
·6876 words·33 mins·
loading
·
loading
AI Generated
AI Theory
Representation Learning
🏢 University of Edinburgh
Novel metric space magnitude measures rigorously quantify the diversity of latent representations across multiple scales, showing superior performance in detecting mode collapse and characterizing emb…
Metric from Human: Zero-shot Monocular Metric Depth Estimation via Test-time Adaptation
·4145 words·20 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 Carnegie Mellon University
Humans as landmarks: A novel zero-shot monocular metric depth estimation method leverages generative models and human mesh recovery to transfer metric scale information, achieving superior generalizat…
Metric Flow Matching for Smooth Interpolations on the Data Manifold
·2425 words·12 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University of Oxford
METRIC FLOW MATCHING (MFM) generates smooth interpolations on data manifolds by minimizing kinetic energy, outperforming Euclidean methods and achieving state-of-the-art results in single-cell traject…
Meteor: Mamba-based Traversal of Rationale for Large Language and Vision Models
·3263 words·16 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 KAIST
Meteor: Mamba-based Traversal of Rationale achieves significant vision-language improvements by efficiently embedding multifaceted rationales in a large language model, without scaling the model or us…
MetaUAS: Universal Anomaly Segmentation with One-Prompt Meta-Learning
·1904 words·9 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 Tencent Youtu Lab
MetaUAS achieves universal visual anomaly segmentation using only one normal image prompt via a pure vision model, surpassing previous zero-shot, few-shot, and full-shot methods.
MetaCURL: Non-stationary Concave Utility Reinforcement Learning
·362 words·2 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Inria
MetaCURL: First algorithm for non-stationary Concave Utility Reinforcement Learning (CURL), achieving near-optimal dynamic regret by using a meta-algorithm and sleeping experts framework.
Metacognitive Capabilities of LLMs: An Exploration in Mathematical Problem Solving
·2343 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Google DeepMind
LLMs gain math skills via prompt-guided skill labeling and exemplar selection, significantly boosting accuracy.
MetaAligner: Towards Generalizable Multi-Objective Alignment of Language Models
·3742 words·18 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Manchester
MetaAligner: a novel, policy-agnostic, and generalizable method for efficiently aligning LLMs to multiple objectives, even unseen ones, achieving significant and balanced improvements while saving up …
Meta-Reinforcement Learning with Universal Policy Adaptation: Provable Near-Optimality under All-task Optimum Comparator
·1665 words·8 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Pennsylvania State University
Provable near-optimality in meta-RL is achieved using a novel bilevel optimization framework and universal policy adaptation algorithm.
Meta-Learning Universal Priors Using Non-Injective Change of Variables
·2169 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Meta Learning
🏢 University of Minnesota
MetaNCoV: Learn data-driven priors via non-injective change of variables for enhanced few-shot learning.
Meta-Exploiting Frequency Prior for Cross-Domain Few-Shot Learning
·1996 words·10 mins·
loading
·
loading
Computer Vision
Few-Shot Learning
🏢 Northwestern Polytechnical University
Meta-Exploiting Frequency Prior enhances cross-domain few-shot learning by leveraging image frequency decomposition and consistency priors to improve model generalization and efficiency.
Meta-DT: Offline Meta-RL as Conditional Sequence Modeling with World Model Disentanglement
·4081 words·20 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Nanjing University
Meta-DT: Offline meta-RL masters unseen tasks via conditional sequence modeling and world model disentanglement, showcasing superior few-shot and zero-shot generalization.
Meta-Diffu$B$: A Contextualized Sequence-to-Sequence Text Diffusion Model with Meta-Exploration
·3164 words·15 mins·
loading
·
loading
AI Generated
Natural Language Processing
Text Generation
🏢 University of Washington
Meta-DiffuB enhances sequence-to-sequence text diffusion models by using meta-exploration to learn a contextualized noise schedule, resulting in state-of-the-art performance.
Meta-Controller: Few-Shot Imitation of Unseen Embodiments and Tasks in Continuous Control
·3389 words·16 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 School of Computing, KAIST
Meta-Controller: A novel few-shot behavior cloning framework enables robots to generalize to unseen embodiments and tasks using only a few reward-free demonstrations, showcasing superior few-shot gene…
Meta 3D AssetGen: Text-to-Mesh Generation with High-Quality Geometry, Texture, and PBR Materials
·2436 words·12 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Meta AI
Meta 3D AssetGen: High-quality text-to-mesh generation with realistic PBR materials and lighting, exceeding prior methods in speed and accuracy.
MeshXL: Neural Coordinate Field for Generative 3D Foundation Models
·2662 words·13 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 Tencent PCG
MeshXL: Autoregressively generating high-quality 3D meshes using a novel Neural Coordinate Field (NeurCF) representation and large language model approaches.
Mesa-Extrapolation: A Weave Position Encoding Method for Enhanced Extrapolation in LLMs
·3226 words·16 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Tsinghua University
Mesa-Extrapolation enhances LLM extrapolation using a novel weave position encoding method, boosting performance while significantly reducing memory and inference time.
MemVLT: Vision-Language Tracking with Adaptive Memory-based Prompts
·2803 words·14 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 School of Artificial Intelligence, University of Chinese Academy of Sciences
MemVLT: Adaptive Vision-Language Tracking leverages memory to generate dynamic prompts, surpassing existing methods by adapting to changing target appearances.