Posters
2024
Interfacing Foundation Models' Embeddings
·2676 words·13 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 UW-Madison
FIND, a lightweight transformer interface, seamlessly aligns foundation models’ embeddings for unified image and dataset-level understanding, enabling generalizable, interleaved performance on segment…
InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction
·2177 words·11 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Illinois Urbana-Champaign
InterDreamer: Zero-shot text-guided 3D human-object interaction generation without paired data, achieved via decoupled semantic and dynamic modeling, using LLMs and a physics-based world model.
InterControl: Zero-shot Human Interaction Generation by Controlling Every Joint
·2703 words·13 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 Chinese University of Hong Kong
InterControl: Zero-shot multi-person interaction generation by precisely controlling every joint using only single-person data.
Interactive Deep Clustering via Value Mining
·1729 words·9 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Sichuan University
Interactive Deep Clustering (IDC) significantly boosts deep clustering performance by strategically incorporating minimal user interaction to resolve ambiguous sample classifications.
Interaction-Force Transport Gradient Flows
·1588 words·8 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Humboldt University of Berlin
New gradient flow geometry improves MMD-based sampling by teleporting particle mass, guaranteeing global exponential convergence, and yielding superior empirical results.
Integrating Suboptimal Human Knowledge with Hierarchical Reinforcement Learning for Large-Scale Multiagent Systems
·2222 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Wollongong
Hierarchical Human Knowledge-guided MARL (hhk-MARL) framework accelerates large-scale multi-agent training by integrating suboptimal human knowledge, significantly improving performance and scalabilit…
Integrating GNN and Neural ODEs for Estimating Non-Reciprocal Two-Body Interactions in Mixed-Species Collective Motion
·1573 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 University of Tokyo
Deep learning framework integrating GNNs and neural ODEs precisely estimates non-reciprocal two-body interactions in mixed-species collective motion, accurately replicating both individual and collect…
Integrating Deep Metric Learning with Coreset for Active Learning in 3D Segmentation
·2210 words·11 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 UC Los Angeles
Deep metric learning and Coreset integration enables efficient slice-based active learning for 3D medical segmentation, surpassing existing methods in performance with low annotation budgets.
Instructor-inspired Machine Learning for Robust Molecular Property Prediction
·2041 words·10 mins·
loading
·
loading
Machine Learning
Semi-Supervised Learning
🏢 Stanford University
InstructMol, a novel semi-supervised learning algorithm, leverages unlabeled data and an instructor model to significantly improve the accuracy and robustness of molecular property prediction, even wi…
Instruction-Guided Visual Masking
·3666 words·18 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
🏢 Tsinghua University
Instruction-Guided Visual Masking (IVM) boosts multimodal instruction following by precisely focusing models on relevant image regions via visual masking, achieving state-of-the-art results on multipl…
Instruction Tuning With Loss Over Instructions
·4022 words·19 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University College London
Boost LLM performance with INSTRUCTION MODELLING: a simple yet effective instruction tuning method that improves model outputs by over 100% in some cases by applying loss to both instructions and outp…
InstructG2I: Synthesizing Images from Multimodal Attributed Graphs
·1973 words·10 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 University of Illinois at Urbana-Champaign
INSTRUCTG2I: a novel graph context-conditioned diffusion model, generates images from multimodal attributed graphs, addressing challenges in graph size, dependencies, and controllability.
Instance-Specific Asymmetric Sensitivity in Differential Privacy
·1985 words·10 mins·
loading
·
loading
AI Theory
Privacy
🏢 Mozilla
New algorithm improves differentially private estimations by adapting to dataset hardness, enhancing accuracy for variance, classification, and regression tasks.
Instance-Optimal Private Density Estimation in the Wasserstein Distance
·338 words·2 mins·
loading
·
loading
AI Theory
Privacy
🏢 Apple
Instance-optimal private density estimation algorithms, adapting to data characteristics for improved accuracy in the Wasserstein distance, are introduced.
Instance-adaptive Zero-shot Chain-of-Thought Prompting
·2222 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 College of Computer Science and Technology, Jilin University
Instance-adaptive prompting significantly improves zero-shot Chain-of-Thought reasoning in LLMs by dynamically selecting prompts tailored to each instance, leading to consistent performance gains acro…
Injecting Undetectable Backdoors in Obfuscated Neural Networks and Language Models
·372 words·2 mins·
loading
·
loading
AI Theory
Robustness
🏢 Yale University
Researchers developed a novel method to inject undetectable backdoors into obfuscated neural networks and language models, even with white-box access, posing significant security risks.
Initializing Variable-sized Vision Transformers from Learngene with Learnable Transformation
·2536 words·12 mins·
loading
·
loading
AI Generated
Computer Vision
Image Classification
🏢 School of Computer Science and Engineering, Southeast University
LeTs: Learnable Transformation efficiently initializes variable-sized Vision Transformers by learning adaptable transformations from a compact learngene module, outperforming from-scratch training.
Initializing Services in Interactive ML Systems for Diverse Users
·1498 words·8 mins·
loading
·
loading
Machine Learning
Federated Learning
🏢 University of Washington
Adaptively initializing multi-service ML systems for diverse users using minimal data, this paper introduces a randomized algorithm achieving near-optimal loss with provable guarantees.
Initialization is Critical to Whether Transformers Fit Composite Functions by Reasoning or Memorizing
·2409 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Shanghai Jiao Tong University
Transformer model initialization dramatically affects whether it reasons or memorizes, impacting performance on compositional tasks.
Infusing Self-Consistency into Density Functional Theory Hamiltonian Prediction via Deep Equilibrium Models
·1907 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Microsoft Research
Deep Equilibrium Models (DEQs) infused into DFT Hamiltonian prediction achieves self-consistency, accelerating large-scale materials simulations.