Posters
2024
Transfer Learning for Latent Variable Network Models
·1891 words·9 mins·
loading
·
loading
Machine Learning
Transfer Learning
🏢 University of Texas at Austin
This paper presents efficient algorithms for transfer learning in latent variable network models, achieving vanishing error under specific conditions, and attaining minimax optimal rates for stochasti…
Transductive Learning is Compact
·325 words·2 mins·
loading
·
loading
AI Theory
Optimization
🏢 USC
Supervised learning’s sample complexity is compact: a hypothesis class is learnable if and only if all its finite projections are learnable, simplifying complexity analysis.
Transductive Active Learning: Theory and Applications
·3403 words·16 mins·
loading
·
loading
Machine Learning
Active Learning
🏢 ETH Zurich
This paper introduces transductive active learning, proving its efficiency in minimizing uncertainty and achieving state-of-the-art results in neural network fine-tuning and safe Bayesian optimization…
Transcendence: Generative Models Can Outperform The Experts That Train Them
·2384 words·12 mins·
loading
·
loading
AI Theory
Generalization
🏢 OpenAI
Generative models can outperform their human trainers: A groundbreaking study shows how autoregressive transformers, trained on chess game data, can achieve higher game ratings than any of the human …
TransAgent: Transfer Vision-Language Foundation Models with Heterogeneous Agent Collaboration
·1922 words·10 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences
TransAgent empowers vision-language models by collaborating with diverse expert agents, achieving state-of-the-art performance in low-shot visual recognition.
Trajectory Diffusion for ObjectGoal Navigation
·2125 words·10 mins·
loading
·
loading
Multimodal Learning
Embodied AI
🏢 University of Chinese Academy of Sciences
Trajectory Diffusion (T-Diff) significantly improves object goal navigation by learning sequential planning through trajectory diffusion, resulting in more accurate and efficient navigation.
Trajectory Data Suffices for Statistically Efficient Learning in Offline RL with Linear q^π-Realizability and Concentrability
·479 words·3 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 University of Alberta
Offline RL with trajectory data achieves statistically efficient learning under linear q*-realizability and concentrability, solving a previously deemed impossible problem.
TrajCLIP: Pedestrian trajectory prediction method using contrastive learning and idempotent networks
·1889 words·9 mins·
loading
·
loading
AI Applications
Robotics
🏢 Institute of Computing Technology, University of Chinese Academy of Sciences
TrajCLIP: a novel pedestrian trajectory prediction method using contrastive learning and idempotent networks to achieve state-of-the-art performance and enhance generalization across diverse scenarios…
Training-Free Open-Ended Object Detection and Segmentation via Attention as Prompts
·1784 words·9 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Peking University
VL-SAM: Training-free open-ended object detection & segmentation using attention maps as prompts, surpassing previous methods on LVIS and CODA datasets.
Training-Free Adaptive Diffusion with Bounded Difference Approximation Strategy
·2325 words·11 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 Shanghai Artificial Intelligence Laboratory
AdaptiveDiffusion accelerates diffusion model inference by adaptively skipping noise prediction steps, achieving 2-5x speedup without quality loss.
Training for Stable Explanation for Free
·2565 words·13 mins·
loading
·
loading
AI Theory
Interpretability
🏢 Hong Kong University of Science and Technology
R2ET: training for robust ranking explanations by an effective regularizer.
Training Dynamics of Transformers to Recognize Word Co-occurrence via Gradient Flow Analysis
·426 words·2 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Princeton University
Researchers reveal how transformers learn word co-occurrence using a novel gradient flow analysis, uncovering a two-phase training process that leads to near-minimum loss and improved model performanc…
Training Binary Neural Networks via Gaussian Variational Inference and Low-Rank Semidefinite Programming
·1655 words·8 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
🏢 University of Chicago
VISPA, a novel BNN training framework using Gaussian variational inference and low-rank SDP, achieves state-of-the-art accuracy on various benchmarks.
Training an Open-Vocabulary Monocular 3D Detection Model without 3D Data
·3285 words·16 mins·
loading
·
loading
AI Generated
Computer Vision
3D Vision
🏢 Tsinghua University
Train open-vocabulary 3D object detectors using only RGB images and large language models, achieving state-of-the-art performance without expensive LiDAR data.
Train-Attention: Meta-Learning Where to Focus in Continual Knowledge Learning
·330 words·2 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Yonsei University
Train-Attention (TAALM) tackles catastrophic forgetting in LLMs by dynamically weighting tokens during training, boosting learning efficiency and knowledge retention, outperforming existing methods on…
Trading off Consistency and Dimensionality of Convex Surrogates for Multiclass Classification
·1544 words·8 mins·
loading
·
loading
AI Theory
Optimization
🏢 Harvard University
Researchers achieve a balance between accuracy and efficiency in multiclass classification by introducing partially consistent surrogate losses and novel methods.
Trade-Offs of Diagonal Fisher Information Matrix Estimators
·2685 words·13 mins·
loading
·
loading
AI Theory
Optimization
🏢 Australian National University
This paper examines the trade-offs between two popular diagonal Fisher Information Matrix (FIM) estimators in neural networks, deriving variance bounds and highlighting the importance of considering e…
TrAct: Making First-layer Pre-Activations Trainable
·2254 words·11 mins·
loading
·
loading
Computer Vision
Image Classification
🏢 Stanford University
TrAct boosts vision model training by directly optimizing first-layer activations, leading to significant speedups (1.25x-4x) and improved accuracy.
Trace is the Next AutoDiff: Generative Optimization with Rich Feedback, Execution Traces, and LLMs
·2686 words·13 mins·
loading
·
loading
AI Theory
Optimization
🏢 Microsoft Research
Trace: Automating AI workflow design with LLMs.
TPR: Topology-Preserving Reservoirs for Generalized Zero-Shot Learning
·2613 words·13 mins·
loading
·
loading
Multimodal Learning
Vision-Language Models
🏢 Xi'an Jiaotong University
Topology-Preserving Reservoirs (TPR) enhances CLIP’s zero-shot learning by using a dual-space alignment and a topology-preserving objective to improve generalization to unseen classes, achieving state…