Posters
2024
Catastrophic Goodhart: regularizing RLHF with KL divergence does not mitigate heavy-tailed reward misspecification
·2257 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Independent / FAR Labs
RLHF’s KL regularization fails to prevent ‘catastrophic Goodhart’—policies achieving high proxy reward but low actual utility—when reward errors have heavy tails.
CAT: Coordinating Anatomical-Textual Prompts for Multi-Organ and Tumor Segmentation
·2794 words·14 mins·
loading
·
loading
AI Generated
Computer Vision
Image Segmentation
🏢 Qing Yuan Research Institute, Shanghai Jiao Tong University
CAT: A novel dual-prompt model coordinates anatomical and textual prompts for superior multi-organ & tumor segmentation in medical imaging, overcoming limitations of single-prompt methods.
Cascade Speculative Drafting for Even Faster LLM Inference
·1806 words·9 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 University of Illinois at Urbana-Champaign
Cascade Speculative Drafting (CS Drafting) dramatically speeds up large language model inference by using a multi-stage drafting process, optimizing both time allocation and autoregressive generation.
Cascade of phase transitions in the training of energy-based models
·1743 words·9 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Université PSL
Energy-based models’ training reveals a cascade of phase transitions, progressively learning data features, offering new insights into deep learning dynamics.
Carrot and Stick: Eliciting Comparison Data and Beyond
·1825 words·9 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
🏢 Harvard University
Truthful comparison data is hard to obtain without ground truth. This paper presents novel peer prediction mechanisms using bonus-penalty payments that incentivize truthful comparisons, even in networ…
Cardinality-Aware Set Prediction and Top-$k$ Classification
·1676 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Google Research
This paper proposes cardinality-aware top-k classification, improving accuracy and efficiency by dynamically adjusting prediction set sizes.
Capturing the denoising effect of PCA via compression ratio
·2544 words·12 mins·
loading
·
loading
Machine Learning
Unsupervised Learning
🏢 Computer Science, University of Southern California
PCA’s denoising effect is quantified via a novel metric: compression ratio. This metric reveals PCA’s ability to reduce intra-community distances while preserving inter-community distances in noisy d…
Can We Leave Deepfake Data Behind in Training Deepfake Detector?
·2627 words·13 mins·
loading
·
loading
Computer Vision
Face Recognition
🏢 Tencent AI Lab
ProDet: Deepfake detection enhanced by progressively organizing blendfake and deepfake data in the latent space, improving generalization and robustness.
Can Simple Averaging Defeat Modern Watermarks?
·3146 words·15 mins·
loading
·
loading
Computer Vision
Image Generation
🏢 National University of Singapore
Simple averaging of watermarked images reveals hidden patterns, enabling watermark removal and forgery, thus highlighting the vulnerability of content-agnostic watermarking methods.
Can neural operators always be continuously discretized?
·380 words·2 mins·
loading
·
loading
AI Generated
AI Theory
Generalization
🏢 Shimane University
Neural operators’ continuous discretization is proven impossible in general Hilbert spaces, but achievable using strongly monotone operators, opening new avenues for numerical methods in scientific ma…
Can Models Learn Skill Composition from Examples?
·3161 words·15 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Princeton University
Smaller language models can learn skill composition from limited examples, substantially improving their ability to combine skills in novel ways through fine-tuning.
Can LLMs Learn by Teaching for Better Reasoning? A Preliminary Study
·5164 words·25 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Tsinghua University
LLMs can improve reasoning by teaching weaker models, a process called Learning by Teaching (LbT), as shown in this preliminary study. LbT enhances not just student models, but also the teacher model…
Can LLMs Implicitly Learn Numeric Parameter Constraints in Data Science APIs?
·2762 words·13 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Illinois Urbana-Champaign
LLMs struggle to reliably generate valid data science code due to a lack of true understanding of numerical constraints in APIs, despite seemingly mastering common patterns through extensive training.
Can large language models explore in-context?
·4498 words·22 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Microsoft Research
LLMs struggle with in-context exploration, needing substantial prompt engineering or training interventions to effectively explore multi-armed bandit environments.
Can Large Language Model Agents Simulate Human Trust Behavior?
·3567 words·17 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 University of Oxford
LLM agents surprisingly exhibit human-like trust behavior, especially GPT-4, paving the way for simulating complex human interactions in various applications.
Can Language Models Perform Robust Reasoning in Chain-of-thought Prompting with Noisy Rationales?
·7400 words·35 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Hong Kong Baptist University
LLMs struggle with noisy rationales in chain-of-thought prompting. This paper introduces the NoRa dataset, showing that existing methods struggle. A new method, CD-CoT, significantly improves accura…
Can Language Models Learn to Skip Steps?
·2929 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 UC Santa Barbara
Language models learn to skip steps in reasoning, improving efficiency and generalization, showcasing emergent human-like cognitive abilities.
Can Graph Neural Networks Expose Training Data Properties? An Efficient Risk Assessment Approach
·1856 words·9 mins·
loading
·
loading
AI Theory
Privacy
🏢 Zhejiang University
New efficient attack reveals GNN model training data properties.
Can Graph Learning Improve Planning in LLM-based Agents?
·2929 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
GNNs enhance LLM-based task planning by improving the ability to process task graphs, surpassing existing solutions even without training.
Can an AI Agent Safely Run a Government? Existence of Probably Approximately Aligned Policies
·506 words·3 mins·
loading
·
loading
AI Theory
Safety
🏢 ETH Zurich
This paper introduces a novel quantitative definition of AI alignment for social decision-making, proposing probably approximately aligned policies and a method to safeguard any autonomous agent’s act…